Compiler.cpp 19 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530
  1. /*
  2. * Copyright (c) 2023, Andreas Kling <kling@serenityos.org>
  3. *
  4. * SPDX-License-Identifier: BSD-2-Clause
  5. */
  6. #include <AK/OwnPtr.h>
  7. #include <LibJS/Bytecode/Instruction.h>
  8. #include <LibJS/Bytecode/Interpreter.h>
  9. #include <LibJS/JIT/Compiler.h>
  10. #include <LibJS/Runtime/VM.h>
  11. #include <LibJS/Runtime/ValueInlines.h>
  12. #include <sys/mman.h>
  13. #include <unistd.h>
  14. #define TRY_OR_SET_EXCEPTION(expression) \
  15. ({ \
  16. /* Ignore -Wshadow to allow nesting the macro. */ \
  17. AK_IGNORE_DIAGNOSTIC("-Wshadow", \
  18. auto&& _temporary_result = (expression)); \
  19. static_assert(!::AK::Detail::IsLvalueReference<decltype(_temporary_result.release_value())>, \
  20. "Do not return a reference from a fallible expression"); \
  21. if (_temporary_result.is_error()) [[unlikely]] { \
  22. vm.bytecode_interpreter().reg(Bytecode::Register::exception()) = _temporary_result.release_error().value().value(); \
  23. return {}; \
  24. } \
  25. _temporary_result.release_value(); \
  26. })
  27. namespace JS::JIT {
  28. void Compiler::store_vm_register(Bytecode::Register dst, Assembler::Reg src)
  29. {
  30. m_assembler.mov(
  31. Assembler::Operand::Mem64BaseAndOffset(REGISTER_ARRAY_BASE, dst.index() * sizeof(Value)),
  32. Assembler::Operand::Register(src));
  33. }
  34. void Compiler::load_vm_register(Assembler::Reg dst, Bytecode::Register src)
  35. {
  36. m_assembler.mov(
  37. Assembler::Operand::Register(dst),
  38. Assembler::Operand::Mem64BaseAndOffset(REGISTER_ARRAY_BASE, src.index() * sizeof(Value)));
  39. }
  40. void Compiler::store_vm_local(size_t dst, Assembler::Reg src)
  41. {
  42. m_assembler.mov(
  43. Assembler::Operand::Mem64BaseAndOffset(LOCALS_ARRAY_BASE, dst * sizeof(Value)),
  44. Assembler::Operand::Register(src));
  45. }
  46. void Compiler::load_vm_local(Assembler::Reg dst, size_t src)
  47. {
  48. m_assembler.mov(
  49. Assembler::Operand::Register(dst),
  50. Assembler::Operand::Mem64BaseAndOffset(LOCALS_ARRAY_BASE, src * sizeof(Value)));
  51. }
  52. void Compiler::compile_load_immediate(Bytecode::Op::LoadImmediate const& op)
  53. {
  54. m_assembler.mov(
  55. Assembler::Operand::Register(GPR0),
  56. Assembler::Operand::Imm64(op.value().encoded()));
  57. store_vm_register(Bytecode::Register::accumulator(), GPR0);
  58. }
  59. void Compiler::compile_load(Bytecode::Op::Load const& op)
  60. {
  61. load_vm_register(GPR0, op.src());
  62. store_vm_register(Bytecode::Register::accumulator(), GPR0);
  63. }
  64. void Compiler::compile_store(Bytecode::Op::Store const& op)
  65. {
  66. load_vm_register(GPR0, Bytecode::Register::accumulator());
  67. store_vm_register(op.dst(), GPR0);
  68. }
  69. void Compiler::compile_get_local(Bytecode::Op::GetLocal const& op)
  70. {
  71. load_vm_local(GPR0, op.index());
  72. store_vm_register(Bytecode::Register::accumulator(), GPR0);
  73. }
  74. void Compiler::compile_set_local(Bytecode::Op::SetLocal const& op)
  75. {
  76. load_vm_register(GPR0, Bytecode::Register::accumulator());
  77. store_vm_local(op.index(), GPR0);
  78. }
  79. void Compiler::compile_jump(Bytecode::Op::Jump const& op)
  80. {
  81. m_assembler.jump(const_cast<Bytecode::BasicBlock&>(op.true_target()->block()));
  82. }
  83. static bool cxx_to_boolean(VM&, Value value)
  84. {
  85. return value.to_boolean();
  86. }
  87. void Compiler::compile_to_boolean(Assembler::Reg dst, Assembler::Reg src)
  88. {
  89. // dst = src;
  90. m_assembler.mov(
  91. Assembler::Operand::Register(dst),
  92. Assembler::Operand::Register(src));
  93. // dst >>= 48;
  94. m_assembler.shift_right(
  95. Assembler::Operand::Register(dst),
  96. Assembler::Operand::Imm8(48));
  97. // if (dst != BOOLEAN_TAG) goto slow_case;
  98. auto slow_case = m_assembler.make_label();
  99. m_assembler.jump_if_not_equal(
  100. Assembler::Operand::Register(dst),
  101. Assembler::Operand::Imm32(BOOLEAN_TAG),
  102. slow_case);
  103. // Fast path for JS::Value booleans.
  104. // dst = src;
  105. m_assembler.mov(
  106. Assembler::Operand::Register(dst),
  107. Assembler::Operand::Register(src));
  108. // dst &= 1;
  109. m_assembler.bitwise_and(
  110. Assembler::Operand::Register(dst),
  111. Assembler::Operand::Imm32(1));
  112. // goto end;
  113. auto end = m_assembler.jump();
  114. // slow_case: // call C++ helper
  115. slow_case.link(m_assembler);
  116. m_assembler.mov(
  117. Assembler::Operand::Register(ARG1),
  118. Assembler::Operand::Register(src));
  119. m_assembler.native_call((void*)cxx_to_boolean);
  120. m_assembler.mov(
  121. Assembler::Operand::Register(dst),
  122. Assembler::Operand::Register(RET));
  123. // end:
  124. end.link(m_assembler);
  125. }
  126. void Compiler::compile_jump_conditional(Bytecode::Op::JumpConditional const& op)
  127. {
  128. load_vm_register(GPR1, Bytecode::Register::accumulator());
  129. compile_to_boolean(GPR0, GPR1);
  130. m_assembler.jump_conditional(GPR0,
  131. const_cast<Bytecode::BasicBlock&>(op.true_target()->block()),
  132. const_cast<Bytecode::BasicBlock&>(op.false_target()->block()));
  133. }
  134. [[maybe_unused]] static Value cxx_less_than(VM& vm, Value lhs, Value rhs)
  135. {
  136. return TRY_OR_SET_EXCEPTION(less_than(vm, lhs, rhs));
  137. }
  138. void Compiler::compile_less_than(Bytecode::Op::LessThan const& op)
  139. {
  140. load_vm_register(ARG1, op.lhs());
  141. load_vm_register(ARG2, Bytecode::Register::accumulator());
  142. m_assembler.native_call((void*)cxx_less_than);
  143. store_vm_register(Bytecode::Register::accumulator(), RET);
  144. check_exception();
  145. }
  146. [[maybe_unused]] static Value cxx_increment(VM& vm, Value value)
  147. {
  148. auto old_value = TRY_OR_SET_EXCEPTION(value.to_numeric(vm));
  149. if (old_value.is_number())
  150. return Value(old_value.as_double() + 1);
  151. return BigInt::create(vm, old_value.as_bigint().big_integer().plus(Crypto::SignedBigInteger { 1 }));
  152. }
  153. void Compiler::compile_increment(Bytecode::Op::Increment const&)
  154. {
  155. load_vm_register(ARG1, Bytecode::Register::accumulator());
  156. m_assembler.native_call((void*)cxx_increment);
  157. store_vm_register(Bytecode::Register::accumulator(), RET);
  158. check_exception();
  159. }
  160. void Compiler::check_exception()
  161. {
  162. // if (exception.is_empty()) goto no_exception;
  163. load_vm_register(GPR0, Bytecode::Register::exception());
  164. m_assembler.mov(Assembler::Operand::Register(GPR1), Assembler::Operand::Imm64(Value().encoded()));
  165. auto no_exception = m_assembler.make_label();
  166. m_assembler.jump_if_equal(Assembler::Operand::Register(GPR0), Assembler::Operand::Register(GPR1), no_exception);
  167. // We have an exception!
  168. // if (!unwind_context.valid) return;
  169. auto handle_exception = m_assembler.make_label();
  170. m_assembler.mov(
  171. Assembler::Operand::Register(GPR0),
  172. Assembler::Operand::Mem64BaseAndOffset(UNWIND_CONTEXT_BASE, 0));
  173. m_assembler.jump_if_not_equal(
  174. Assembler::Operand::Register(GPR0),
  175. Assembler::Operand::Imm32(0),
  176. handle_exception);
  177. m_assembler.exit();
  178. // handle_exception:
  179. handle_exception.link(m_assembler);
  180. // if (unwind_context.handler) {
  181. // accumulator = exception;
  182. // exception = Value();
  183. // goto handler;
  184. // }
  185. auto no_handler = m_assembler.make_label();
  186. m_assembler.mov(
  187. Assembler::Operand::Register(GPR0),
  188. Assembler::Operand::Mem64BaseAndOffset(UNWIND_CONTEXT_BASE, 8));
  189. m_assembler.jump_if_equal(
  190. Assembler::Operand::Register(GPR0),
  191. Assembler::Operand::Imm32(0),
  192. no_handler);
  193. load_vm_register(GPR1, Bytecode::Register::exception());
  194. store_vm_register(Bytecode::Register::accumulator(), GPR1);
  195. m_assembler.mov(
  196. Assembler::Operand::Register(GPR1),
  197. Assembler::Operand::Imm64(Value().encoded()));
  198. store_vm_register(Bytecode::Register::exception(), GPR1);
  199. m_assembler.jump(Assembler::Operand::Register(GPR0));
  200. // no_handler:
  201. no_handler.link(m_assembler);
  202. // if (unwind_context.finalizer) goto finalizer;
  203. auto no_finalizer = m_assembler.make_label();
  204. m_assembler.mov(
  205. Assembler::Operand::Register(GPR0),
  206. Assembler::Operand::Mem64BaseAndOffset(UNWIND_CONTEXT_BASE, 16));
  207. m_assembler.jump_if_equal(
  208. Assembler::Operand::Register(GPR0),
  209. Assembler::Operand::Imm32(0),
  210. no_finalizer);
  211. m_assembler.jump(Assembler::Operand::Register(GPR0));
  212. // no_finalizer:
  213. // NOTE: No catch and no finally!? Crash.
  214. no_finalizer.link(m_assembler);
  215. m_assembler.verify_not_reached();
  216. // no_exception:
  217. no_exception.link(m_assembler);
  218. }
  219. void Compiler::push_unwind_context(bool valid, Optional<Bytecode::Label> const& handler, Optional<Bytecode::Label> const& finalizer)
  220. {
  221. // Put this on the stack, and then point UNWIND_CONTEXT_BASE at it.
  222. // struct {
  223. // u64 valid;
  224. // u64 handler;
  225. // u64 finalizer;
  226. // };
  227. // push finalizer (patched later)
  228. m_assembler.mov(
  229. Assembler::Operand::Register(GPR0),
  230. Assembler::Operand::Imm64(0));
  231. if (finalizer.has_value())
  232. const_cast<Bytecode::BasicBlock&>(finalizer.value().block()).absolute_references_to_here.append(m_assembler.m_output.size() - 8);
  233. m_assembler.push(Assembler::Operand::Register(GPR0));
  234. // push handler (patched later)
  235. m_assembler.mov(
  236. Assembler::Operand::Register(GPR0),
  237. Assembler::Operand::Imm64(0));
  238. if (handler.has_value())
  239. const_cast<Bytecode::BasicBlock&>(handler.value().block()).absolute_references_to_here.append(m_assembler.m_output.size() - 8);
  240. m_assembler.push(Assembler::Operand::Register(GPR0));
  241. // push valid
  242. m_assembler.push(Assembler::Operand::Imm32(valid));
  243. // UNWIND_CONTEXT_BASE = STACK_POINTER
  244. m_assembler.mov(
  245. Assembler::Operand::Register(UNWIND_CONTEXT_BASE),
  246. Assembler::Operand::Register(STACK_POINTER));
  247. // align stack pointer
  248. m_assembler.sub(Assembler::Operand::Register(STACK_POINTER), Assembler::Operand::Imm8(8));
  249. }
  250. void Compiler::pop_unwind_context()
  251. {
  252. m_assembler.add(Assembler::Operand::Register(STACK_POINTER), Assembler::Operand::Imm8(32));
  253. m_assembler.add(Assembler::Operand::Register(UNWIND_CONTEXT_BASE), Assembler::Operand::Imm8(32));
  254. }
  255. void Compiler::compile_enter_unwind_context(Bytecode::Op::EnterUnwindContext const& op)
  256. {
  257. push_unwind_context(true, op.handler_target(), op.finalizer_target());
  258. m_assembler.jump(const_cast<Bytecode::BasicBlock&>(op.entry_point().block()));
  259. }
  260. void Compiler::compile_leave_unwind_context(Bytecode::Op::LeaveUnwindContext const&)
  261. {
  262. pop_unwind_context();
  263. }
  264. void Compiler::compile_throw(Bytecode::Op::Throw const&)
  265. {
  266. load_vm_register(GPR0, Bytecode::Register::accumulator());
  267. store_vm_register(Bytecode::Register::exception(), GPR0);
  268. check_exception();
  269. }
  270. static Value cxx_add(VM& vm, Value lhs, Value rhs)
  271. {
  272. return TRY_OR_SET_EXCEPTION(add(vm, lhs, rhs));
  273. }
  274. void Compiler::compile_add(Bytecode::Op::Add const& op)
  275. {
  276. load_vm_register(ARG1, op.lhs());
  277. load_vm_register(ARG2, Bytecode::Register::accumulator());
  278. m_assembler.native_call((void*)cxx_add);
  279. store_vm_register(Bytecode::Register::accumulator(), RET);
  280. check_exception();
  281. }
  282. static Value cxx_sub(VM& vm, Value lhs, Value rhs)
  283. {
  284. return TRY_OR_SET_EXCEPTION(sub(vm, lhs, rhs));
  285. }
  286. void Compiler::compile_sub(Bytecode::Op::Sub const& op)
  287. {
  288. load_vm_register(ARG1, op.lhs());
  289. load_vm_register(ARG2, Bytecode::Register::accumulator());
  290. m_assembler.native_call((void*)cxx_sub);
  291. store_vm_register(Bytecode::Register::accumulator(), RET);
  292. check_exception();
  293. }
  294. static Value cxx_mul(VM& vm, Value lhs, Value rhs)
  295. {
  296. return TRY_OR_SET_EXCEPTION(mul(vm, lhs, rhs));
  297. }
  298. void Compiler::compile_mul(Bytecode::Op::Mul const& op)
  299. {
  300. load_vm_register(ARG1, op.lhs());
  301. load_vm_register(ARG2, Bytecode::Register::accumulator());
  302. m_assembler.native_call((void*)cxx_mul);
  303. store_vm_register(Bytecode::Register::accumulator(), RET);
  304. check_exception();
  305. }
  306. static Value cxx_div(VM& vm, Value lhs, Value rhs)
  307. {
  308. return TRY_OR_SET_EXCEPTION(div(vm, lhs, rhs));
  309. }
  310. void Compiler::compile_div(Bytecode::Op::Div const& op)
  311. {
  312. load_vm_register(ARG1, op.lhs());
  313. load_vm_register(ARG2, Bytecode::Register::accumulator());
  314. m_assembler.native_call((void*)cxx_div);
  315. store_vm_register(Bytecode::Register::accumulator(), RET);
  316. check_exception();
  317. }
  318. void Compiler::compile_return(Bytecode::Op::Return const&)
  319. {
  320. load_vm_register(GPR0, Bytecode::Register::accumulator());
  321. store_vm_register(Bytecode::Register::return_value(), GPR0);
  322. m_assembler.exit();
  323. }
  324. static Value cxx_new_string(VM& vm, DeprecatedString const& string)
  325. {
  326. return PrimitiveString::create(vm, string);
  327. }
  328. void Compiler::compile_new_string(Bytecode::Op::NewString const& op)
  329. {
  330. auto const& string = m_bytecode_executable.string_table->get(op.index());
  331. m_assembler.mov(
  332. Assembler::Operand::Register(ARG1),
  333. Assembler::Operand::Imm64(bit_cast<u64>(&string)));
  334. m_assembler.native_call((void*)cxx_new_string);
  335. store_vm_register(Bytecode::Register::accumulator(), RET);
  336. }
  337. OwnPtr<NativeExecutable> Compiler::compile(Bytecode::Executable& bytecode_executable)
  338. {
  339. if (getenv("LIBJS_NO_JIT"))
  340. return nullptr;
  341. Compiler compiler { bytecode_executable };
  342. compiler.m_assembler.enter();
  343. compiler.m_assembler.mov(
  344. Assembler::Operand::Register(REGISTER_ARRAY_BASE),
  345. Assembler::Operand::Register(ARG1));
  346. compiler.m_assembler.mov(
  347. Assembler::Operand::Register(LOCALS_ARRAY_BASE),
  348. Assembler::Operand::Register(ARG2));
  349. compiler.push_unwind_context(false, {}, {});
  350. for (auto& block : bytecode_executable.basic_blocks) {
  351. block->offset = compiler.m_output.size();
  352. auto it = Bytecode::InstructionStreamIterator(block->instruction_stream());
  353. while (!it.at_end()) {
  354. auto const& op = *it;
  355. switch (op.type()) {
  356. case Bytecode::Instruction::Type::LoadImmediate:
  357. compiler.compile_load_immediate(static_cast<Bytecode::Op::LoadImmediate const&>(op));
  358. break;
  359. case Bytecode::Instruction::Type::Store:
  360. compiler.compile_store(static_cast<Bytecode::Op::Store const&>(op));
  361. break;
  362. case Bytecode::Instruction::Type::Load:
  363. compiler.compile_load(static_cast<Bytecode::Op::Load const&>(op));
  364. break;
  365. case Bytecode::Instruction::Type::GetLocal:
  366. compiler.compile_get_local(static_cast<Bytecode::Op::GetLocal const&>(op));
  367. break;
  368. case Bytecode::Instruction::Type::SetLocal:
  369. compiler.compile_set_local(static_cast<Bytecode::Op::SetLocal const&>(op));
  370. break;
  371. case Bytecode::Instruction::Type::Jump:
  372. compiler.compile_jump(static_cast<Bytecode::Op::Jump const&>(op));
  373. break;
  374. case Bytecode::Instruction::Type::JumpConditional:
  375. compiler.compile_jump_conditional(static_cast<Bytecode::Op::JumpConditional const&>(op));
  376. break;
  377. case Bytecode::Instruction::Type::LessThan:
  378. compiler.compile_less_than(static_cast<Bytecode::Op::LessThan const&>(op));
  379. break;
  380. case Bytecode::Instruction::Type::Increment:
  381. compiler.compile_increment(static_cast<Bytecode::Op::Increment const&>(op));
  382. break;
  383. case Bytecode::Instruction::Type::EnterUnwindContext:
  384. compiler.compile_enter_unwind_context(static_cast<Bytecode::Op::EnterUnwindContext const&>(op));
  385. break;
  386. case Bytecode::Instruction::Type::LeaveUnwindContext:
  387. compiler.compile_leave_unwind_context(static_cast<Bytecode::Op::LeaveUnwindContext const&>(op));
  388. break;
  389. case Bytecode::Instruction::Type::Throw:
  390. compiler.compile_throw(static_cast<Bytecode::Op::Throw const&>(op));
  391. break;
  392. case Bytecode::Instruction::Type::Add:
  393. compiler.compile_add(static_cast<Bytecode::Op::Add const&>(op));
  394. break;
  395. case Bytecode::Instruction::Type::Sub:
  396. compiler.compile_sub(static_cast<Bytecode::Op::Sub const&>(op));
  397. break;
  398. case Bytecode::Instruction::Type::Mul:
  399. compiler.compile_mul(static_cast<Bytecode::Op::Mul const&>(op));
  400. break;
  401. case Bytecode::Instruction::Type::Div:
  402. compiler.compile_div(static_cast<Bytecode::Op::Div const&>(op));
  403. break;
  404. case Bytecode::Instruction::Type::Return:
  405. compiler.compile_return(static_cast<Bytecode::Op::Return const&>(op));
  406. break;
  407. case Bytecode::Instruction::Type::NewString:
  408. compiler.compile_new_string(static_cast<Bytecode::Op::NewString const&>(op));
  409. break;
  410. default:
  411. dbgln("JIT compilation failed: {}", bytecode_executable.name);
  412. dbgln("Unsupported bytecode op: {}", op.to_deprecated_string(bytecode_executable));
  413. return nullptr;
  414. }
  415. ++it;
  416. }
  417. if (!block->is_terminated())
  418. compiler.m_assembler.exit();
  419. }
  420. auto* executable_memory = mmap(nullptr, compiler.m_output.size(), PROT_READ | PROT_WRITE, MAP_ANONYMOUS | MAP_PRIVATE, 0, 0);
  421. if (executable_memory == MAP_FAILED) {
  422. perror("mmap");
  423. return nullptr;
  424. }
  425. for (auto& block : bytecode_executable.basic_blocks) {
  426. // Patch up all the jumps
  427. for (auto& jump : block->jumps_to_here) {
  428. auto offset = block->offset - jump - 4;
  429. compiler.m_output[jump + 0] = (offset >> 0) & 0xff;
  430. compiler.m_output[jump + 1] = (offset >> 8) & 0xff;
  431. compiler.m_output[jump + 2] = (offset >> 16) & 0xff;
  432. compiler.m_output[jump + 3] = (offset >> 24) & 0xff;
  433. }
  434. // Patch up all the absolute references
  435. for (auto& absolute_reference : block->absolute_references_to_here) {
  436. auto offset = bit_cast<u64>(executable_memory) + block->offset;
  437. compiler.m_output[absolute_reference + 0] = (offset >> 0) & 0xff;
  438. compiler.m_output[absolute_reference + 1] = (offset >> 8) & 0xff;
  439. compiler.m_output[absolute_reference + 2] = (offset >> 16) & 0xff;
  440. compiler.m_output[absolute_reference + 3] = (offset >> 24) & 0xff;
  441. compiler.m_output[absolute_reference + 4] = (offset >> 32) & 0xff;
  442. compiler.m_output[absolute_reference + 5] = (offset >> 40) & 0xff;
  443. compiler.m_output[absolute_reference + 6] = (offset >> 48) & 0xff;
  444. compiler.m_output[absolute_reference + 7] = (offset >> 56) & 0xff;
  445. }
  446. }
  447. size_t res = write(STDOUT_FILENO, compiler.m_output.data(), compiler.m_output.size());
  448. if (!res) { }
  449. memcpy(executable_memory, compiler.m_output.data(), compiler.m_output.size());
  450. mprotect(executable_memory, compiler.m_output.size(), PROT_READ | PROT_EXEC);
  451. return make<NativeExecutable>(executable_memory, compiler.m_output.size());
  452. }
  453. }