Compiler.cpp 16 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423
  1. /*
  2. * Copyright (c) 2023, Andreas Kling <kling@serenityos.org>
  3. *
  4. * SPDX-License-Identifier: BSD-2-Clause
  5. */
  6. #include <AK/OwnPtr.h>
  7. #include <LibJS/Bytecode/Instruction.h>
  8. #include <LibJS/Bytecode/Interpreter.h>
  9. #include <LibJS/JIT/Compiler.h>
  10. #include <LibJS/Runtime/VM.h>
  11. #include <LibJS/Runtime/ValueInlines.h>
  12. #include <sys/mman.h>
  13. #include <unistd.h>
  14. #define TRY_OR_SET_EXCEPTION(expression) \
  15. ({ \
  16. /* Ignore -Wshadow to allow nesting the macro. */ \
  17. AK_IGNORE_DIAGNOSTIC("-Wshadow", \
  18. auto&& _temporary_result = (expression)); \
  19. static_assert(!::AK::Detail::IsLvalueReference<decltype(_temporary_result.release_value())>, \
  20. "Do not return a reference from a fallible expression"); \
  21. if (_temporary_result.is_error()) [[unlikely]] { \
  22. vm.bytecode_interpreter().reg(Bytecode::Register::exception()) = _temporary_result.release_error().value().value(); \
  23. return {}; \
  24. } \
  25. _temporary_result.release_value(); \
  26. })
  27. namespace JS::JIT {
  28. void Compiler::store_vm_register(Bytecode::Register dst, Assembler::Reg src)
  29. {
  30. m_assembler.mov(
  31. Assembler::Operand::Mem64BaseAndOffset(REGISTER_ARRAY_BASE, dst.index() * sizeof(Value)),
  32. Assembler::Operand::Register(src));
  33. }
  34. void Compiler::load_vm_register(Assembler::Reg dst, Bytecode::Register src)
  35. {
  36. m_assembler.mov(
  37. Assembler::Operand::Register(dst),
  38. Assembler::Operand::Mem64BaseAndOffset(REGISTER_ARRAY_BASE, src.index() * sizeof(Value)));
  39. }
  40. void Compiler::store_vm_local(size_t dst, Assembler::Reg src)
  41. {
  42. m_assembler.mov(
  43. Assembler::Operand::Mem64BaseAndOffset(LOCALS_ARRAY_BASE, dst * sizeof(Value)),
  44. Assembler::Operand::Register(src));
  45. }
  46. void Compiler::load_vm_local(Assembler::Reg dst, size_t src)
  47. {
  48. m_assembler.mov(
  49. Assembler::Operand::Register(dst),
  50. Assembler::Operand::Mem64BaseAndOffset(LOCALS_ARRAY_BASE, src * sizeof(Value)));
  51. }
  52. void Compiler::compile_load_immediate(Bytecode::Op::LoadImmediate const& op)
  53. {
  54. m_assembler.mov(
  55. Assembler::Operand::Register(GPR0),
  56. Assembler::Operand::Imm64(op.value().encoded()));
  57. store_vm_register(Bytecode::Register::accumulator(), GPR0);
  58. }
  59. void Compiler::compile_load(Bytecode::Op::Load const& op)
  60. {
  61. load_vm_register(GPR0, op.src());
  62. store_vm_register(Bytecode::Register::accumulator(), GPR0);
  63. }
  64. void Compiler::compile_store(Bytecode::Op::Store const& op)
  65. {
  66. load_vm_register(GPR0, Bytecode::Register::accumulator());
  67. store_vm_register(op.dst(), GPR0);
  68. }
  69. void Compiler::compile_get_local(Bytecode::Op::GetLocal const& op)
  70. {
  71. load_vm_local(GPR0, op.index());
  72. store_vm_register(Bytecode::Register::accumulator(), GPR0);
  73. }
  74. void Compiler::compile_set_local(Bytecode::Op::SetLocal const& op)
  75. {
  76. load_vm_register(GPR0, Bytecode::Register::accumulator());
  77. store_vm_local(op.index(), GPR0);
  78. }
  79. void Compiler::compile_jump(Bytecode::Op::Jump const& op)
  80. {
  81. m_assembler.jump(const_cast<Bytecode::BasicBlock&>(op.true_target()->block()));
  82. }
  83. static bool cxx_to_boolean(VM&, Value value)
  84. {
  85. return value.to_boolean();
  86. }
  87. void Compiler::compile_to_boolean(Assembler::Reg dst, Assembler::Reg src)
  88. {
  89. // dst = src;
  90. m_assembler.mov(
  91. Assembler::Operand::Register(dst),
  92. Assembler::Operand::Register(src));
  93. // dst >>= 48;
  94. m_assembler.shift_right(
  95. Assembler::Operand::Register(dst),
  96. Assembler::Operand::Imm8(48));
  97. // if (dst != BOOLEAN_TAG) goto slow_case;
  98. auto slow_case = m_assembler.make_label();
  99. m_assembler.jump_if_not_equal(
  100. Assembler::Operand::Register(dst),
  101. Assembler::Operand::Imm32(BOOLEAN_TAG),
  102. slow_case);
  103. // Fast path for JS::Value booleans.
  104. // dst = src;
  105. m_assembler.mov(
  106. Assembler::Operand::Register(dst),
  107. Assembler::Operand::Register(src));
  108. // dst &= 1;
  109. m_assembler.bitwise_and(
  110. Assembler::Operand::Register(dst),
  111. Assembler::Operand::Imm32(1));
  112. // goto end;
  113. auto end = m_assembler.jump();
  114. // slow_case: // call C++ helper
  115. slow_case.link(m_assembler);
  116. m_assembler.mov(
  117. Assembler::Operand::Register(ARG1),
  118. Assembler::Operand::Register(src));
  119. m_assembler.native_call((void*)cxx_to_boolean);
  120. m_assembler.mov(
  121. Assembler::Operand::Register(dst),
  122. Assembler::Operand::Register(RET));
  123. // end:
  124. end.link(m_assembler);
  125. }
  126. void Compiler::compile_jump_conditional(Bytecode::Op::JumpConditional const& op)
  127. {
  128. load_vm_register(GPR1, Bytecode::Register::accumulator());
  129. compile_to_boolean(GPR0, GPR1);
  130. m_assembler.jump_conditional(GPR0,
  131. const_cast<Bytecode::BasicBlock&>(op.true_target()->block()),
  132. const_cast<Bytecode::BasicBlock&>(op.false_target()->block()));
  133. }
  134. [[maybe_unused]] static Value cxx_less_than(VM& vm, Value lhs, Value rhs)
  135. {
  136. return TRY_OR_SET_EXCEPTION(less_than(vm, lhs, rhs));
  137. }
  138. void Compiler::compile_less_than(Bytecode::Op::LessThan const& op)
  139. {
  140. load_vm_register(ARG1, op.lhs());
  141. load_vm_register(ARG2, Bytecode::Register::accumulator());
  142. m_assembler.native_call((void*)cxx_less_than);
  143. store_vm_register(Bytecode::Register::accumulator(), RET);
  144. check_exception();
  145. }
  146. [[maybe_unused]] static Value cxx_increment(VM& vm, Value value)
  147. {
  148. auto old_value = TRY_OR_SET_EXCEPTION(value.to_numeric(vm));
  149. if (old_value.is_number())
  150. return Value(old_value.as_double() + 1);
  151. return BigInt::create(vm, old_value.as_bigint().big_integer().plus(Crypto::SignedBigInteger { 1 }));
  152. }
  153. void Compiler::compile_increment(Bytecode::Op::Increment const&)
  154. {
  155. load_vm_register(ARG1, Bytecode::Register::accumulator());
  156. m_assembler.native_call((void*)cxx_increment);
  157. store_vm_register(Bytecode::Register::accumulator(), RET);
  158. check_exception();
  159. }
  160. void Compiler::check_exception()
  161. {
  162. // if (exception.is_empty()) goto no_exception;
  163. load_vm_register(GPR0, Bytecode::Register::exception());
  164. m_assembler.mov(Assembler::Operand::Register(GPR1), Assembler::Operand::Imm64(Value().encoded()));
  165. auto no_exception = m_assembler.make_label();
  166. m_assembler.jump_if_equal(Assembler::Operand::Register(GPR0), Assembler::Operand::Register(GPR1), no_exception);
  167. // We have an exception!
  168. // if (!unwind_context.valid) return;
  169. auto handle_exception = m_assembler.make_label();
  170. m_assembler.mov(
  171. Assembler::Operand::Register(GPR0),
  172. Assembler::Operand::Mem64BaseAndOffset(UNWIND_CONTEXT_BASE, 0));
  173. m_assembler.jump_if_not_equal(
  174. Assembler::Operand::Register(GPR0),
  175. Assembler::Operand::Imm32(0),
  176. handle_exception);
  177. m_assembler.exit();
  178. // handle_exception:
  179. handle_exception.link(m_assembler);
  180. // if (unwind_context.handler) {
  181. // accumulator = exception;
  182. // exception = Value();
  183. // goto handler;
  184. // }
  185. auto no_handler = m_assembler.make_label();
  186. m_assembler.mov(
  187. Assembler::Operand::Register(GPR0),
  188. Assembler::Operand::Mem64BaseAndOffset(UNWIND_CONTEXT_BASE, 8));
  189. m_assembler.jump_if_equal(
  190. Assembler::Operand::Register(GPR0),
  191. Assembler::Operand::Imm32(0),
  192. no_handler);
  193. load_vm_register(GPR1, Bytecode::Register::exception());
  194. store_vm_register(Bytecode::Register::accumulator(), GPR1);
  195. m_assembler.mov(
  196. Assembler::Operand::Register(GPR1),
  197. Assembler::Operand::Imm64(Value().encoded()));
  198. store_vm_register(Bytecode::Register::exception(), GPR1);
  199. m_assembler.jump(Assembler::Operand::Register(GPR0));
  200. // no_handler:
  201. no_handler.link(m_assembler);
  202. // if (unwind_context.finalizer) goto finalizer;
  203. auto no_finalizer = m_assembler.make_label();
  204. m_assembler.mov(
  205. Assembler::Operand::Register(GPR0),
  206. Assembler::Operand::Mem64BaseAndOffset(UNWIND_CONTEXT_BASE, 16));
  207. m_assembler.jump_if_equal(
  208. Assembler::Operand::Register(GPR0),
  209. Assembler::Operand::Imm32(0),
  210. no_finalizer);
  211. m_assembler.jump(Assembler::Operand::Register(GPR0));
  212. // no_finalizer:
  213. // NOTE: No catch and no finally!? Crash.
  214. no_finalizer.link(m_assembler);
  215. m_assembler.verify_not_reached();
  216. // no_exception:
  217. no_exception.link(m_assembler);
  218. }
  219. void Compiler::push_unwind_context(bool valid, Optional<Bytecode::Label> const& handler, Optional<Bytecode::Label> const& finalizer)
  220. {
  221. // Put this on the stack, and then point UNWIND_CONTEXT_BASE at it.
  222. // struct {
  223. // u64 valid;
  224. // u64 handler;
  225. // u64 finalizer;
  226. // };
  227. // push finalizer (patched later)
  228. m_assembler.mov(
  229. Assembler::Operand::Register(GPR0),
  230. Assembler::Operand::Imm64(0));
  231. if (finalizer.has_value())
  232. const_cast<Bytecode::BasicBlock&>(finalizer.value().block()).absolute_references_to_here.append(m_assembler.m_output.size() - 8);
  233. m_assembler.push(Assembler::Operand::Register(GPR0));
  234. // push handler (patched later)
  235. m_assembler.mov(
  236. Assembler::Operand::Register(GPR0),
  237. Assembler::Operand::Imm64(0));
  238. if (handler.has_value())
  239. const_cast<Bytecode::BasicBlock&>(handler.value().block()).absolute_references_to_here.append(m_assembler.m_output.size() - 8);
  240. m_assembler.push(Assembler::Operand::Register(GPR0));
  241. // push valid
  242. m_assembler.push(Assembler::Operand::Imm32(valid));
  243. // UNWIND_CONTEXT_BASE = STACK_POINTER
  244. m_assembler.mov(
  245. Assembler::Operand::Register(UNWIND_CONTEXT_BASE),
  246. Assembler::Operand::Register(STACK_POINTER));
  247. // align stack pointer
  248. m_assembler.sub(Assembler::Operand::Register(STACK_POINTER), Assembler::Operand::Imm8(8));
  249. }
  250. void Compiler::pop_unwind_context()
  251. {
  252. m_assembler.add(Assembler::Operand::Register(STACK_POINTER), Assembler::Operand::Imm8(32));
  253. m_assembler.add(Assembler::Operand::Register(UNWIND_CONTEXT_BASE), Assembler::Operand::Imm8(32));
  254. }
  255. void Compiler::compile_enter_unwind_context(Bytecode::Op::EnterUnwindContext const& op)
  256. {
  257. push_unwind_context(true, op.handler_target(), op.finalizer_target());
  258. m_assembler.jump(const_cast<Bytecode::BasicBlock&>(op.entry_point().block()));
  259. }
  260. void Compiler::compile_leave_unwind_context(Bytecode::Op::LeaveUnwindContext const&)
  261. {
  262. pop_unwind_context();
  263. }
  264. OwnPtr<NativeExecutable> Compiler::compile(Bytecode::Executable const& bytecode_executable)
  265. {
  266. if (getenv("LIBJS_NO_JIT"))
  267. return nullptr;
  268. Compiler compiler;
  269. compiler.m_assembler.enter();
  270. compiler.m_assembler.mov(
  271. Assembler::Operand::Register(REGISTER_ARRAY_BASE),
  272. Assembler::Operand::Register(ARG1));
  273. compiler.m_assembler.mov(
  274. Assembler::Operand::Register(LOCALS_ARRAY_BASE),
  275. Assembler::Operand::Register(ARG2));
  276. compiler.push_unwind_context(false, {}, {});
  277. for (auto& block : bytecode_executable.basic_blocks) {
  278. block->offset = compiler.m_output.size();
  279. auto it = Bytecode::InstructionStreamIterator(block->instruction_stream());
  280. while (!it.at_end()) {
  281. auto const& op = *it;
  282. switch (op.type()) {
  283. case Bytecode::Instruction::Type::LoadImmediate:
  284. compiler.compile_load_immediate(static_cast<Bytecode::Op::LoadImmediate const&>(op));
  285. break;
  286. case Bytecode::Instruction::Type::Store:
  287. compiler.compile_store(static_cast<Bytecode::Op::Store const&>(op));
  288. break;
  289. case Bytecode::Instruction::Type::Load:
  290. compiler.compile_load(static_cast<Bytecode::Op::Load const&>(op));
  291. break;
  292. case Bytecode::Instruction::Type::GetLocal:
  293. compiler.compile_get_local(static_cast<Bytecode::Op::GetLocal const&>(op));
  294. break;
  295. case Bytecode::Instruction::Type::SetLocal:
  296. compiler.compile_set_local(static_cast<Bytecode::Op::SetLocal const&>(op));
  297. break;
  298. case Bytecode::Instruction::Type::Jump:
  299. compiler.compile_jump(static_cast<Bytecode::Op::Jump const&>(op));
  300. break;
  301. case Bytecode::Instruction::Type::JumpConditional:
  302. compiler.compile_jump_conditional(static_cast<Bytecode::Op::JumpConditional const&>(op));
  303. break;
  304. case Bytecode::Instruction::Type::LessThan:
  305. compiler.compile_less_than(static_cast<Bytecode::Op::LessThan const&>(op));
  306. break;
  307. case Bytecode::Instruction::Type::Increment:
  308. compiler.compile_increment(static_cast<Bytecode::Op::Increment const&>(op));
  309. break;
  310. case Bytecode::Instruction::Type::EnterUnwindContext:
  311. compiler.compile_enter_unwind_context(static_cast<Bytecode::Op::EnterUnwindContext const&>(op));
  312. break;
  313. case Bytecode::Instruction::Type::LeaveUnwindContext:
  314. compiler.compile_leave_unwind_context(static_cast<Bytecode::Op::LeaveUnwindContext const&>(op));
  315. break;
  316. default:
  317. dbgln("JIT compilation failed: {}", bytecode_executable.name);
  318. dbgln("Unsupported bytecode op: {}", op.to_deprecated_string(bytecode_executable));
  319. return nullptr;
  320. }
  321. ++it;
  322. }
  323. if (!block->is_terminated())
  324. compiler.m_assembler.exit();
  325. }
  326. auto* executable_memory = mmap(nullptr, compiler.m_output.size(), PROT_READ | PROT_WRITE, MAP_ANONYMOUS | MAP_PRIVATE, 0, 0);
  327. if (executable_memory == MAP_FAILED) {
  328. perror("mmap");
  329. return nullptr;
  330. }
  331. for (auto& block : bytecode_executable.basic_blocks) {
  332. // Patch up all the jumps
  333. for (auto& jump : block->jumps_to_here) {
  334. auto offset = block->offset - jump - 4;
  335. compiler.m_output[jump + 0] = (offset >> 0) & 0xff;
  336. compiler.m_output[jump + 1] = (offset >> 8) & 0xff;
  337. compiler.m_output[jump + 2] = (offset >> 16) & 0xff;
  338. compiler.m_output[jump + 3] = (offset >> 24) & 0xff;
  339. }
  340. // Patch up all the absolute references
  341. for (auto& absolute_reference : block->absolute_references_to_here) {
  342. auto offset = bit_cast<u64>(executable_memory) + block->offset;
  343. compiler.m_output[absolute_reference + 0] = (offset >> 0) & 0xff;
  344. compiler.m_output[absolute_reference + 1] = (offset >> 8) & 0xff;
  345. compiler.m_output[absolute_reference + 2] = (offset >> 16) & 0xff;
  346. compiler.m_output[absolute_reference + 3] = (offset >> 24) & 0xff;
  347. compiler.m_output[absolute_reference + 4] = (offset >> 32) & 0xff;
  348. compiler.m_output[absolute_reference + 5] = (offset >> 40) & 0xff;
  349. compiler.m_output[absolute_reference + 6] = (offset >> 48) & 0xff;
  350. compiler.m_output[absolute_reference + 7] = (offset >> 56) & 0xff;
  351. }
  352. }
  353. write(STDOUT_FILENO, compiler.m_output.data(), compiler.m_output.size());
  354. memcpy(executable_memory, compiler.m_output.data(), compiler.m_output.size());
  355. mprotect(executable_memory, compiler.m_output.size(), PROT_READ | PROT_EXEC);
  356. return make<NativeExecutable>(executable_memory, compiler.m_output.size());
  357. }
  358. }