// Copyright 2015 the V8 project authors. All rights reserved. // Use of this source code is governed by a BSD-style license that can be // found in the LICENSE file. #ifndef V8_COMPILER_WASM_COMPILER_H_ #define V8_COMPILER_WASM_COMPILER_H_ #include // Clients of this interface shouldn't depend on lots of compiler internals. // Do not include anything from src/compiler here! #include "src/compiler.h" #include "src/optimized-compilation-info.h" #include "src/trap-handler/trap-handler.h" #include "src/wasm/baseline/liftoff-assembler.h" #include "src/wasm/function-body-decoder.h" #include "src/wasm/wasm-module.h" #include "src/wasm/wasm-opcodes.h" #include "src/wasm/wasm-result.h" #include "src/zone/zone.h" namespace v8 { namespace internal { namespace compiler { // Forward declarations for some compiler data structures. class Node; class JSGraph; class Graph; class Operator; class SourcePositionTable; } // namespace compiler namespace wasm { struct DecodeStruct; class SignatureMap; // Expose {Node} and {Graph} opaquely as {wasm::TFNode} and {wasm::TFGraph}. typedef compiler::Node TFNode; typedef compiler::JSGraph TFGraph; class NativeModule; class WasmCode; } // namespace wasm namespace compiler { // The {ModuleEnv} encapsulates the module data that is used by the // {WasmGraphBuilder} during graph building. // ModuleEnvs are shareable across multiple compilations. struct ModuleEnv { // A pointer to the decoded module's static representation. const wasm::WasmModule* module; // True if trap handling should be used in compiled code, rather than // compiling in bounds checks for each memory access. const bool use_trap_handler; ModuleEnv(const wasm::WasmModule* module, bool use_trap_handler) : module(module), use_trap_handler(use_trap_handler) {} }; enum RuntimeExceptionSupport : bool { kRuntimeExceptionSupport = true, kNoRuntimeExceptionSupport = false }; // Information about Wasm compilation that needs to be plumbed through the // different layers of the compiler. class WasmCompilationData { public: explicit WasmCompilationData(RuntimeExceptionSupport); void AddProtectedInstruction(uint32_t instr_offset, uint32_t landing_offset); std::unique_ptr> ReleaseProtectedInstructions() { return std::move(protected_instructions_); } RuntimeExceptionSupport runtime_exception_support() const { return runtime_exception_support_; } private: std::unique_ptr> protected_instructions_; // See WasmGraphBuilder::runtime_exception_support_. const RuntimeExceptionSupport runtime_exception_support_; DISALLOW_COPY_AND_ASSIGN(WasmCompilationData); }; class WasmCompilationUnit final { public: enum class CompilationMode : uint8_t { kLiftoff, kTurbofan }; static CompilationMode GetDefaultCompilationMode(); // If constructing from a background thread, pass in a Counters*, and ensure // that the Counters live at least as long as this compilation unit (which // typically means to hold a std::shared_ptr). // If no such pointer is passed, Isolate::counters() will be called. This is // only allowed to happen on the foreground thread. WasmCompilationUnit(Isolate*, ModuleEnv*, wasm::NativeModule*, wasm::FunctionBody, wasm::WasmName, int index, Handle centry_stub, CompilationMode = GetDefaultCompilationMode(), Counters* = nullptr, RuntimeExceptionSupport = kRuntimeExceptionSupport, bool lower_simd = false); ~WasmCompilationUnit(); void ExecuteCompilation(); wasm::WasmCode* FinishCompilation(wasm::ErrorThrower* thrower); static wasm::WasmCode* CompileWasmFunction( wasm::NativeModule* native_module, wasm::ErrorThrower* thrower, Isolate* isolate, const wasm::ModuleWireBytes& wire_bytes, ModuleEnv* env, const wasm::WasmFunction* function, CompilationMode = GetDefaultCompilationMode()); size_t memory_cost() const { return memory_cost_; } wasm::NativeModule* native_module() const { return native_module_; } private: struct LiftoffData { wasm::LiftoffAssembler asm_; int safepoint_table_offset_; SourcePositionTableBuilder source_position_table_builder_; // The {codegen_zone_} needs to survive until FinishCompilation. It's only // rarely used (e.g. for runtime calls), so it's only allocated when needed. std::unique_ptr codegen_zone_; explicit LiftoffData(Isolate* isolate) : asm_(isolate) {} }; struct TurbofanData { // The graph zone is deallocated at the end of ExecuteCompilation by virtue // of it being zone allocated. JSGraph* jsgraph_ = nullptr; // The compilation_zone_, info_, and job_ fields need to survive past // ExecuteCompilation, onto FinishCompilation (which happens on the main // thread). std::unique_ptr compilation_zone_; std::unique_ptr info_; std::unique_ptr job_; wasm::Result graph_construction_result_; }; // Turbofan. SourcePositionTable* BuildGraphForWasmFunction(double* decode_ms); void ExecuteTurbofanCompilation(); wasm::WasmCode* FinishTurbofanCompilation(wasm::ErrorThrower*); // Liftoff. bool ExecuteLiftoffCompilation(); wasm::WasmCode* FinishLiftoffCompilation(wasm::ErrorThrower*); Isolate* isolate_; ModuleEnv* env_; wasm::FunctionBody func_body_; wasm::WasmName func_name_; Counters* counters_; Handle centry_stub_; int func_index_; bool ok_ = true; size_t memory_cost_ = 0; wasm::NativeModule* native_module_; bool lower_simd_; WasmCompilationData wasm_compilation_data_; CompilationMode mode_; // {liftoff_} is valid if mode_ == kLiftoff, tf_ if mode_ == kTurbofan. union { LiftoffData liftoff_; TurbofanData tf_; }; Counters* counters() { return counters_; } DISALLOW_COPY_AND_ASSIGN(WasmCompilationUnit); }; // Wraps a JS function, producing a code object that can be called from wasm. Handle CompileWasmToJSWrapper(Isolate* isolate, Handle target, wasm::FunctionSig* sig, uint32_t index, wasm::ModuleOrigin origin, bool use_trap_handler); // Wraps a given wasm code object, producing a code object. V8_EXPORT_PRIVATE Handle CompileJSToWasmWrapper( Isolate* isolate, wasm::WasmModule* module, Handle weak_instance, wasm::WasmCode* wasm_code, uint32_t index, bool use_trap_handler); // Compiles a stub that redirects a call to a wasm function to the wasm // interpreter. It's ABI compatible with the compiled wasm function. Handle CompileWasmInterpreterEntry(Isolate* isolate, uint32_t func_index, wasm::FunctionSig* sig); // Helper function to get the offset into a fixed array for a given {index}. // TODO(titzer): access-builder.h is not accessible outside compiler. Move? int FixedArrayOffsetMinusTag(uint32_t index); enum CWasmEntryParameters { kCodeObject, kWasmInstance, kArgumentsBuffer, // marker: kNumParameters }; // Compiles a stub with JS linkage, taking parameters as described by // {CWasmEntryParameters}. It loads the wasm parameters from the argument // buffer and calls the wasm function given as first parameter. Handle CompileCWasmEntry(Isolate* isolate, wasm::FunctionSig* sig); // Values from the instance object are cached between WASM-level function calls. // This struct allows the SSA environment handling this cache to be defined // and manipulated in wasm-compiler.{h,cc} instead of inside the WASM decoder. // (Note that currently, the globals base is immutable, so not cached here.) struct WasmInstanceCacheNodes { Node* mem_start; Node* mem_size; Node* mem_mask; }; // Abstracts details of building TurboFan graph nodes for wasm to separate // the wasm decoder from the internal details of TurboFan. typedef ZoneVector NodeVector; class WasmGraphBuilder { public: enum EnforceBoundsCheck : bool { kNeedsBoundsCheck = true, kCanOmitBoundsCheck = false }; enum UseRetpoline : bool { kRetpoline = true, kNoRetpoline = false }; WasmGraphBuilder(ModuleEnv* env, Zone* zone, JSGraph* graph, Handle centry_stub, Handle anyref_null, wasm::FunctionSig* sig, compiler::SourcePositionTable* spt = nullptr, RuntimeExceptionSupport res = kRuntimeExceptionSupport); Node** Buffer(size_t count) { if (count > cur_bufsize_) { size_t new_size = count + cur_bufsize_ + 5; cur_buffer_ = reinterpret_cast(zone_->New(new_size * sizeof(Node*))); cur_bufsize_ = new_size; } return cur_buffer_; } //----------------------------------------------------------------------- // Operations independent of {control} or {effect}. //----------------------------------------------------------------------- Node* Error(); Node* Start(unsigned params); Node* Param(unsigned index); Node* Loop(Node* entry); Node* Terminate(Node* effect, Node* control); Node* Merge(unsigned count, Node** controls); Node* Phi(wasm::ValueType type, unsigned count, Node** vals, Node* control); Node* CreateOrMergeIntoPhi(wasm::ValueType type, Node* merge, Node* tnode, Node* fnode); Node* CreateOrMergeIntoEffectPhi(Node* merge, Node* tnode, Node* fnode); Node* EffectPhi(unsigned count, Node** effects, Node* control); Node* NumberConstant(int32_t value); Node* Uint32Constant(uint32_t value); Node* Int32Constant(int32_t value); Node* Int64Constant(int64_t value); Node* IntPtrConstant(intptr_t value); Node* Float32Constant(float value); Node* Float64Constant(double value); Node* RefNull() { return anyref_null_node_; } Node* HeapConstant(Handle value); Node* Binop(wasm::WasmOpcode opcode, Node* left, Node* right, wasm::WasmCodePosition position = wasm::kNoCodePosition); Node* Unop(wasm::WasmOpcode opcode, Node* input, wasm::WasmCodePosition position = wasm::kNoCodePosition); Node* GrowMemory(Node* input); Node* Throw(uint32_t tag, const wasm::WasmException* exception, const Vector values); Node* Rethrow(); Node* ConvertExceptionTagToRuntimeId(uint32_t tag); Node* GetExceptionRuntimeId(); Node** GetExceptionValues(const wasm::WasmException* except_decl); bool IsPhiWithMerge(Node* phi, Node* merge); bool ThrowsException(Node* node, Node** if_success, Node** if_exception); void AppendToMerge(Node* merge, Node* from); void AppendToPhi(Node* phi, Node* from); void StackCheck(wasm::WasmCodePosition position, Node** effect = nullptr, Node** control = nullptr); void PatchInStackCheckIfNeeded(); //----------------------------------------------------------------------- // Operations that read and/or write {control} and {effect}. //----------------------------------------------------------------------- Node* BranchNoHint(Node* cond, Node** true_node, Node** false_node); Node* BranchExpectTrue(Node* cond, Node** true_node, Node** false_node); Node* BranchExpectFalse(Node* cond, Node** true_node, Node** false_node); Node* TrapIfTrue(wasm::TrapReason reason, Node* cond, wasm::WasmCodePosition position); Node* TrapIfFalse(wasm::TrapReason reason, Node* cond, wasm::WasmCodePosition position); Node* TrapIfEq32(wasm::TrapReason reason, Node* node, int32_t val, wasm::WasmCodePosition position); Node* ZeroCheck32(wasm::TrapReason reason, Node* node, wasm::WasmCodePosition position); Node* TrapIfEq64(wasm::TrapReason reason, Node* node, int64_t val, wasm::WasmCodePosition position); Node* ZeroCheck64(wasm::TrapReason reason, Node* node, wasm::WasmCodePosition position); Node* Switch(unsigned count, Node* key); Node* IfValue(int32_t value, Node* sw); Node* IfDefault(Node* sw); Node* Return(unsigned count, Node** nodes); template Node* Return(Node* fst, Nodes*... more) { Node* arr[] = {fst, more...}; return Return(arraysize(arr), arr); } Node* ReturnVoid(); Node* Unreachable(wasm::WasmCodePosition position); Node* CallDirect(uint32_t index, Node** args, Node*** rets, wasm::WasmCodePosition position); Node* CallIndirect(uint32_t index, Node** args, Node*** rets, wasm::WasmCodePosition position); void BuildJSToWasmWrapper(Handle weak_instance, wasm::WasmCode* wasm_code); bool BuildWasmToJSWrapper(Handle target, int index); void BuildWasmInterpreterEntry(uint32_t func_index); void BuildCWasmEntry(); Node* ToJS(Node* node, wasm::ValueType type); Node* FromJS(Node* node, Node* js_context, wasm::ValueType type); Node* Invert(Node* node); //----------------------------------------------------------------------- // Operations that concern the linear memory. //----------------------------------------------------------------------- Node* CurrentMemoryPages(); Node* GetGlobal(uint32_t index); Node* SetGlobal(uint32_t index, Node* val); Node* TraceMemoryOperation(bool is_store, MachineRepresentation, Node* index, uint32_t offset, wasm::WasmCodePosition); Node* LoadMem(wasm::ValueType type, MachineType memtype, Node* index, uint32_t offset, uint32_t alignment, wasm::WasmCodePosition position); Node* StoreMem(MachineRepresentation mem_rep, Node* index, uint32_t offset, uint32_t alignment, Node* val, wasm::WasmCodePosition position, wasm::ValueType type); static void PrintDebugName(Node* node); void set_instance_node(Node* instance_node) { this->instance_node_ = instance_node; } Node* Control() { return *control_; } Node* Effect() { return *effect_; } void set_control_ptr(Node** control) { this->control_ = control; } void set_effect_ptr(Node** effect) { this->effect_ = effect; } void GetGlobalBaseAndOffset(MachineType mem_type, uint32_t offset, Node** base_node, Node** offset_node); // Utilities to manipulate sets of instance cache nodes. void InitInstanceCache(WasmInstanceCacheNodes* instance_cache); void PrepareInstanceCacheForLoop(WasmInstanceCacheNodes* instance_cache, Node* control); void NewInstanceCacheMerge(WasmInstanceCacheNodes* to, WasmInstanceCacheNodes* from, Node* merge); void MergeInstanceCacheInto(WasmInstanceCacheNodes* to, WasmInstanceCacheNodes* from, Node* merge); void set_instance_cache(WasmInstanceCacheNodes* instance_cache) { this->instance_cache_ = instance_cache; } wasm::FunctionSig* GetFunctionSignature() { return sig_; } void LowerInt64(); void SimdScalarLoweringForTesting(); void SetSourcePosition(Node* node, wasm::WasmCodePosition position); Node* S128Zero(); Node* S1x4Zero(); Node* S1x8Zero(); Node* S1x16Zero(); Node* SimdOp(wasm::WasmOpcode opcode, Node* const* inputs); Node* SimdLaneOp(wasm::WasmOpcode opcode, uint8_t lane, Node* const* inputs); Node* SimdShiftOp(wasm::WasmOpcode opcode, uint8_t shift, Node* const* inputs); Node* Simd8x16ShuffleOp(const uint8_t shuffle[16], Node* const* inputs); Node* AtomicOp(wasm::WasmOpcode opcode, Node* const* inputs, uint32_t alignment, uint32_t offset, wasm::WasmCodePosition position); bool has_simd() const { return has_simd_; } const wasm::WasmModule* module() { return env_ ? env_->module : nullptr; } bool use_trap_handler() const { return env_ && env_->use_trap_handler; } JSGraph* jsgraph() { return jsgraph_; } Graph* graph(); private: static const int kDefaultBufferSize = 16; Zone* const zone_; JSGraph* const jsgraph_; Node* const centry_stub_node_; Node* const anyref_null_node_; // env_ == nullptr means we're not compiling Wasm functions, such as for // wrappers or interpreter stubs. ModuleEnv* const env_ = nullptr; SetOncePointer instance_node_; struct FunctionTableNodes { Node* table_addr; Node* size; }; Node** control_ = nullptr; Node** effect_ = nullptr; WasmInstanceCacheNodes* instance_cache_ = nullptr; SetOncePointer globals_start_; Node** cur_buffer_; size_t cur_bufsize_; Node* def_buffer_[kDefaultBufferSize]; bool has_simd_ = false; bool needs_stack_check_ = false; const bool untrusted_code_mitigations_ = true; // If the runtime doesn't support exception propagation, // we won't generate stack checks, and trap handling will also // be generated differently. const RuntimeExceptionSupport runtime_exception_support_; wasm::FunctionSig* const sig_; SetOncePointer allocate_heap_number_operator_; compiler::SourcePositionTable* const source_position_table_ = nullptr; Node* String(const char* string); Node* MemBuffer(uint32_t offset); // BoundsCheckMem receives a uint32 {index} node and returns a ptrsize index. Node* BoundsCheckMem(uint8_t access_size, Node* index, uint32_t offset, wasm::WasmCodePosition, EnforceBoundsCheck); Node* Uint32ToUintptr(Node*); const Operator* GetSafeLoadOperator(int offset, wasm::ValueType type); const Operator* GetSafeStoreOperator(int offset, wasm::ValueType type); Node* BuildChangeEndiannessStore(Node* node, MachineRepresentation rep, wasm::ValueType wasmtype = wasm::kWasmStmt); Node* BuildChangeEndiannessLoad(Node* node, MachineType type, wasm::ValueType wasmtype = wasm::kWasmStmt); Node* MaskShiftCount32(Node* node); Node* MaskShiftCount64(Node* node); template Node* BuildCCall(MachineSignature* sig, Node* function, Args... args); Node* BuildWasmCall(wasm::FunctionSig* sig, Node** args, Node*** rets, wasm::WasmCodePosition position, Node* instance_node, UseRetpoline use_retpoline); Node* BuildF32CopySign(Node* left, Node* right); Node* BuildF64CopySign(Node* left, Node* right); Node* BuildIntConvertFloat(Node* input, wasm::WasmCodePosition position, wasm::WasmOpcode); Node* BuildI32Ctz(Node* input); Node* BuildI32Popcnt(Node* input); Node* BuildI64Ctz(Node* input); Node* BuildI64Popcnt(Node* input); Node* BuildBitCountingCall(Node* input, ExternalReference ref, MachineRepresentation input_type); Node* BuildCFuncInstruction(ExternalReference ref, MachineType type, Node* input0, Node* input1 = nullptr); Node* BuildF32Trunc(Node* input); Node* BuildF32Floor(Node* input); Node* BuildF32Ceil(Node* input); Node* BuildF32NearestInt(Node* input); Node* BuildF64Trunc(Node* input); Node* BuildF64Floor(Node* input); Node* BuildF64Ceil(Node* input); Node* BuildF64NearestInt(Node* input); Node* BuildI32Rol(Node* left, Node* right); Node* BuildI64Rol(Node* left, Node* right); Node* BuildF64Acos(Node* input); Node* BuildF64Asin(Node* input); Node* BuildF64Pow(Node* left, Node* right); Node* BuildF64Mod(Node* left, Node* right); Node* BuildIntToFloatConversionInstruction( Node* input, ExternalReference ref, MachineRepresentation parameter_representation, const MachineType result_type); Node* BuildF32SConvertI64(Node* input); Node* BuildF32UConvertI64(Node* input); Node* BuildF64SConvertI64(Node* input); Node* BuildF64UConvertI64(Node* input); Node* BuildCcallConvertFloat(Node* input, wasm::WasmCodePosition position, wasm::WasmOpcode opcode); Node* BuildI32DivS(Node* left, Node* right, wasm::WasmCodePosition position); Node* BuildI32RemS(Node* left, Node* right, wasm::WasmCodePosition position); Node* BuildI32DivU(Node* left, Node* right, wasm::WasmCodePosition position); Node* BuildI32RemU(Node* left, Node* right, wasm::WasmCodePosition position); Node* BuildI64DivS(Node* left, Node* right, wasm::WasmCodePosition position); Node* BuildI64RemS(Node* left, Node* right, wasm::WasmCodePosition position); Node* BuildI64DivU(Node* left, Node* right, wasm::WasmCodePosition position); Node* BuildI64RemU(Node* left, Node* right, wasm::WasmCodePosition position); Node* BuildDiv64Call(Node* left, Node* right, ExternalReference ref, MachineType result_type, int trap_zero, wasm::WasmCodePosition position); Node* BuildJavaScriptToNumber(Node* node, Node* js_context); Node* BuildChangeInt32ToTagged(Node* value); Node* BuildChangeFloat64ToTagged(Node* value); Node* BuildChangeTaggedToFloat64(Node* value); Node* BuildChangeInt32ToSmi(Node* value); Node* BuildChangeSmiToInt32(Node* value); Node* BuildChangeUint32ToSmi(Node* value); Node* BuildChangeSmiToFloat64(Node* value); Node* BuildTestNotSmi(Node* value); Node* BuildSmiShiftBitsConstant(); Node* BuildAllocateHeapNumberWithValue(Node* value, Node* control); Node* BuildLoadHeapNumberValue(Node* value, Node* control); Node* BuildHeapNumberValueIndexConstant(); // Asm.js specific functionality. Node* BuildI32AsmjsSConvertF32(Node* input); Node* BuildI32AsmjsSConvertF64(Node* input); Node* BuildI32AsmjsUConvertF32(Node* input); Node* BuildI32AsmjsUConvertF64(Node* input); Node* BuildI32AsmjsDivS(Node* left, Node* right); Node* BuildI32AsmjsRemS(Node* left, Node* right); Node* BuildI32AsmjsDivU(Node* left, Node* right); Node* BuildI32AsmjsRemU(Node* left, Node* right); Node* BuildAsmjsLoadMem(MachineType type, Node* index); Node* BuildAsmjsStoreMem(MachineType type, Node* index, Node* val); uint32_t GetExceptionEncodedSize(const wasm::WasmException* exception) const; void BuildEncodeException32BitValue(uint32_t* index, Node* value); Node* BuildDecodeException32BitValue(Node* const* values, uint32_t* index); Node** Realloc(Node* const* buffer, size_t old_count, size_t new_count) { Node** buf = Buffer(new_count); if (buf != buffer) memcpy(buf, buffer, old_count * sizeof(Node*)); return buf; } int AddParameterNodes(Node** args, int pos, int param_count, wasm::FunctionSig* sig); void SetNeedsStackCheck() { needs_stack_check_ = true; } //----------------------------------------------------------------------- // Operations involving the CEntryStub, a dependency we want to remove // to get off the GC heap. //----------------------------------------------------------------------- Node* BuildCallToRuntime(Runtime::FunctionId f, Node** parameters, int parameter_count); Node* BuildCallToRuntimeWithContext(Runtime::FunctionId f, Node* js_context, Node** parameters, int parameter_count); Node* BuildCallToRuntimeWithContextFromJS(Runtime::FunctionId f, Node* js_context, Node* const* parameters, int parameter_count); Node* BuildModifyThreadInWasmFlag(bool new_value); Builtins::Name GetBuiltinIdForTrap(wasm::TrapReason reason); }; // The parameter index where the instance parameter should be placed in wasm // call descriptors. This is used by the Int64Lowering::LowerNode method. constexpr int kWasmInstanceParameterIndex = 0; V8_EXPORT_PRIVATE CallDescriptor* GetWasmCallDescriptor( Zone* zone, wasm::FunctionSig* signature, WasmGraphBuilder::UseRetpoline use_retpoline = WasmGraphBuilder::kNoRetpoline); V8_EXPORT_PRIVATE CallDescriptor* GetI32WasmCallDescriptor( Zone* zone, CallDescriptor* call_descriptor); V8_EXPORT_PRIVATE CallDescriptor* GetI32WasmCallDescriptorForSimd( Zone* zone, CallDescriptor* call_descriptor); } // namespace compiler } // namespace internal } // namespace v8 #endif // V8_COMPILER_WASM_COMPILER_H_