Index: runtime/vm/intermediate_language.cc |
=================================================================== |
--- runtime/vm/intermediate_language.cc (revision 8343) |
+++ runtime/vm/intermediate_language.cc (working copy) |
@@ -626,17 +626,17 @@ |
#define __ compiler->assembler()-> |
-void GraphEntryInstr::PrepareEntry(FlowGraphCompiler* compiler) { |
+void GraphEntryInstr::PrepareEntry(FlowGraphCompilerShared* compiler) { |
// Nothing to do. |
} |
-void JoinEntryInstr::PrepareEntry(FlowGraphCompiler* compiler) { |
+void JoinEntryInstr::PrepareEntry(FlowGraphCompilerShared* compiler) { |
__ Bind(compiler->GetBlockLabel(this)); |
} |
-void TargetEntryInstr::PrepareEntry(FlowGraphCompiler* compiler) { |
+void TargetEntryInstr::PrepareEntry(FlowGraphCompilerShared* compiler) { |
__ Bind(compiler->GetBlockLabel(this)); |
if (HasTryIndex()) { |
compiler->AddExceptionHandler(try_index(), |
@@ -653,7 +653,7 @@ |
-void ThrowInstr::EmitNativeCode(FlowGraphCompiler* compiler) { |
+void ThrowInstr::EmitNativeCode(FlowGraphCompilerShared* compiler) { |
ASSERT(exception()->IsUse()); |
compiler->GenerateCallRuntime(cid(), |
token_index(), |
@@ -670,7 +670,7 @@ |
} |
-void ReThrowInstr::EmitNativeCode(FlowGraphCompiler* compiler) { |
+void ReThrowInstr::EmitNativeCode(FlowGraphCompilerShared* compiler) { |
ASSERT(exception()->IsUse()); |
ASSERT(stack_trace()->IsUse()); |
compiler->GenerateCallRuntime(cid(), |
@@ -690,7 +690,7 @@ |
} |
-void BranchInstr::EmitNativeCode(FlowGraphCompiler* compiler) { |
+void BranchInstr::EmitNativeCode(FlowGraphCompilerShared* compiler) { |
Register value = locs()->in(0).reg(); |
__ CompareObject(value, Bool::ZoneHandle(Bool::True())); |
if (compiler->IsNextBlock(false_successor())) { |
@@ -711,7 +711,7 @@ |
} |
-void CurrentContextComp::EmitNativeCode(FlowGraphCompiler* compiler) { |
+void CurrentContextComp::EmitNativeCode(FlowGraphCompilerShared* compiler) { |
__ MoveRegister(locs()->out().reg(), CTX); |
} |
@@ -725,7 +725,7 @@ |
} |
-void StoreContextComp::EmitNativeCode(FlowGraphCompiler* compiler) { |
+void StoreContextComp::EmitNativeCode(FlowGraphCompilerShared* compiler) { |
// Nothing to do. Context register were loaded by register allocator. |
ASSERT(locs()->in(0).reg() == CTX); |
} |
@@ -736,7 +736,7 @@ |
} |
-void StrictCompareComp::EmitNativeCode(FlowGraphCompiler* compiler) { |
+void StrictCompareComp::EmitNativeCode(FlowGraphCompilerShared* compiler) { |
const Bool& bool_true = Bool::ZoneHandle(Bool::True()); |
const Bool& bool_false = Bool::ZoneHandle(Bool::False()); |
@@ -760,7 +760,7 @@ |
} |
-void ClosureCallComp::EmitNativeCode(FlowGraphCompiler* compiler) { |
+void ClosureCallComp::EmitNativeCode(FlowGraphCompilerShared* compiler) { |
ASSERT(VerifyCallComputation(this)); |
// The arguments to the stub include the closure. The arguments |
// descriptor describes the closure's arguments (and so does not include |
@@ -785,7 +785,7 @@ |
} |
-void InstanceCallComp::EmitNativeCode(FlowGraphCompiler* compiler) { |
+void InstanceCallComp::EmitNativeCode(FlowGraphCompilerShared* compiler) { |
ASSERT(VerifyCallComputation(this)); |
compiler->AddCurrentDescriptor(PcDescriptors::kDeopt, |
cid(), |
@@ -806,7 +806,7 @@ |
} |
-void StaticCallComp::EmitNativeCode(FlowGraphCompiler* compiler) { |
+void StaticCallComp::EmitNativeCode(FlowGraphCompilerShared* compiler) { |
ASSERT(VerifyCallComputation(this)); |
compiler->GenerateStaticCall(cid(), |
token_index(), |
@@ -822,12 +822,12 @@ |
} |
-void UseVal::EmitNativeCode(FlowGraphCompiler* compiler) { |
+void UseVal::EmitNativeCode(FlowGraphCompilerShared* compiler) { |
UNIMPLEMENTED(); |
} |
-void AssertAssignableComp::EmitNativeCode(FlowGraphCompiler* compiler) { |
+void AssertAssignableComp::EmitNativeCode(FlowGraphCompilerShared* compiler) { |
compiler->GenerateAssertAssignable(cid(), |
token_index(), |
try_index(), |
@@ -847,7 +847,7 @@ |
} |
-void StoreInstanceFieldComp::EmitNativeCode(FlowGraphCompiler* compiler) { |
+void StoreInstanceFieldComp::EmitNativeCode(FlowGraphCompilerShared* compiler) { |
ASSERT(VerifyValues(instance(), value())); |
Register instance = locs()->in(0).reg(); |
Register value = locs()->in(1).reg(); |
@@ -870,7 +870,7 @@ |
} |
-void StoreStaticFieldComp::EmitNativeCode(FlowGraphCompiler* compiler) { |
+void StoreStaticFieldComp::EmitNativeCode(FlowGraphCompilerShared* compiler) { |
Register value = locs()->in(0).reg(); |
Register temp = locs()->temp(0).reg(); |
ASSERT(locs()->out().reg() == value); |
@@ -885,7 +885,7 @@ |
} |
-void BooleanNegateComp::EmitNativeCode(FlowGraphCompiler* compiler) { |
+void BooleanNegateComp::EmitNativeCode(FlowGraphCompilerShared* compiler) { |
Register value = locs()->in(0).reg(); |
Register result = locs()->out().reg(); |
@@ -905,7 +905,7 @@ |
} |
-void ChainContextComp::EmitNativeCode(FlowGraphCompiler* compiler) { |
+void ChainContextComp::EmitNativeCode(FlowGraphCompilerShared* compiler) { |
Register context_value = locs()->in(0).reg(); |
// Chain the new context in context_value to its parent in CTX. |
@@ -922,7 +922,7 @@ |
} |
-void StoreVMFieldComp::EmitNativeCode(FlowGraphCompiler* compiler) { |
+void StoreVMFieldComp::EmitNativeCode(FlowGraphCompilerShared* compiler) { |
Register value_reg = locs()->in(0).reg(); |
Register dest_reg = locs()->in(1).reg(); |
ASSERT(value_reg == locs()->out().reg()); |
@@ -937,7 +937,7 @@ |
} |
-void AllocateObjectComp::EmitNativeCode(FlowGraphCompiler* compiler) { |
+void AllocateObjectComp::EmitNativeCode(FlowGraphCompilerShared* compiler) { |
const Class& cls = Class::ZoneHandle(constructor().owner()); |
const Code& stub = Code::Handle(StubCode::GetAllocationStubForClass(cls)); |
const ExternalLabel label(cls.ToCString(), stub.EntryPoint()); |
@@ -954,7 +954,7 @@ |
} |
-void CreateClosureComp::EmitNativeCode(FlowGraphCompiler* compiler) { |
+void CreateClosureComp::EmitNativeCode(FlowGraphCompilerShared* compiler) { |
const Function& closure_function = function(); |
const Code& stub = Code::Handle( |
StubCode::GetAllocationStubForClosure(closure_function)); |