| Index: runtime/vm/flow_graph_compiler_ia32.cc
|
| diff --git a/runtime/vm/flow_graph_compiler_ia32.cc b/runtime/vm/flow_graph_compiler_ia32.cc
|
| index 8253b34248613cd9d38221917692475a0d7ebc8f..24bdb8168c3251ebad563e594697f5aa0e953b3b 100644
|
| --- a/runtime/vm/flow_graph_compiler_ia32.cc
|
| +++ b/runtime/vm/flow_graph_compiler_ia32.cc
|
| @@ -880,6 +880,7 @@ void FlowGraphCompiler::GenerateInlinedMathSqrt(Label* done) {
|
| AssemblerMacros::TryAllocate(assembler_,
|
| double_class_,
|
| &call_method,
|
| + Assembler::kNearJump,
|
| EAX); // Result register.
|
| __ movsd(FieldAddress(EAX, Double::value_offset()), XMM0);
|
| __ Drop(1);
|
| @@ -1106,12 +1107,56 @@ void FlowGraphCompiler::LoadDoubleOrSmiToXmm(XmmRegister result,
|
| }
|
|
|
|
|
| +void FlowGraphCompiler::SaveLiveRegisters(LocationSummary* locs) {
|
| + // TODO(vegorov): consider saving only caller save (volatile) registers.
|
| + const intptr_t xmm_regs_count = locs->live_registers()->xmm_regs_count();
|
| + if (xmm_regs_count > 0) {
|
| + intptr_t stack_offs = (StackSize() + 1) * kWordSize;
|
| + for (intptr_t reg_idx = 0; reg_idx < kNumberOfXmmRegisters; ++reg_idx) {
|
| + XmmRegister xmm_reg = static_cast<XmmRegister>(reg_idx);
|
| + if (locs->live_registers()->ContainsXmmRegister(xmm_reg)) {
|
| + __ movsd(Address(EBP, -stack_offs), xmm_reg);
|
| + stack_offs += kDoubleSize;
|
| + }
|
| + }
|
| + }
|
| +
|
| + for (intptr_t reg_idx = 0; reg_idx < kNumberOfCpuRegisters; ++reg_idx) {
|
| + Register reg = static_cast<Register>(reg_idx);
|
| + if (locs->live_registers()->ContainsRegister(reg)) {
|
| + __ pushl(reg);
|
| + }
|
| + }
|
| +}
|
| +
|
| +
|
| +void FlowGraphCompiler::RestoreLiveRegisters(LocationSummary* locs) {
|
| + for (intptr_t reg_idx = kNumberOfCpuRegisters - 1; reg_idx >= 0; --reg_idx) {
|
| + Register reg = static_cast<Register>(reg_idx);
|
| + if (locs->live_registers()->ContainsRegister(reg)) {
|
| + __ popl(reg);
|
| + }
|
| + }
|
| +
|
| + const intptr_t xmm_regs_count = locs->live_registers()->xmm_regs_count();
|
| + if (xmm_regs_count > 0) {
|
| + intptr_t stack_offs = (StackSize() + 1) * kWordSize;
|
| + for (intptr_t reg_idx = 0; reg_idx < kNumberOfXmmRegisters; ++reg_idx) {
|
| + XmmRegister xmm_reg = static_cast<XmmRegister>(reg_idx);
|
| + if (locs->live_registers()->ContainsXmmRegister(xmm_reg)) {
|
| + __ movsd(xmm_reg, Address(EBP, -stack_offs));
|
| + stack_offs += kDoubleSize;
|
| + }
|
| + }
|
| + }
|
| +}
|
| +
|
| +
|
| #undef __
|
| #define __ compiler_->assembler()->
|
|
|
|
|
| static Address ToStackSlotAddress(Location loc) {
|
| - ASSERT(loc.IsStackSlot());
|
| const intptr_t index = loc.stack_index();
|
| if (index < 0) {
|
| const intptr_t offset = (1 - index) * kWordSize;
|
| @@ -1144,6 +1189,23 @@ void ParallelMoveResolver::EmitMove(int index) {
|
| MoveMemoryToMemory(ToStackSlotAddress(destination),
|
| ToStackSlotAddress(source));
|
| }
|
| + } else if (source.IsXmmRegister()) {
|
| + if (destination.IsXmmRegister()) {
|
| + // Optimization manual recommends using MOVAPS for register
|
| + // to register moves.
|
| + __ movaps(destination.xmm_reg(), source.xmm_reg());
|
| + } else {
|
| + ASSERT(destination.IsDoubleStackSlot());
|
| + __ movsd(ToStackSlotAddress(destination), source.xmm_reg());
|
| + }
|
| + } else if (source.IsDoubleStackSlot()) {
|
| + if (destination.IsXmmRegister()) {
|
| + __ movsd(destination.xmm_reg(), ToStackSlotAddress(source));
|
| + } else {
|
| + ASSERT(destination.IsDoubleStackSlot());
|
| + __ movsd(XMM0, ToStackSlotAddress(source));
|
| + __ movsd(ToStackSlotAddress(destination), XMM0);
|
| + }
|
| } else {
|
| ASSERT(source.IsConstant());
|
| if (destination.IsRegister()) {
|
| @@ -1171,6 +1233,20 @@ void ParallelMoveResolver::EmitSwap(int index) {
|
| Exchange(destination.reg(), ToStackSlotAddress(source));
|
| } else if (source.IsStackSlot() && destination.IsStackSlot()) {
|
| Exchange(ToStackSlotAddress(destination), ToStackSlotAddress(source));
|
| + } else if (source.IsXmmRegister() && destination.IsXmmRegister()) {
|
| + __ movaps(XMM0, source.xmm_reg());
|
| + __ movaps(source.xmm_reg(), destination.xmm_reg());
|
| + __ movaps(destination.xmm_reg(), XMM0);
|
| + } else if (source.IsXmmRegister() || destination.IsXmmRegister()) {
|
| + ASSERT(destination.IsDoubleStackSlot() || source.IsDoubleStackSlot());
|
| + XmmRegister reg = source.IsXmmRegister() ? source.xmm_reg()
|
| + : destination.xmm_reg();
|
| + Address slot_address =
|
| + ToStackSlotAddress(source.IsXmmRegister() ? destination : source);
|
| +
|
| + __ movsd(XMM0, slot_address);
|
| + __ movsd(slot_address, reg);
|
| + __ movaps(reg, XMM0);
|
| } else {
|
| UNREACHABLE();
|
| }
|
|
|