| Index: runtime/vm/flow_graph_compiler_ia32.cc | 
| diff --git a/runtime/vm/flow_graph_compiler_ia32.cc b/runtime/vm/flow_graph_compiler_ia32.cc | 
| index 8253b34248613cd9d38221917692475a0d7ebc8f..24bdb8168c3251ebad563e594697f5aa0e953b3b 100644 | 
| --- a/runtime/vm/flow_graph_compiler_ia32.cc | 
| +++ b/runtime/vm/flow_graph_compiler_ia32.cc | 
| @@ -880,6 +880,7 @@ void FlowGraphCompiler::GenerateInlinedMathSqrt(Label* done) { | 
| AssemblerMacros::TryAllocate(assembler_, | 
| double_class_, | 
| &call_method, | 
| +                               Assembler::kNearJump, | 
| EAX);  // Result register. | 
| __ movsd(FieldAddress(EAX, Double::value_offset()), XMM0); | 
| __ Drop(1); | 
| @@ -1106,12 +1107,56 @@ void FlowGraphCompiler::LoadDoubleOrSmiToXmm(XmmRegister result, | 
| } | 
|  | 
|  | 
| +void FlowGraphCompiler::SaveLiveRegisters(LocationSummary* locs) { | 
| +  // TODO(vegorov): consider saving only caller save (volatile) registers. | 
| +  const intptr_t xmm_regs_count = locs->live_registers()->xmm_regs_count(); | 
| +  if (xmm_regs_count > 0) { | 
| +    intptr_t stack_offs = (StackSize() + 1) * kWordSize; | 
| +    for (intptr_t reg_idx = 0; reg_idx < kNumberOfXmmRegisters; ++reg_idx) { | 
| +      XmmRegister xmm_reg = static_cast<XmmRegister>(reg_idx); | 
| +      if (locs->live_registers()->ContainsXmmRegister(xmm_reg)) { | 
| +        __ movsd(Address(EBP, -stack_offs), xmm_reg); | 
| +        stack_offs += kDoubleSize; | 
| +      } | 
| +    } | 
| +  } | 
| + | 
| +  for (intptr_t reg_idx = 0; reg_idx < kNumberOfCpuRegisters; ++reg_idx) { | 
| +    Register reg = static_cast<Register>(reg_idx); | 
| +    if (locs->live_registers()->ContainsRegister(reg)) { | 
| +      __ pushl(reg); | 
| +    } | 
| +  } | 
| +} | 
| + | 
| + | 
| +void FlowGraphCompiler::RestoreLiveRegisters(LocationSummary* locs) { | 
| +  for (intptr_t reg_idx = kNumberOfCpuRegisters - 1; reg_idx >= 0; --reg_idx) { | 
| +    Register reg = static_cast<Register>(reg_idx); | 
| +    if (locs->live_registers()->ContainsRegister(reg)) { | 
| +      __ popl(reg); | 
| +    } | 
| +  } | 
| + | 
| +  const intptr_t xmm_regs_count = locs->live_registers()->xmm_regs_count(); | 
| +  if (xmm_regs_count > 0) { | 
| +    intptr_t stack_offs = (StackSize() + 1) * kWordSize; | 
| +    for (intptr_t reg_idx = 0; reg_idx < kNumberOfXmmRegisters; ++reg_idx) { | 
| +      XmmRegister xmm_reg = static_cast<XmmRegister>(reg_idx); | 
| +      if (locs->live_registers()->ContainsXmmRegister(xmm_reg)) { | 
| +        __ movsd(xmm_reg, Address(EBP, -stack_offs)); | 
| +        stack_offs += kDoubleSize; | 
| +      } | 
| +    } | 
| +  } | 
| +} | 
| + | 
| + | 
| #undef __ | 
| #define __ compiler_->assembler()-> | 
|  | 
|  | 
| static Address ToStackSlotAddress(Location loc) { | 
| -  ASSERT(loc.IsStackSlot()); | 
| const intptr_t index = loc.stack_index(); | 
| if (index < 0) { | 
| const intptr_t offset = (1 - index)  * kWordSize; | 
| @@ -1144,6 +1189,23 @@ void ParallelMoveResolver::EmitMove(int index) { | 
| MoveMemoryToMemory(ToStackSlotAddress(destination), | 
| ToStackSlotAddress(source)); | 
| } | 
| +  } else if (source.IsXmmRegister()) { | 
| +    if (destination.IsXmmRegister()) { | 
| +      // Optimization manual recommends using MOVAPS for register | 
| +      // to register moves. | 
| +      __ movaps(destination.xmm_reg(), source.xmm_reg()); | 
| +    } else { | 
| +      ASSERT(destination.IsDoubleStackSlot()); | 
| +      __ movsd(ToStackSlotAddress(destination), source.xmm_reg()); | 
| +    } | 
| +  } else if (source.IsDoubleStackSlot()) { | 
| +    if (destination.IsXmmRegister()) { | 
| +      __ movsd(destination.xmm_reg(), ToStackSlotAddress(source)); | 
| +    } else { | 
| +      ASSERT(destination.IsDoubleStackSlot()); | 
| +      __ movsd(XMM0, ToStackSlotAddress(source)); | 
| +      __ movsd(ToStackSlotAddress(destination), XMM0); | 
| +    } | 
| } else { | 
| ASSERT(source.IsConstant()); | 
| if (destination.IsRegister()) { | 
| @@ -1171,6 +1233,20 @@ void ParallelMoveResolver::EmitSwap(int index) { | 
| Exchange(destination.reg(), ToStackSlotAddress(source)); | 
| } else if (source.IsStackSlot() && destination.IsStackSlot()) { | 
| Exchange(ToStackSlotAddress(destination), ToStackSlotAddress(source)); | 
| +  } else if (source.IsXmmRegister() && destination.IsXmmRegister()) { | 
| +    __ movaps(XMM0, source.xmm_reg()); | 
| +    __ movaps(source.xmm_reg(), destination.xmm_reg()); | 
| +    __ movaps(destination.xmm_reg(), XMM0); | 
| +  } else if (source.IsXmmRegister() || destination.IsXmmRegister()) { | 
| +    ASSERT(destination.IsDoubleStackSlot() || source.IsDoubleStackSlot()); | 
| +    XmmRegister reg = source.IsXmmRegister() ? source.xmm_reg() | 
| +                                             : destination.xmm_reg(); | 
| +    Address slot_address = | 
| +        ToStackSlotAddress(source.IsXmmRegister() ? destination : source); | 
| + | 
| +    __ movsd(XMM0, slot_address); | 
| +    __ movsd(slot_address, reg); | 
| +    __ movaps(reg, XMM0); | 
| } else { | 
| UNREACHABLE(); | 
| } | 
|  |