| OLD | NEW |
| 1 // Copyright 2011 the V8 project authors. All rights reserved. | 1 // Copyright 2011 the V8 project authors. All rights reserved. |
| 2 // Redistribution and use in source and binary forms, with or without | 2 // Redistribution and use in source and binary forms, with or without |
| 3 // modification, are permitted provided that the following conditions are | 3 // modification, are permitted provided that the following conditions are |
| 4 // met: | 4 // met: |
| 5 // | 5 // |
| 6 // * Redistributions of source code must retain the above copyright | 6 // * Redistributions of source code must retain the above copyright |
| 7 // notice, this list of conditions and the following disclaimer. | 7 // notice, this list of conditions and the following disclaimer. |
| 8 // * Redistributions in binary form must reproduce the above | 8 // * Redistributions in binary form must reproduce the above |
| 9 // copyright notice, this list of conditions and the following | 9 // copyright notice, this list of conditions and the following |
| 10 // disclaimer in the documentation and/or other materials provided | 10 // disclaimer in the documentation and/or other materials provided |
| (...skipping 191 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 202 } else { | 202 } else { |
| 203 __ Move(dst, cgen_->ToHandle(constant_source)); | 203 __ Move(dst, cgen_->ToHandle(constant_source)); |
| 204 } | 204 } |
| 205 } else if (destination->IsDoubleRegister()) { | 205 } else if (destination->IsDoubleRegister()) { |
| 206 double v = cgen_->ToDouble(constant_source); | 206 double v = cgen_->ToDouble(constant_source); |
| 207 uint64_t int_val = BitCast<uint64_t, double>(v); | 207 uint64_t int_val = BitCast<uint64_t, double>(v); |
| 208 XMMRegister dst = cgen_->ToDoubleRegister(destination); | 208 XMMRegister dst = cgen_->ToDoubleRegister(destination); |
| 209 if (int_val == 0) { | 209 if (int_val == 0) { |
| 210 __ xorps(dst, dst); | 210 __ xorps(dst, dst); |
| 211 } else { | 211 } else { |
| 212 __ movq(kScratchRegister, int_val, RelocInfo::NONE64); | 212 __ Set(kScratchRegister, int_val); |
| 213 __ movq(dst, kScratchRegister); | 213 __ movq(dst, kScratchRegister); |
| 214 } | 214 } |
| 215 } else { | 215 } else { |
| 216 ASSERT(destination->IsStackSlot()); | 216 ASSERT(destination->IsStackSlot()); |
| 217 Operand dst = cgen_->ToOperand(destination); | 217 Operand dst = cgen_->ToOperand(destination); |
| 218 if (cgen_->IsSmiConstant(constant_source)) { | 218 if (cgen_->IsSmiConstant(constant_source)) { |
| 219 __ Move(dst, cgen_->ToSmi(constant_source)); | 219 __ Move(dst, cgen_->ToSmi(constant_source)); |
| 220 } else if (cgen_->IsInteger32Constant(constant_source)) { | 220 } else if (cgen_->IsInteger32Constant(constant_source)) { |
| 221 // Zero top 32 bits of a 64 bit spill slot that holds a 32 bit untagged | 221 // Zero top 32 bits of a 64 bit spill slot that holds a 32 bit untagged |
| 222 // value. | 222 // value. |
| (...skipping 75 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 298 ASSERT((source->IsDoubleRegister() && destination->IsDoubleStackSlot()) || | 298 ASSERT((source->IsDoubleRegister() && destination->IsDoubleStackSlot()) || |
| 299 (source->IsDoubleStackSlot() && destination->IsDoubleRegister())); | 299 (source->IsDoubleStackSlot() && destination->IsDoubleRegister())); |
| 300 XMMRegister reg = cgen_->ToDoubleRegister(source->IsDoubleRegister() | 300 XMMRegister reg = cgen_->ToDoubleRegister(source->IsDoubleRegister() |
| 301 ? source | 301 ? source |
| 302 : destination); | 302 : destination); |
| 303 LOperand* other = source->IsDoubleRegister() ? destination : source; | 303 LOperand* other = source->IsDoubleRegister() ? destination : source; |
| 304 ASSERT(other->IsDoubleStackSlot()); | 304 ASSERT(other->IsDoubleStackSlot()); |
| 305 Operand other_operand = cgen_->ToOperand(other); | 305 Operand other_operand = cgen_->ToOperand(other); |
| 306 __ movsd(xmm0, other_operand); | 306 __ movsd(xmm0, other_operand); |
| 307 __ movsd(other_operand, reg); | 307 __ movsd(other_operand, reg); |
| 308 __ movsd(reg, xmm0); | 308 __ movaps(reg, xmm0); |
| 309 | 309 |
| 310 } else { | 310 } else { |
| 311 // No other combinations are possible. | 311 // No other combinations are possible. |
| 312 UNREACHABLE(); | 312 UNREACHABLE(); |
| 313 } | 313 } |
| 314 | 314 |
| 315 // The swap of source and destination has executed a move from source to | 315 // The swap of source and destination has executed a move from source to |
| 316 // destination. | 316 // destination. |
| 317 moves_[index].Eliminate(); | 317 moves_[index].Eliminate(); |
| 318 | 318 |
| 319 // Any unperformed (including pending) move with a source of either | 319 // Any unperformed (including pending) move with a source of either |
| 320 // this move's source or destination needs to have their source | 320 // this move's source or destination needs to have their source |
| 321 // changed to reflect the state of affairs after the swap. | 321 // changed to reflect the state of affairs after the swap. |
| 322 for (int i = 0; i < moves_.length(); ++i) { | 322 for (int i = 0; i < moves_.length(); ++i) { |
| 323 LMoveOperands other_move = moves_[i]; | 323 LMoveOperands other_move = moves_[i]; |
| 324 if (other_move.Blocks(source)) { | 324 if (other_move.Blocks(source)) { |
| 325 moves_[i].set_source(destination); | 325 moves_[i].set_source(destination); |
| 326 } else if (other_move.Blocks(destination)) { | 326 } else if (other_move.Blocks(destination)) { |
| 327 moves_[i].set_source(source); | 327 moves_[i].set_source(source); |
| 328 } | 328 } |
| 329 } | 329 } |
| 330 } | 330 } |
| 331 | 331 |
| 332 #undef __ | 332 #undef __ |
| 333 | 333 |
| 334 } } // namespace v8::internal | 334 } } // namespace v8::internal |
| 335 | 335 |
| 336 #endif // V8_TARGET_ARCH_X64 | 336 #endif // V8_TARGET_ARCH_X64 |
| OLD | NEW |