| OLD | NEW |
| 1 // Copyright 2012 the V8 project authors. All rights reserved. | 1 // Copyright 2012 the V8 project authors. All rights reserved. |
| 2 // Redistribution and use in source and binary forms, with or without | 2 // Redistribution and use in source and binary forms, with or without |
| 3 // modification, are permitted provided that the following conditions are | 3 // modification, are permitted provided that the following conditions are |
| 4 // met: | 4 // met: |
| 5 // | 5 // |
| 6 // * Redistributions of source code must retain the above copyright | 6 // * Redistributions of source code must retain the above copyright |
| 7 // notice, this list of conditions and the following disclaimer. | 7 // notice, this list of conditions and the following disclaimer. |
| 8 // * Redistributions in binary form must reproduce the above | 8 // * Redistributions in binary form must reproduce the above |
| 9 // copyright notice, this list of conditions and the following | 9 // copyright notice, this list of conditions and the following |
| 10 // disclaimer in the documentation and/or other materials provided | 10 // disclaimer in the documentation and/or other materials provided |
| (...skipping 93 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 104 code->set_safepoint_table_offset(safepoints_.GetCodeOffset()); | 104 code->set_safepoint_table_offset(safepoints_.GetCodeOffset()); |
| 105 if (FLAG_weak_embedded_maps_in_optimized_code) { | 105 if (FLAG_weak_embedded_maps_in_optimized_code) { |
| 106 RegisterDependentCodeForEmbeddedMaps(code); | 106 RegisterDependentCodeForEmbeddedMaps(code); |
| 107 } | 107 } |
| 108 PopulateDeoptimizationData(code); | 108 PopulateDeoptimizationData(code); |
| 109 if (!info()->IsStub()) { | 109 if (!info()->IsStub()) { |
| 110 Deoptimizer::EnsureRelocSpaceForLazyDeoptimization(code); | 110 Deoptimizer::EnsureRelocSpaceForLazyDeoptimization(code); |
| 111 } | 111 } |
| 112 info()->CommitDependentMaps(code); | 112 info()->CommitDependentMaps(code); |
| 113 | 113 |
| 114 for (int i = 0 ; i < transition_maps_.length(); i++) { | |
| 115 transition_maps_.at(i)->AddDependentCode( | |
| 116 DependentCode::kTransitionGroup, code); | |
| 117 } | |
| 118 if (graph()->depends_on_empty_array_proto_elements()) { | 114 if (graph()->depends_on_empty_array_proto_elements()) { |
| 119 isolate()->initial_object_prototype()->map()->AddDependentCode( | 115 isolate()->initial_object_prototype()->map()->AddDependentCode( |
| 120 DependentCode::kElementsCantBeAddedGroup, code); | 116 DependentCode::kElementsCantBeAddedGroup, code); |
| 121 isolate()->initial_array_prototype()->map()->AddDependentCode( | 117 isolate()->initial_array_prototype()->map()->AddDependentCode( |
| 122 DependentCode::kElementsCantBeAddedGroup, code); | 118 DependentCode::kElementsCantBeAddedGroup, code); |
| 123 } | 119 } |
| 124 } | 120 } |
| 125 | 121 |
| 126 | 122 |
| 127 void LCodeGen::Abort(const char* reason) { | 123 void LCodeGen::Abort(const char* reason) { |
| (...skipping 4173 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 4301 CpuFeatureScope scope(masm(), SSE2); | 4297 CpuFeatureScope scope(masm(), SSE2); |
| 4302 XMMRegister value = ToDoubleRegister(instr->value()); | 4298 XMMRegister value = ToDoubleRegister(instr->value()); |
| 4303 __ movdbl(FieldOperand(object, offset), value); | 4299 __ movdbl(FieldOperand(object, offset), value); |
| 4304 } else { | 4300 } else { |
| 4305 __ fstp_d(FieldOperand(object, offset)); | 4301 __ fstp_d(FieldOperand(object, offset)); |
| 4306 } | 4302 } |
| 4307 return; | 4303 return; |
| 4308 } | 4304 } |
| 4309 | 4305 |
| 4310 if (!transition.is_null()) { | 4306 if (!transition.is_null()) { |
| 4311 if (transition->CanBeDeprecated()) { | |
| 4312 transition_maps_.Add(transition, info()->zone()); | |
| 4313 } | |
| 4314 if (!instr->hydrogen()->NeedsWriteBarrierForMap()) { | 4307 if (!instr->hydrogen()->NeedsWriteBarrierForMap()) { |
| 4315 __ mov(FieldOperand(object, HeapObject::kMapOffset), transition); | 4308 __ mov(FieldOperand(object, HeapObject::kMapOffset), transition); |
| 4316 } else { | 4309 } else { |
| 4317 Register temp = ToRegister(instr->temp()); | 4310 Register temp = ToRegister(instr->temp()); |
| 4318 Register temp_map = ToRegister(instr->temp_map()); | 4311 Register temp_map = ToRegister(instr->temp_map()); |
| 4319 __ mov(temp_map, transition); | 4312 __ mov(temp_map, transition); |
| 4320 __ mov(FieldOperand(object, HeapObject::kMapOffset), temp_map); | 4313 __ mov(FieldOperand(object, HeapObject::kMapOffset), temp_map); |
| 4321 // Update the write barrier for the map field. | 4314 // Update the write barrier for the map field. |
| 4322 __ RecordWriteField(object, | 4315 __ RecordWriteField(object, |
| 4323 HeapObject::kMapOffset, | 4316 HeapObject::kMapOffset, |
| (...skipping 2184 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 6508 FixedArray::kHeaderSize - kPointerSize)); | 6501 FixedArray::kHeaderSize - kPointerSize)); |
| 6509 __ bind(&done); | 6502 __ bind(&done); |
| 6510 } | 6503 } |
| 6511 | 6504 |
| 6512 | 6505 |
| 6513 #undef __ | 6506 #undef __ |
| 6514 | 6507 |
| 6515 } } // namespace v8::internal | 6508 } } // namespace v8::internal |
| 6516 | 6509 |
| 6517 #endif // V8_TARGET_ARCH_IA32 | 6510 #endif // V8_TARGET_ARCH_IA32 |
| OLD | NEW |