OLD | NEW |
1 // Copyright 2012 the V8 project authors. All rights reserved. | 1 // Copyright 2012 the V8 project authors. All rights reserved. |
2 // Redistribution and use in source and binary forms, with or without | 2 // Redistribution and use in source and binary forms, with or without |
3 // modification, are permitted provided that the following conditions are | 3 // modification, are permitted provided that the following conditions are |
4 // met: | 4 // met: |
5 // | 5 // |
6 // * Redistributions of source code must retain the above copyright | 6 // * Redistributions of source code must retain the above copyright |
7 // notice, this list of conditions and the following disclaimer. | 7 // notice, this list of conditions and the following disclaimer. |
8 // * Redistributions in binary form must reproduce the above | 8 // * Redistributions in binary form must reproduce the above |
9 // copyright notice, this list of conditions and the following | 9 // copyright notice, this list of conditions and the following |
10 // disclaimer in the documentation and/or other materials provided | 10 // disclaimer in the documentation and/or other materials provided |
(...skipping 93 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
104 code->set_safepoint_table_offset(safepoints_.GetCodeOffset()); | 104 code->set_safepoint_table_offset(safepoints_.GetCodeOffset()); |
105 if (FLAG_weak_embedded_maps_in_optimized_code) { | 105 if (FLAG_weak_embedded_maps_in_optimized_code) { |
106 RegisterDependentCodeForEmbeddedMaps(code); | 106 RegisterDependentCodeForEmbeddedMaps(code); |
107 } | 107 } |
108 PopulateDeoptimizationData(code); | 108 PopulateDeoptimizationData(code); |
109 if (!info()->IsStub()) { | 109 if (!info()->IsStub()) { |
110 Deoptimizer::EnsureRelocSpaceForLazyDeoptimization(code); | 110 Deoptimizer::EnsureRelocSpaceForLazyDeoptimization(code); |
111 } | 111 } |
112 info()->CommitDependentMaps(code); | 112 info()->CommitDependentMaps(code); |
113 | 113 |
114 for (int i = 0 ; i < transition_maps_.length(); i++) { | |
115 transition_maps_.at(i)->AddDependentCode( | |
116 DependentCode::kTransitionGroup, code); | |
117 } | |
118 if (graph()->depends_on_empty_array_proto_elements()) { | 114 if (graph()->depends_on_empty_array_proto_elements()) { |
119 isolate()->initial_object_prototype()->map()->AddDependentCode( | 115 isolate()->initial_object_prototype()->map()->AddDependentCode( |
120 DependentCode::kElementsCantBeAddedGroup, code); | 116 DependentCode::kElementsCantBeAddedGroup, code); |
121 isolate()->initial_array_prototype()->map()->AddDependentCode( | 117 isolate()->initial_array_prototype()->map()->AddDependentCode( |
122 DependentCode::kElementsCantBeAddedGroup, code); | 118 DependentCode::kElementsCantBeAddedGroup, code); |
123 } | 119 } |
124 } | 120 } |
125 | 121 |
126 | 122 |
127 void LCodeGen::Abort(const char* reason) { | 123 void LCodeGen::Abort(const char* reason) { |
(...skipping 4173 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
4301 CpuFeatureScope scope(masm(), SSE2); | 4297 CpuFeatureScope scope(masm(), SSE2); |
4302 XMMRegister value = ToDoubleRegister(instr->value()); | 4298 XMMRegister value = ToDoubleRegister(instr->value()); |
4303 __ movdbl(FieldOperand(object, offset), value); | 4299 __ movdbl(FieldOperand(object, offset), value); |
4304 } else { | 4300 } else { |
4305 __ fstp_d(FieldOperand(object, offset)); | 4301 __ fstp_d(FieldOperand(object, offset)); |
4306 } | 4302 } |
4307 return; | 4303 return; |
4308 } | 4304 } |
4309 | 4305 |
4310 if (!transition.is_null()) { | 4306 if (!transition.is_null()) { |
4311 if (transition->CanBeDeprecated()) { | |
4312 transition_maps_.Add(transition, info()->zone()); | |
4313 } | |
4314 if (!instr->hydrogen()->NeedsWriteBarrierForMap()) { | 4307 if (!instr->hydrogen()->NeedsWriteBarrierForMap()) { |
4315 __ mov(FieldOperand(object, HeapObject::kMapOffset), transition); | 4308 __ mov(FieldOperand(object, HeapObject::kMapOffset), transition); |
4316 } else { | 4309 } else { |
4317 Register temp = ToRegister(instr->temp()); | 4310 Register temp = ToRegister(instr->temp()); |
4318 Register temp_map = ToRegister(instr->temp_map()); | 4311 Register temp_map = ToRegister(instr->temp_map()); |
4319 __ mov(temp_map, transition); | 4312 __ mov(temp_map, transition); |
4320 __ mov(FieldOperand(object, HeapObject::kMapOffset), temp_map); | 4313 __ mov(FieldOperand(object, HeapObject::kMapOffset), temp_map); |
4321 // Update the write barrier for the map field. | 4314 // Update the write barrier for the map field. |
4322 __ RecordWriteField(object, | 4315 __ RecordWriteField(object, |
4323 HeapObject::kMapOffset, | 4316 HeapObject::kMapOffset, |
(...skipping 2184 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
6508 FixedArray::kHeaderSize - kPointerSize)); | 6501 FixedArray::kHeaderSize - kPointerSize)); |
6509 __ bind(&done); | 6502 __ bind(&done); |
6510 } | 6503 } |
6511 | 6504 |
6512 | 6505 |
6513 #undef __ | 6506 #undef __ |
6514 | 6507 |
6515 } } // namespace v8::internal | 6508 } } // namespace v8::internal |
6516 | 6509 |
6517 #endif // V8_TARGET_ARCH_IA32 | 6510 #endif // V8_TARGET_ARCH_IA32 |
OLD | NEW |