OLD | NEW |
1 // Copyright 2012 the V8 project authors. All rights reserved. | 1 // Copyright 2012 the V8 project authors. All rights reserved. |
2 // Redistribution and use in source and binary forms, with or without | 2 // Redistribution and use in source and binary forms, with or without |
3 // modification, are permitted provided that the following conditions are | 3 // modification, are permitted provided that the following conditions are |
4 // met: | 4 // met: |
5 // | 5 // |
6 // * Redistributions of source code must retain the above copyright | 6 // * Redistributions of source code must retain the above copyright |
7 // notice, this list of conditions and the following disclaimer. | 7 // notice, this list of conditions and the following disclaimer. |
8 // * Redistributions in binary form must reproduce the above | 8 // * Redistributions in binary form must reproduce the above |
9 // copyright notice, this list of conditions and the following | 9 // copyright notice, this list of conditions and the following |
10 // disclaimer in the documentation and/or other materials provided | 10 // disclaimer in the documentation and/or other materials provided |
(...skipping 422 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
433 | 433 |
434 // Dispatch on the indirect string shape: slice or cons. | 434 // Dispatch on the indirect string shape: slice or cons. |
435 Label cons_string; | 435 Label cons_string; |
436 __ tst(result, Operand(kSlicedNotConsMask)); | 436 __ tst(result, Operand(kSlicedNotConsMask)); |
437 __ b(eq, &cons_string); | 437 __ b(eq, &cons_string); |
438 | 438 |
439 // Handle slices. | 439 // Handle slices. |
440 Label indirect_string_loaded; | 440 Label indirect_string_loaded; |
441 __ ldr(result, FieldMemOperand(string, SlicedString::kOffsetOffset)); | 441 __ ldr(result, FieldMemOperand(string, SlicedString::kOffsetOffset)); |
442 __ ldr(string, FieldMemOperand(string, SlicedString::kParentOffset)); | 442 __ ldr(string, FieldMemOperand(string, SlicedString::kParentOffset)); |
443 __ add(index, index, Operand(result, ASR, kSmiTagSize)); | 443 __ add(index, index, Operand::SmiUntag(result)); |
444 __ jmp(&indirect_string_loaded); | 444 __ jmp(&indirect_string_loaded); |
445 | 445 |
446 // Handle cons strings. | 446 // Handle cons strings. |
447 // Check whether the right hand side is the empty string (i.e. if | 447 // Check whether the right hand side is the empty string (i.e. if |
448 // this is really a flat string in a cons string). If that is not | 448 // this is really a flat string in a cons string). If that is not |
449 // the case we would rather go to the runtime system now to flatten | 449 // the case we would rather go to the runtime system now to flatten |
450 // the string. | 450 // the string. |
451 __ bind(&cons_string); | 451 __ bind(&cons_string); |
452 __ ldr(result, FieldMemOperand(string, ConsString::kSecondOffset)); | 452 __ ldr(result, FieldMemOperand(string, ConsString::kSecondOffset)); |
453 __ CompareRoot(result, Heap::kempty_stringRootIndex); | 453 __ CompareRoot(result, Heap::kempty_stringRootIndex); |
(...skipping 49 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
503 __ bind(&done); | 503 __ bind(&done); |
504 } | 504 } |
505 | 505 |
506 | 506 |
507 void SeqStringSetCharGenerator::Generate(MacroAssembler* masm, | 507 void SeqStringSetCharGenerator::Generate(MacroAssembler* masm, |
508 String::Encoding encoding, | 508 String::Encoding encoding, |
509 Register string, | 509 Register string, |
510 Register index, | 510 Register index, |
511 Register value) { | 511 Register value) { |
512 if (FLAG_debug_code) { | 512 if (FLAG_debug_code) { |
513 __ tst(index, Operand(kSmiTagMask)); | 513 __ SmiTst(index); |
514 __ Check(eq, "Non-smi index"); | 514 __ Check(eq, "Non-smi index"); |
515 __ tst(value, Operand(kSmiTagMask)); | 515 __ SmiTst(value); |
516 __ Check(eq, "Non-smi value"); | 516 __ Check(eq, "Non-smi value"); |
517 | 517 |
518 __ ldr(ip, FieldMemOperand(string, String::kLengthOffset)); | 518 __ ldr(ip, FieldMemOperand(string, String::kLengthOffset)); |
519 __ cmp(index, ip); | 519 __ cmp(index, ip); |
520 __ Check(lt, "Index is too large"); | 520 __ Check(lt, "Index is too large"); |
521 | 521 |
522 __ cmp(index, Operand(Smi::FromInt(0))); | 522 __ cmp(index, Operand(Smi::FromInt(0))); |
523 __ Check(ge, "Index is negative"); | 523 __ Check(ge, "Index is negative"); |
524 | 524 |
525 __ ldr(ip, FieldMemOperand(string, HeapObject::kMapOffset)); | 525 __ ldr(ip, FieldMemOperand(string, HeapObject::kMapOffset)); |
526 __ ldrb(ip, FieldMemOperand(ip, Map::kInstanceTypeOffset)); | 526 __ ldrb(ip, FieldMemOperand(ip, Map::kInstanceTypeOffset)); |
527 | 527 |
528 __ and_(ip, ip, Operand(kStringRepresentationMask | kStringEncodingMask)); | 528 __ and_(ip, ip, Operand(kStringRepresentationMask | kStringEncodingMask)); |
529 static const uint32_t one_byte_seq_type = kSeqStringTag | kOneByteStringTag; | 529 static const uint32_t one_byte_seq_type = kSeqStringTag | kOneByteStringTag; |
530 static const uint32_t two_byte_seq_type = kSeqStringTag | kTwoByteStringTag; | 530 static const uint32_t two_byte_seq_type = kSeqStringTag | kTwoByteStringTag; |
531 __ cmp(ip, Operand(encoding == String::ONE_BYTE_ENCODING | 531 __ cmp(ip, Operand(encoding == String::ONE_BYTE_ENCODING |
532 ? one_byte_seq_type : two_byte_seq_type)); | 532 ? one_byte_seq_type : two_byte_seq_type)); |
533 __ Check(eq, "Unexpected string type"); | 533 __ Check(eq, "Unexpected string type"); |
534 } | 534 } |
535 | 535 |
536 __ add(ip, | 536 __ add(ip, |
537 string, | 537 string, |
538 Operand(SeqString::kHeaderSize - kHeapObjectTag)); | 538 Operand(SeqString::kHeaderSize - kHeapObjectTag)); |
539 __ SmiUntag(value, value); | 539 __ SmiUntag(value, value); |
540 STATIC_ASSERT(kSmiTagSize == 1 && kSmiTag == 0); | 540 STATIC_ASSERT(kSmiTagSize == 1 && kSmiTag == 0); |
541 if (encoding == String::ONE_BYTE_ENCODING) { | 541 if (encoding == String::ONE_BYTE_ENCODING) { |
542 // Smis are tagged by left shift by 1, thus LSR by 1 to smi-untag inline. | 542 // Smis are tagged by left shift by 1, thus LSR by 1 to smi-untag inline. |
543 __ strb(value, MemOperand(ip, index, LSR, 1)); | 543 __ strb(value, MemOperand(ip, index, LSR, kSmiTagSize)); |
544 } else { | 544 } else { |
545 // No need to untag a smi for two-byte addressing. | 545 // No need to untag a smi for two-byte addressing. |
546 __ strh(value, MemOperand(ip, index)); | 546 __ strh(value, MemOperand(ip, index)); // LSL(1 - kSmiTagSize). |
547 } | 547 } |
548 } | 548 } |
549 | 549 |
550 | 550 |
551 static MemOperand ExpConstant(int index, Register base) { | 551 static MemOperand ExpConstant(int index, Register base) { |
552 return MemOperand(base, index * kDoubleSize); | 552 return MemOperand(base, index * kDoubleSize); |
553 } | 553 } |
554 | 554 |
555 | 555 |
556 void MathExpGenerator::EmitMathExp(MacroAssembler* masm, | 556 void MathExpGenerator::EmitMathExp(MacroAssembler* masm, |
(...skipping 124 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
681 patcher.masm()->add(r0, pc, Operand(-8)); | 681 patcher.masm()->add(r0, pc, Operand(-8)); |
682 patcher.masm()->ldr(pc, MemOperand(pc, -4)); | 682 patcher.masm()->ldr(pc, MemOperand(pc, -4)); |
683 patcher.masm()->dd(reinterpret_cast<uint32_t>(stub->instruction_start())); | 683 patcher.masm()->dd(reinterpret_cast<uint32_t>(stub->instruction_start())); |
684 } | 684 } |
685 } | 685 } |
686 | 686 |
687 | 687 |
688 } } // namespace v8::internal | 688 } } // namespace v8::internal |
689 | 689 |
690 #endif // V8_TARGET_ARCH_ARM | 690 #endif // V8_TARGET_ARCH_ARM |
OLD | NEW |