Index: third_party/gsutil/boto/tests/unit/glacier/test_writer.py |
diff --git a/third_party/gsutil/boto/tests/unit/glacier/test_writer.py b/third_party/gsutil/boto/tests/unit/glacier/test_writer.py |
new file mode 100644 |
index 0000000000000000000000000000000000000000..41ddfb092ef76feeb6393554ad8851076989e9da |
--- /dev/null |
+++ b/third_party/gsutil/boto/tests/unit/glacier/test_writer.py |
@@ -0,0 +1,185 @@ |
+# Copyright (c) 2012 Amazon.com, Inc. or its affiliates. All Rights Reserved |
+# |
+# Permission is hereby granted, free of charge, to any person obtaining a |
+# copy of this software and associated documentation files (the |
+# "Software"), to deal in the Software without restriction, including |
+# without limitation the rights to use, copy, modify, merge, publish, dis- |
+# tribute, sublicense, and/or sell copies of the Software, and to permit |
+# persons to whom the Software is furnished to do so, subject to the fol- |
+# lowing conditions: |
+# |
+# The above copyright notice and this permission notice shall be included |
+# in all copies or substantial portions of the Software. |
+# |
+# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS |
+# OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABIL- |
+# ITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT |
+# SHALL THE AUTHOR BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, |
+# WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, |
+# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS |
+# IN THE SOFTWARE. |
+# |
+from hashlib import sha256 |
+import itertools |
+from StringIO import StringIO |
+ |
+from tests.unit import unittest |
+from mock import ( |
+ call, |
+ Mock, |
+ patch, |
+ sentinel, |
+) |
+from nose.tools import assert_equal |
+ |
+from boto.glacier.layer1 import Layer1 |
+from boto.glacier.vault import Vault |
+from boto.glacier.writer import Writer, resume_file_upload |
+from boto.glacier.utils import bytes_to_hex, chunk_hashes, tree_hash |
+ |
+ |
+def create_mock_vault(): |
+ vault = Mock(spec=Vault) |
+ vault.layer1 = Mock(spec=Layer1) |
+ vault.layer1.complete_multipart_upload.return_value = dict( |
+ ArchiveId=sentinel.archive_id) |
+ vault.name = sentinel.vault_name |
+ return vault |
+ |
+ |
+def partify(data, part_size): |
+ for i in itertools.count(0): |
+ start = i * part_size |
+ part = data[start:start+part_size] |
+ if part: |
+ yield part |
+ else: |
+ return |
+ |
+ |
+def calculate_mock_vault_calls(data, part_size, chunk_size): |
+ upload_part_calls = [] |
+ data_tree_hashes = [] |
+ for i, data_part in enumerate(partify(data, part_size)): |
+ start = i * part_size |
+ end = start + len(data_part) |
+ data_part_tree_hash_blob = tree_hash( |
+ chunk_hashes(data_part, chunk_size)) |
+ data_part_tree_hash = bytes_to_hex(data_part_tree_hash_blob) |
+ data_part_linear_hash = sha256(data_part).hexdigest() |
+ upload_part_calls.append( |
+ call.layer1.upload_part( |
+ sentinel.vault_name, sentinel.upload_id, |
+ data_part_linear_hash, data_part_tree_hash, |
+ (start, end - 1), data_part)) |
+ data_tree_hashes.append(data_part_tree_hash_blob) |
+ |
+ return upload_part_calls, data_tree_hashes |
+ |
+ |
+def check_mock_vault_calls(vault, upload_part_calls, data_tree_hashes, |
+ data_len): |
+ vault.layer1.upload_part.assert_has_calls( |
+ upload_part_calls, any_order=True) |
+ assert_equal( |
+ len(upload_part_calls), vault.layer1.upload_part.call_count) |
+ |
+ data_tree_hash = bytes_to_hex(tree_hash(data_tree_hashes)) |
+ vault.layer1.complete_multipart_upload.assert_called_once_with( |
+ sentinel.vault_name, sentinel.upload_id, data_tree_hash, data_len) |
+ |
+ |
+class TestWriter(unittest.TestCase): |
+ def setUp(self): |
+ super(TestWriter, self).setUp() |
+ self.vault = create_mock_vault() |
+ self.chunk_size = 2 # power of 2 |
+ self.part_size = 4 # power of 2 |
+ upload_id = sentinel.upload_id |
+ self.writer = Writer( |
+ self.vault, upload_id, self.part_size, self.chunk_size) |
+ |
+ def check_write(self, write_list): |
+ for write_data in write_list: |
+ self.writer.write(write_data) |
+ self.writer.close() |
+ |
+ data = ''.join(write_list) |
+ upload_part_calls, data_tree_hashes = calculate_mock_vault_calls( |
+ data, self.part_size, self.chunk_size) |
+ check_mock_vault_calls( |
+ self.vault, upload_part_calls, data_tree_hashes, len(data)) |
+ |
+ def test_single_byte_write(self): |
+ self.check_write(['1']) |
+ |
+ def test_one_part_write(self): |
+ self.check_write(['1234']) |
+ |
+ def test_split_write_1(self): |
+ self.check_write(['1', '234']) |
+ |
+ def test_split_write_2(self): |
+ self.check_write(['12', '34']) |
+ |
+ def test_split_write_3(self): |
+ self.check_write(['123', '4']) |
+ |
+ def test_one_part_plus_one_write(self): |
+ self.check_write(['12345']) |
+ |
+ def test_returns_archive_id(self): |
+ self.writer.write('1') |
+ self.writer.close() |
+ self.assertEquals(sentinel.archive_id, self.writer.get_archive_id()) |
+ |
+ def test_upload_id(self): |
+ self.assertEquals(sentinel.upload_id, self.writer.upload_id) |
+ |
+ |
+class TestResume(unittest.TestCase): |
+ def setUp(self): |
+ super(TestResume, self).setUp() |
+ self.vault = create_mock_vault() |
+ self.chunk_size = 2 # power of 2 |
+ self.part_size = 4 # power of 2 |
+ |
+ def check_no_resume(self, data, resume_set=set()): |
+ fobj = StringIO(data) |
+ part_hash_map = {} |
+ for part_index in resume_set: |
+ start = self.part_size * part_index |
+ end = start + self.part_size |
+ part_data = data[start:end] |
+ part_hash_map[part_index] = tree_hash( |
+ chunk_hashes(part_data, self.chunk_size)) |
+ |
+ resume_file_upload( |
+ self.vault, sentinel.upload_id, self.part_size, fobj, |
+ part_hash_map, self.chunk_size) |
+ |
+ upload_part_calls, data_tree_hashes = calculate_mock_vault_calls( |
+ data, self.part_size, self.chunk_size) |
+ resume_upload_part_calls = [ |
+ call for part_index, call in enumerate(upload_part_calls) |
+ if part_index not in resume_set] |
+ check_mock_vault_calls( |
+ self.vault, resume_upload_part_calls, data_tree_hashes, len(data)) |
+ |
+ def test_one_part_no_resume(self): |
+ self.check_no_resume('1234') |
+ |
+ def test_two_parts_no_resume(self): |
+ self.check_no_resume('12345678') |
+ |
+ def test_one_part_resume(self): |
+ self.check_no_resume('1234', resume_set=set([0])) |
+ |
+ def test_two_parts_one_resume(self): |
+ self.check_no_resume('12345678', resume_set=set([1])) |
+ |
+ def test_returns_archive_id(self): |
+ archive_id = resume_file_upload( |
+ self.vault, sentinel.upload_id, self.part_size, StringIO('1'), {}, |
+ self.chunk_size) |
+ self.assertEquals(sentinel.archive_id, archive_id) |