Index: third_party/gsutil/boto/tests/integration/s3/test_connection.py |
diff --git a/third_party/gsutil/boto/tests/integration/s3/test_connection.py b/third_party/gsutil/boto/tests/integration/s3/test_connection.py |
new file mode 100644 |
index 0000000000000000000000000000000000000000..b6733036bb4fcaf9252597ae370d019b9d31e6c2 |
--- /dev/null |
+++ b/third_party/gsutil/boto/tests/integration/s3/test_connection.py |
@@ -0,0 +1,245 @@ |
+# -*- coding: utf-8 -*- |
+# Copyright (c) 2006-2011 Mitch Garnaat http://garnaat.org/ |
+# |
+# Permission is hereby granted, free of charge, to any person obtaining a |
+# copy of this software and associated documentation files (the |
+# "Software"), to deal in the Software without restriction, including |
+# without limitation the rights to use, copy, modify, merge, publish, dis- |
+# tribute, sublicense, and/or sell copies of the Software, and to permit |
+# persons to whom the Software is furnished to do so, subject to the fol- |
+# lowing conditions: |
+# |
+# The above copyright notice and this permission notice shall be included |
+# in all copies or substantial portions of the Software. |
+# |
+# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS |
+# OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABIL- |
+# ITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT |
+# SHALL THE AUTHOR BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, |
+# WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, |
+# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS |
+# IN THE SOFTWARE. |
+ |
+""" |
+Some unit tests for the S3Connection |
+""" |
+ |
+import unittest |
+import time |
+import os |
+import urllib |
+import urlparse |
+import httplib |
+from boto.s3.connection import S3Connection |
+from boto.s3.bucket import Bucket |
+from boto.exception import S3PermissionsError, S3ResponseError |
+ |
+ |
+class S3ConnectionTest (unittest.TestCase): |
+ s3 = True |
+ |
+ def test_1_basic(self): |
+ print '--- running S3Connection tests ---' |
+ c = S3Connection() |
+ # create a new, empty bucket |
+ bucket_name = 'test-%d' % int(time.time()) |
+ bucket = c.create_bucket(bucket_name) |
+ # now try a get_bucket call and see if it's really there |
+ bucket = c.get_bucket(bucket_name) |
+ # test logging |
+ logging_bucket = c.create_bucket(bucket_name + '-log') |
+ logging_bucket.set_as_logging_target() |
+ bucket.enable_logging(target_bucket=logging_bucket, target_prefix=bucket.name) |
+ bucket.disable_logging() |
+ c.delete_bucket(logging_bucket) |
+ k = bucket.new_key('foobar') |
+ s1 = 'This is a test of file upload and download' |
+ s2 = 'This is a second string to test file upload and download' |
+ k.set_contents_from_string(s1) |
+ fp = open('foobar', 'wb') |
+ # now get the contents from s3 to a local file |
+ k.get_contents_to_file(fp) |
+ fp.close() |
+ fp = open('foobar') |
+ # check to make sure content read from s3 is identical to original |
+ assert s1 == fp.read(), 'corrupted file' |
+ fp.close() |
+ # test generated URLs |
+ url = k.generate_url(3600) |
+ file = urllib.urlopen(url) |
+ assert s1 == file.read(), 'invalid URL %s' % url |
+ url = k.generate_url(3600, force_http=True) |
+ file = urllib.urlopen(url) |
+ assert s1 == file.read(), 'invalid URL %s' % url |
+ url = k.generate_url(3600, force_http=True, headers={'x-amz-x-token' : 'XYZ'}) |
+ file = urllib.urlopen(url) |
+ assert s1 == file.read(), 'invalid URL %s' % url |
+ rh = {'response-content-disposition': 'attachment; filename="foo.txt"'} |
+ url = k.generate_url(60, response_headers=rh) |
+ file = urllib.urlopen(url) |
+ assert s1 == file.read(), 'invalid URL %s' % url |
+ #test whether amperands and to-be-escaped characters work in header filename |
+ rh = {'response-content-disposition': 'attachment; filename="foo&z%20ar&ar&zar&bar.txt"'} |
+ url = k.generate_url(60, response_headers=rh, force_http=True) |
+ file = urllib.urlopen(url) |
+ assert s1 == file.read(), 'invalid URL %s' % url |
+ # overwrite foobar contents with a PUT |
+ url = k.generate_url(3600, 'PUT', force_http=True, policy='private', reduced_redundancy=True) |
+ up = urlparse.urlsplit(url) |
+ con = httplib.HTTPConnection(up.hostname, up.port) |
+ con.request("PUT", up.path + '?' + up.query, body="hello there") |
+ resp = con.getresponse() |
+ assert 200 == resp.status |
+ assert "hello there" == k.get_contents_as_string() |
+ bucket.delete_key(k) |
+ # test a few variations on get_all_keys - first load some data |
+ # for the first one, let's override the content type |
+ phony_mimetype = 'application/x-boto-test' |
+ headers = {'Content-Type': phony_mimetype} |
+ k.name = 'foo/bar' |
+ k.set_contents_from_string(s1, headers) |
+ k.name = 'foo/bas' |
+ k.set_contents_from_filename('foobar') |
+ k.name = 'foo/bat' |
+ k.set_contents_from_string(s1) |
+ k.name = 'fie/bar' |
+ k.set_contents_from_string(s1) |
+ k.name = 'fie/bas' |
+ k.set_contents_from_string(s1) |
+ k.name = 'fie/bat' |
+ k.set_contents_from_string(s1) |
+ # try resetting the contents to another value |
+ md5 = k.md5 |
+ k.set_contents_from_string(s2) |
+ assert k.md5 != md5 |
+ os.unlink('foobar') |
+ all = bucket.get_all_keys() |
+ assert len(all) == 6 |
+ rs = bucket.get_all_keys(prefix='foo') |
+ assert len(rs) == 3 |
+ rs = bucket.get_all_keys(prefix='', delimiter='/') |
+ assert len(rs) == 2 |
+ rs = bucket.get_all_keys(maxkeys=5) |
+ assert len(rs) == 5 |
+ # test the lookup method |
+ k = bucket.lookup('foo/bar') |
+ assert isinstance(k, bucket.key_class) |
+ assert k.content_type == phony_mimetype |
+ k = bucket.lookup('notthere') |
+ assert k == None |
+ # try some metadata stuff |
+ k = bucket.new_key('has_metadata') |
+ mdkey1 = 'meta1' |
+ mdval1 = 'This is the first metadata value' |
+ k.set_metadata(mdkey1, mdval1) |
+ mdkey2 = 'meta2' |
+ mdval2 = 'This is the second metadata value' |
+ k.set_metadata(mdkey2, mdval2) |
+ # try a unicode metadata value |
+ mdval3 = u'föö' |
+ mdkey3 = 'meta3' |
+ k.set_metadata(mdkey3, mdval3) |
+ k.set_contents_from_string(s1) |
+ k = bucket.lookup('has_metadata') |
+ assert k.get_metadata(mdkey1) == mdval1 |
+ assert k.get_metadata(mdkey2) == mdval2 |
+ assert k.get_metadata(mdkey3) == mdval3 |
+ k = bucket.new_key('has_metadata') |
+ k.get_contents_as_string() |
+ assert k.get_metadata(mdkey1) == mdval1 |
+ assert k.get_metadata(mdkey2) == mdval2 |
+ assert k.get_metadata(mdkey3) == mdval3 |
+ bucket.delete_key(k) |
+ # test list and iterator |
+ rs1 = bucket.list() |
+ num_iter = 0 |
+ for r in rs1: |
+ num_iter = num_iter + 1 |
+ rs = bucket.get_all_keys() |
+ num_keys = len(rs) |
+ assert num_iter == num_keys |
+ # try a key with a funny character |
+ k = bucket.new_key('testnewline\n') |
+ k.set_contents_from_string('This is a test') |
+ rs = bucket.get_all_keys() |
+ assert len(rs) == num_keys + 1 |
+ bucket.delete_key(k) |
+ rs = bucket.get_all_keys() |
+ assert len(rs) == num_keys |
+ # try some acl stuff |
+ bucket.set_acl('public-read') |
+ policy = bucket.get_acl() |
+ assert len(policy.acl.grants) == 2 |
+ bucket.set_acl('private') |
+ policy = bucket.get_acl() |
+ assert len(policy.acl.grants) == 1 |
+ k = bucket.lookup('foo/bar') |
+ k.set_acl('public-read') |
+ policy = k.get_acl() |
+ assert len(policy.acl.grants) == 2 |
+ k.set_acl('private') |
+ policy = k.get_acl() |
+ assert len(policy.acl.grants) == 1 |
+ # try the convenience methods for grants |
+ bucket.add_user_grant('FULL_CONTROL', |
+ 'c1e724fbfa0979a4448393c59a8c055011f739b6d102fb37a65f26414653cd67') |
+ try: |
+ bucket.add_email_grant('foobar', 'foo@bar.com') |
+ except S3PermissionsError: |
+ pass |
+ # now try to create an RRS key |
+ k = bucket.new_key('reduced_redundancy') |
+ k.set_contents_from_string('This key has reduced redundancy', |
+ reduced_redundancy=True) |
+ |
+ # now try to inject a response header |
+ data = k.get_contents_as_string(response_headers={'response-content-type' : 'foo/bar'}) |
+ assert k.content_type == 'foo/bar' |
+ |
+ # now delete all keys in bucket |
+ for k in bucket: |
+ if k.name == 'reduced_redundancy': |
+ assert k.storage_class == 'REDUCED_REDUNDANCY' |
+ bucket.delete_key(k) |
+ # now delete bucket |
+ time.sleep(5) |
+ c.delete_bucket(bucket) |
+ print '--- tests completed ---' |
+ |
+ def test_basic_anon(self): |
+ auth_con = S3Connection() |
+ # create a new, empty bucket |
+ bucket_name = 'test-%d' % int(time.time()) |
+ auth_bucket = auth_con.create_bucket(bucket_name) |
+ |
+ # try read the bucket anonymously |
+ anon_con = S3Connection(anon=True) |
+ anon_bucket = Bucket(anon_con, bucket_name) |
+ try: |
+ iter(anon_bucket.list()).next() |
+ self.fail("anon bucket list should fail") |
+ except S3ResponseError: |
+ pass |
+ |
+ # give bucket anon user access and anon read again |
+ auth_bucket.set_acl('public-read') |
+ try: |
+ iter(anon_bucket.list()).next() |
+ self.fail("not expecting contents") |
+ except S3ResponseError, e: |
+ self.fail("We should have public-read access, but received " |
+ "an error: %s" % e) |
+ except StopIteration: |
+ pass |
+ |
+ # cleanup |
+ auth_con.delete_bucket(auth_bucket) |
+ |
+ def test_error_code_populated(self): |
+ c = S3Connection() |
+ try: |
+ c.create_bucket('bad$bucket$name') |
+ except S3ResponseError, e: |
+ self.assertEqual(e.error_code, 'InvalidBucketName') |
+ else: |
+ self.fail("S3ResponseError not raised.") |