Memory leakage is detected via memory_profiler
. Since such big file will be uploaded from 128MB
GCF or f1-micro
GCE, how could I prevent this memory leakage?
✗ python -m memory_profiler tests/test_gcp_storage.py
67108864
Filename: tests/test_gcp_storage.py
Line # Mem usage Increment Line Contents
================================================
48 35.586 MiB 35.586 MiB @profile
49 def test_upload_big_file():
50 35.586 MiB 0.000 MiB from google.cloud import storage
51 35.609 MiB 0.023 MiB client = storage.Client()
52
53 35.609 MiB 0.000 MiB m_bytes = 64
54 35.609 MiB 0.000 MiB filename = int(datetime.utcnow().timestamp())
55 35.609 MiB 0.000 MiB blob_name = f'test/{filename}'
56 35.609 MiB 0.000 MiB bucket_name = 'my_bucket'
57 38.613 MiB 3.004 MiB bucket = client.get_bucket(bucket_name)
58
59 38.613 MiB 0.000 MiB with open(f'/tmp/{filename}', 'wb+') as file_obj:
60 38.613 MiB 0.000 MiB file_obj.seek(m_bytes * 1024 * 1024 - 1)
61 38.613 MiB 0.000 MiB file_obj.write(b'\0')
62 38.613 MiB 0.000 MiB file_obj.seek(0)
63
64 38.613 MiB 0.000 MiB blob = bucket.blob(blob_name)
65 102.707 MiB 64.094 MiB blob.upload_from_file(file_obj)
66
67 102.715 MiB 0.008 MiB blob = bucket.get_blob(blob_name)
68 102.719 MiB 0.004 MiB print(blob.size)
Moreover, if the file is not open with binary mode, the memory leakage will be twice as the file size.
67108864 Filename: tests/test_gcp_storage.py Line # Mem usage Increment Line Contents ================================================ 48 35.410 MiB 35.410 MiB @profile 49 def test_upload_big_file(): 50 35.410 MiB 0.000 MiB from google.cloud import storage 51 35.441 MiB 0.031 MiB client = storage.Client() 52 53 35.441 MiB 0.000 MiB m_bytes = 64 54 35.441 MiB 0.000 MiB filename = int(datetime.utcnow().timestamp()) 55 35.441 MiB 0.000 MiB blob_name = f'test/{filename}' 56 35.441 MiB 0.000 MiB bucket_name = 'my_bucket' 57 38.512 MiB 3.070 MiB bucket = client.get_bucket(bucket_name) 58 59 38.512 MiB 0.000 MiB with open(f'/tmp/{filename}', 'w+') as file_obj: 60 38.512 MiB 0.000 MiB file_obj.seek(m_bytes * 1024 * 1024 - 1) 61 38.512 MiB 0.000 MiB file_obj.write('\0') 62 38.512 MiB 0.000 MiB file_obj.seek(0) 63 64 38.512 MiB 0.000 MiB blob = bucket.blob(blob_name) 65 152.250 MiB 113.738 MiB blob.upload_from_file(file_obj) 66 67 152.699 MiB 0.449 MiB blob = bucket.get_blob(blob_name) 68 152.703 MiB 0.004 MiB print(blob.size)
GIST: https://gist.github.com/northtree/8b560a6b552a975640ec406c9f701731