Skip to content

LargeFileUploadTask.upload() throws a 400 error when file size is smaller than max_chunk_size #718

Open
@itsmariodias

Description

@itsmariodias

Describe the bug

While uploading a large file using LargeFileUploadTask to a SharePoint drive location, if the overall file size is less than the max_chunk_size defined (e.g uploading a 5 KB file but the chunk size is 10 MB), the async LargeFileUploadTask.upload() method uploads the file but throws below error:

APIError:
        APIError
        Code: 400
        message: The server returned an unexpected status code and no error class is registered for this code 400

Expected behavior

The upload task should complete without any errors OR the error message should be meaningful e.g. File is too small.

How to reproduce

destination_path = "path/to/your_file.txt" # path on SharePoint drive where I want to upload the file
file_path = "path/to/your_file.txt" # A small file like 10 KB in my local system

async def upload_large_file(graph_client, drive_id):
    try:
        file = open(file_path, 'rb')
        uploadable_properties = DriveItemUploadableProperties(
            additional_data={'@microsoft.graph.conflictBehavior': 'replace'}
        )
        upload_session_request_body = CreateUploadSessionPostRequestBody(item=uploadable_properties)
        print(f"Uploadable Properties: {uploadable_properties.additional_data}")
        # can be used for normal drive uploads
        try:
            upload_session = await graph_client.drives.by_drive_id(
                drive_id
            ).items.by_drive_item_id("root:/my_docs/test_upload.txt:"
                                     ).create_upload_session.post(upload_session_request_body)
            
        except APIError as ex:
            print(f"Error creating upload session: {ex}")

        # to be used for large file uploads
        large_file_upload_session = LargeFileUploadSession(
            upload_url=upload_session.upload_url,
            expiration_date_time=datetime.now() + timedelta(days=1),
            additional_data=upload_session.additional_data,
            is_cancelled=False,
            next_expected_ranges=upload_session.next_expected_ranges
        )

        max_chunk_size = 10 * 1024 * 1024
        task = LargeFileUploadTask(
            upload_session=large_file_upload_session, 
            request_adapter=graph_client.request_adapter, 
            stream=file, 
            parsable_factory=DriveItem, 
            max_chunk_size=max_chunk_size
        )
        total_length = os.path.getsize(file_path)
        
        # Upload the file
        # The callback
        def progress_callback(uploaded_byte_range: tuple[int, int]):
            print(f"Uploaded {uploaded_byte_range[0]} bytes of {total_length} bytes\n\n")

        try:
            upload_result = await task.upload(progress_callback)
            print(f"Upload complete {upload_result}")
        except APIError as ex:
            print(f"Error uploading: {ex.message} - {ex.response_status_code}")
            raise
    except APIError as e:
        print(f"Error: {e}")
        raise

asyncio.run(upload_large_file(graph_client, drive_id))

SDK Version

1.1.7

Latest version known to work for scenario above?

No response

Known Workarounds

Changing the max_chunk_size value to (file_size - 1) seems to prevent the error from occurring.

Debug output

Click to expand log ```
</details>


### Configuration

OS: Windows 10
Architecture: x64
Python version: 3.9.19

### Other information

_No response_

Metadata

Metadata

Assignees

No one assigned

    Labels

    P1Priorityarea:uploadsFocused on functional modules of the productpriority:p1High priority/Major issue but not blocking or Big percentage of customers affected.Bug SLA <=7daystype:bugA broken experience

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions