Skip to content

Writing a dataframe to a mounted s3 bucket in parquet format results in 0-byte files #1901

Open
@Taytay

Description

@Taytay

I've got an S3 bucket mounted. Doing simple touch and echo tests works like a champ. As does writing in other formats.

Then I tried to write a parquet file to the folder and noticed that it often yields 0-byte files with no error

# All of these work:
my_dataframe.to_csv(f"/mounted_folder/my_dataframe.csv")
my_dataframe.to_feather(f"/mounted_folder/my_dataframe.feather")
my_dataframe.to_parquet(f"./tmp/my_dataframe.parquet")

# But this often (not always) yields a zero-byte file:
my_dataframe.to_parquet(f"/mounted_folder/my_dataframe.parquet")

Seems like a bug in the way the flushing is working with the mounted filesystem, but pretty freaky...

Metadata

Metadata

Assignees

No one assigned

    Labels

    P0bugSomething isn't working

    Type

    No type

    Projects

    No projects

    Milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions