Currently, after doing all filters, the csv file that gets generated is of ~82 MB and i get this error while downloading:
RuntimeError: Data of size 82.1MB exceeds write limit of 50.0MB
|https://user-images.githubusercontent.com/4397489/92728231-00a5e880-f38e-11ea-9d50-f54853ffab2e.png)]
Using the standard function to download csv file:
{code:python}def get_table_download_link(df,filename):
"""Generates a link allowing the data in a given panda dataframe to be downloaded
in: dataframe
out: href string
"""
csv = df.to_csv(index=False)
b64 = base64.b64encode(
csv.encode()
).decode() # some strings <-> bytes conversions necessary here
return f'Download csv file'{code}
I am sending the filtered csv which is roughly more than 80 MB. How do change the max size in this case of downloading?
Ideally, i should be able to download any size or an configuration to increase the limit as per the use case.
no
I'm getting a similar error

Hello!
From memory, the Base64 data-encoded method has a size limitation depending on your browser.
For larger dataframes, the preferred method for now is dropping the dataframe into Streamlit's static folder and linking back to it. Several approaches are described in this thread.
Fanilo
Thanks, Fanilo! great to know there's this option. Perhaps, I should also validate the df size before displaying it.
Most helpful comment
Hello!
From memory, the Base64 data-encoded method has a size limitation depending on your browser.
For larger dataframes, the preferred method for now is dropping the dataframe into Streamlit's static folder and linking back to it. Several approaches are described in this thread.
Fanilo