-
Notifications
You must be signed in to change notification settings - Fork 0
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
importing a dataset via an URL #1
Comments
Thanks for your interest and reporting the issue with detailed context! Some extra precaution, just in case. The url needs to be a direct downloadable csv file. If you host your table with Google spreadsheet, use Thanks for catching this bug, we already fixed the instructions in the Colab notebook, and tested with import of 13k rows. We'll revise the default value and improve the error messaging in the next release. Let us know if you have further questions |
Thank you for the reply and for the information. However, I wanted to automate corpus creation so I was running the code via python interpreter, not in the notebook. The idea was to automatically download a huggingface dataset, export it I will try to do the import via the notebook. From the end user perspective, the simplest way would be to make the method for importing from Thanks again. |
Thanks for your information. If you have already downloaded the data and loaded in a dataframe, or you need to pre-process the data, it's probably better to load from your local dataframe. We set the limit to avoid passing huge object in a single REST call, but large dataframes can be processed through batching. We'll consider adding it in the next release. But a
|
Ok, thank you, I'll try it |
Hi, using the [v1.5.4], I can't make the importing of data via URL to work using the Service.import_data_url method.
The error received is a generic server error without any specific details, and the logs don't contain any data on the error.
The error occurs for a range of sizes - 100, 1000, 10000 texts.
I'm confident the setup and the invocation of the method are done properly, since I have an operational
instance of a single-project meganno setup, workable with through both notebook and py modules,
and the import from a dataframe works (but not for large datasets), and this data can be viewed and annotated.
Any suggestion would be most helpful. Thanks!
The text was updated successfully, but these errors were encountered: