Export very large dataset (million or row) in CSV as chunks
Answers
-
Hi @user094816
I don't believe Domo's API supports pagination / chunking data. You could utilize the API to read the entire dataset and then use your python script to do the splitting logic yourself.
**Was this post helpful? Click Agree or Like below**
**Did this solve your problem? Accept it as a solution!**1 -
Hi @GrantSmith , Thank you for your feedback. Since domo is slow while exporting data from cards (applying filer of date range to reduce number of rows < 10 ML) for 2 reasons
1) data is very huge
2) Many filters applied on card
I have solved 2nd option and added all filters on ETL itself to reduce the data and generated a new dataset.
Now I am looking for export this huge dataset 104 M rows with 38.5 GB data size. What is the best approach?
I tried to filter and export from UI in DOMO Dataset, but it's not exporting filtered rows from dataset.
0 -
either use dataset views to construct filtered views of your dataset, or use the domo cli https://knowledge.domo.com/Administer/Other_Administrative_Tools/Command_Line_Interface_(CLI)_Tool#section_34, to query-data or export-data and create filtered exports.
Jae Wilson
Check out my 🎥 Domo Training YouTube Channel 👨💻
**Say "Thanks" by clicking the ❤️ in the post that helped you.
**Please mark the post that solves your problem by clicking on "Accept as Solution"0
Categories
- 10.6K All Categories
- 8 Connect
- 918 Connectors
- 250 Workbench
- 474 Transform
- 1.7K Magic ETL
- 69 SQL DataFlows
- 478 Datasets
- 214 Visualize
- 259 Beast Mode
- 2.1K Charting
- 12 Variables
- 19 Automate
- 356 APIs & Domo Developer
- 89 Apps
- 3 Workflows
- 20 Predict
- 5 Jupyter Workspaces
- 15 R & Python Tiles
- 249 Distribute
- 65 Domo Everywhere
- 243 Scheduled Reports
- 21 Manage
- 42 Governance & Security
- 189 Product Ideas
- 1.2K Ideas Exchange
- 11 Community Forums
- 27 Getting Started
- 14 Community Member Introductions
- 55 Community News
- 4.5K Archive