Store max transaction number
I have a general ledger transaction dataset that is almost 90Mil rows. It takes three hours to do a full replace daily. The dataset has a unique sequential transaction number that grows as transactions are posted.
Most of the time, letting the general ledger just run overnight is sufficient. But several times a month there are instances where I need to be able to refresh the general ledger in the middle of the day and get the results in a timely manner.
Therefore, I want to be able to only pull new transactions since the dataset last ran. I believe there is a way to store the maximum transaction number from the last pull. Which would allow this to become an append of any transaction numbers greater than the stored maximum transaction from the last pull, rather than having to do a full replace.
Best Answers
-
This should get you what you need:
If I solved your problem, please select "yes" above
0 -
I would handle it in an ETL. You have a huge dataset. I deal with 15-20Mil rows. Depending upon your data, you may be able to pull parts of transactions (such as current year, current quarter, etc) and append. I pull past years less frequently than I pull quarterly transactions. I combine the datasets to get a full set for use.
I am currently experimenting with pulling my data through Python because the Domo json connnector is too slow for large data. You probably found what I did - that Postman can pull data much quicker than the oAuth No Code. I'm interested in any tricks you have for that large set btw….do share.** Was this post helpful? Click Agree or Like below. **
** Did this solve your problem? Accept it as a solution! **0 -
@barb_barney_88 I'd also suggest exploring subset processing in MagicETL, which will allow you to request smaller amounts of data on your Workbench dataset (last 30 days, etc.) and merge with the existing output data.
0 -
what you’re referring to is the special lastvalue parameter in a workbench query. Some basic information can be found here as an example
you can ignore the fact that this is OLAP Cube documentation but just search for lastvalue to see the documentation on it. This way you can track the latest id you’ve pulled in and have it pick up from there moving forward**Was this post helpful? Click Agree or Like below**
**Did this solve your problem? Accept it as a solution!**0
Answers
-
This should get you what you need:
If I solved your problem, please select "yes" above
0 -
I would handle it in an ETL. You have a huge dataset. I deal with 15-20Mil rows. Depending upon your data, you may be able to pull parts of transactions (such as current year, current quarter, etc) and append. I pull past years less frequently than I pull quarterly transactions. I combine the datasets to get a full set for use.
I am currently experimenting with pulling my data through Python because the Domo json connnector is too slow for large data. You probably found what I did - that Postman can pull data much quicker than the oAuth No Code. I'm interested in any tricks you have for that large set btw….do share.** Was this post helpful? Click Agree or Like below. **
** Did this solve your problem? Accept it as a solution! **0 -
@barb_barney_88 I'd also suggest exploring subset processing in MagicETL, which will allow you to request smaller amounts of data on your Workbench dataset (last 30 days, etc.) and merge with the existing output data.
0 -
what you’re referring to is the special lastvalue parameter in a workbench query. Some basic information can be found here as an example
you can ignore the fact that this is OLAP Cube documentation but just search for lastvalue to see the documentation on it. This way you can track the latest id you’ve pulled in and have it pick up from there moving forward**Was this post helpful? Click Agree or Like below**
**Did this solve your problem? Accept it as a solution!**0
Categories
- All Categories
- 1.9K Product Ideas
- 1.9K Ideas Exchange
- 1.6K Connect
- 1.3K Connectors
- 303 Workbench
- 6 Cloud Amplifier
- 9 Federated
- 3K Transform
- 104 SQL DataFlows
- 640 Datasets
- 2.2K Magic ETL
- 4K Visualize
- 2.5K Charting
- 769 Beast Mode
- 72 App Studio
- 43 Variables
- 718 Automate
- 185 Apps
- 462 APIs & Domo Developer
- 57 Workflows
- 14 DomoAI
- 40 Predict
- 17 Jupyter Workspaces
- 23 R & Python Tiles
- 402 Distribute
- 116 Domo Everywhere
- 277 Scheduled Reports
- 9 Software Integrations
- 135 Manage
- 132 Governance & Security
- 8 Domo Community Gallery
- 44 Product Releases
- 12 Domo University
- 5.4K Community Forums
- 40 Getting Started
- 30 Community Member Introductions
- 111 Community Announcements
- 4.8K Archive