-
Dataset not updating.
Can someone help me with this issue? I'm trying to pull the data from databricks (using python script) into domo dataset. First I'm trying to create the dataset for a single date using #domo.ds_create(df2,'datasetname') then I'm trying to update the same dataset daily. the below code worked till yesterday and its only…
-
Exception: Error retrieving a Domo API Access Token
I keep getting this error when i call Domo(domo_client_id, domo_client_secret, api_host='api.domo.com') Exception: Error retrieving a Domo API Access Token: {"status":401,"statusReason":"Unauthorized","path":"/oauth/token","toe":"######"}
-
How can I add to a dataset previously created using the DOMO Python SDK?
I am using the Python SDK to upload csv data to domo. I have been using the dataset.py methods and have been successful in getting my data to my domo domain. Although, everytime I upload a csv, it will create a new dataset. How can I make it so that my data is added to a previously created dataset?
-
How come I cant access Domo through the API?
It was working fine before, I get authorized but now my python script is giving me a 403 status ("Forbidden" error) when trying to retrieve a dataset out of Domo. Is there a rate limit I'm hitting? Quite confused. Any assistance is welcome.
-
Multiple outputs for Python modules
Python boxes in the Domo ETLs allowing multiple outputs.
-
Dataset reference in Python by identifier instead of name
To identify input dataset by identifier instead of name, very useful for working with Python modules, as right now you need to update the source code when changing a name.
-
More info in logs
It would be very useful if, when executing a Domo ETL, a log file was generated to see the details of potential failures in the process. Currently, if there is a failure in any of the Domo's own boxes, you see that the process has been interrupted in a part of the process but we don't see the details. Also, if the process…
-
Domo API to MS-SQL
I tried unsuccessfully to find an example of pulling a Domo dataset directly to MS-SQL via the API. So I created my own. I am not a python coder but my script works just fine. Does anyone have a version of this subroutine that does not require defining the column names (ie header)? #Retrieve dataset from Domo API def…
-
ValueError: The truth value ...
Hi Folks, I'm having an issue with a Python tile in Magic ETL , the code works on my machine (pycharm) but throws an error once in Domo (this happens often, what's the difference?) ValueError: The truth value of an array with more than one element is ambiguous. Use a.any() or a.al The goal is to produce a string value…
-
'None' response when calling Domo SDK to append data into dataset
I am trying to append data into a dataset I created using the API but I get this error: Failed to import data into Domo: 'NoneType' object is not subscriptable. The dataset is created successfully but then I get that error message when trying to append the data into it. After some troubleshooting I found that the 'None'…
-
Adding a Python Function or Google Translation into the Magic ETL
Hi, I am currently pulling Datas of Twitter's feed of specific words from the twitter connector on DOMO. As most of the tweets are in Japanese, i was wondering if there a function/calculated field i could use to translate those japanese tweets on the Magic ETL. Do guide me on the process if it is possible. Thank you!
-
NEW Domo Data Science Resources Guide (help w/ Jupyter, scripting tiles, data science tiles, AutoML)
Hi all, I'm a data scientist at Domo on the Data Science Professional Services Team. We have a NEW Domo Data Science Resources Guide. The guide includes a list of all the help resources Domo has created related to data science products (Jupyter Workspaces, Magic ETL R/Python scripting tiles, Magic ETL data science tiles, &…
-
Is it possible to make a HTTP request using the Python Script tile of Magic ETL?
We would like to use data of a dataset in the body of a POST request and we were wondering if the Python Script tile in Magic ETL would work for that. We have analyzed but it does not seem to have libraries for that. We have this solution to export the dataset and make the request in a Python script in a Virtual Machine…
-
write_dataframe fails to write output of pd.pivot_table having Nulls
Domo python tile fails to write the pivoted dataframe with pd.pivot_table if it has nulls while converting - even if you add 'fill_value=np.nan'. See below: The pivot function works perfectly fine - it does generate the df, however, when you try to write it, it fails - see below if I write the unpivoted df, it works fine…
-
Can we get more packages added to Python Scripting tiles?
I could really use pandas-datareader in an ETL. I know it's available in Jupyter Workspace, but its availability in a Magic ETL would be convenient for me.
-
R and Python for determing dataframe equality in ETL
Greetings Domo Community, I have currently been refactoring Magic ETL tiles and have been attempting to use Python and R scripts to test for equality before finalizing the refactored tile components. Using both approaches with Python and R I have had significant difficulties in completing this task, which I will outline…
-
Daily Dataset Update via Python fails on first run each day
I am using the Domo Dimensions Calendar as an input to a Python script that updates an dataset of the next 365 days as columns |2023-01-27 | 2023-01-28 | 2023-01-29 | 2023-01-30 | ........ | 2024-01-27 So, when the script runs tomorrow the first column will be 2023-01-28, and so on. It works, to a point. The first daily…
-
PyDomo - DS_Update to Append Data?
Hello all, I've written a script that ideally will create a datastream into Domo. At the moment though, I'm struggling to find a way to replace the update method of ds_update() from PyDomo. The default appears to be replacing the data, but I need to append the dataset. Does anyone have any experience with this? Any help…
-
Using Python API to Update Already Created Datasets
Hi Everyone, I am trying to find a way to upload information from a python script into a domo dataset weekly. I am finding that you can't update an existing dataset with this method, only create a new one and update that newly created one. Has anyone found a workaround for this? In the documentation it appears this is…
-
pydomo export_dataset - exporting Domo dataset with millions of rows failing
I have the below that exports a dataset to a string, I'm trying to export a dataset with 257 million rows but I don't see any logs as to why it fails It just fails! When I try the below with a smaller dataset (250k rows) I'm able to export, is there a way to programmatically break out the dataset into chunks? domo =…