We have a use case were we want to run queries that may return more than 500K rows and we want to be able to download those datasets.
I set up a docker version of reDash using the setup guide and I pulled in the majestic data set into the docker postgres instance.
After running a couple test queries I noticed that I can run queries that return 10K rows worth of data and properly download them from the UI. When I goto 100K the UI hangs, but I am able to retrieve the data via the API link: <redash_url>/api/queries/<query_id>/results.csv?api_key=<api_key>, which is great, but I noticed that every time I go revisit the query on the UI, since it never really pulled the result into the UI the page would hang and I can’t really modify anything like setting the reoccurring run schedule (using chrome).
Another issue I am having is since this dataset is 1 million rows, I created another query removing the 100K limit and saved it. I have been trying to run this query, but it fails in the UI due to the SIG 9, memory issue. One way I was trying to get around this is to call the refresh api, but that is not working for me. I am using curl -X post <redash_url>/api/queries/<query_id>/refresh?api_key=<api_key> but I keep on getting 400 errors.
My question is whether I am going in the right direction or if there is anything that I am missing on my side.