What are the best options to load 100 GB of data in CSV?
- Run using Notebook
- Studio
- Anything else?
What are the best options to load 100 GB of data in CSV?
depends where the data is located. You need to bring it (upload) to server first if you want to load using Studio. I’m using Python for all my uploads, it can be very fast if you configure it properly. And, it should not run on the server, you need a “jump” machine for it.
Question
Then you can create a loading job, check if the files are readably by server, start the job and disconnet the notebook client. Once started, job is going to run in the background (it’s not REST API driven!)
In addition to executing the load jobs via pyTigerGraph notebook and GraphStudio, you can use gsql client either locally or remotely (as long as the target files are accessible on the server).
after loading job where is the data stored? it is not in the local data source.
@Robert_Hardaway thanks for the reply.
Does TG have any sample cronjob?
@porscheme you can create the query and install it, after that you get an exposed REST endpoint to call that query.
After that, you can crontab a bash script to curl that query.
Can you share more about this?
How to use crontab to run query in TG (if I have 3 query, and i want to run the sequencely)?