I have around 1Tb of data, I have stored this data in vertices and edge files to be loaded in Spark GraphFrame to create a graph and run motif(pattern finding) queries on this graph.
For every batch, this 1Tb of vertices and edge file needs to be loaded in GraphFrame to create graph and query it.
Problem I have is that creation of graph is slow.So I want to store this created graph to S3/disk, so that from next time I will directly load this graph and run queries on it which will be fast. Is there is any way to do it, ie create huge graph with help of graphframe store it on disk, and from next time directly load this graph in graphframe and query it.