I am looking for the best data processing methods. So, this is what I got so far: 1,000,000 nodes of type "A". Each "A" node can be connected to 1-1000 nodes of type "B" and 1-10 nodes of type "C".
I wrote a RESTful service (Java, Jersey) to import data into a neo4j chart. After importing nodes "A" (only nodes with identifiers, no additional data), I notice that neo4j db has grown to ~ 2.4 GB.
Is it good to store additional fields (name, description, ...) in neo4j? Or do I need to configure mongoDB / hadoop to use a combination of keys / values ββto access data?
source share