This section describes in details the options available when using the Neo4j import tool to import data from CSV files.
STRING: arbitrary strings for identifying nodes,
INTEGER: arbitrary integer values for identifying nodes,
ACTUAL: (advanced) actual node ids. Default:
--multiline-fields=truecan severely degrade performance of the importer. Therefore, use it with care, especially with large imports. Default:
\. For example,
\44is equivalent to
\. For example,
\59is equivalent to
""would be interpreted as a literal
". You cannot escape using
20Gfor 20 gigabyte. It can also be specified as a percentage of the available memory, e.g.
truefor SSDs, large raid arrays and network-attached storage.
|Heap size for the import|
You want to set the maximum heap size to a relevant value for the import.
This is done by defining the
If your import data will result in a graph that is larger than 34 billion nodes, 34 billion relationships, or 68 billion properties
you will need to configure the importer to use the high limit record format.
This is achieved by setting the parameter
The location of the import log file can be controlled using the --report-file option.
If you run large imports of CSV files that have low data quality, the import log file can grow very large.
For example, CSV files that contain duplicate node IDs, or that attempt to create relationships between non-existent nodes,
could be classed as having low data quality.
In these cases, you may wish to direct the output to a location that can handle the large log file.
If you are running on a UNIX-like system and you are not interested in the output, you can get rid of it altogether by directing
the report file to
If you need to debug the import it might be useful to collect the stack trace.
This is done by setting the environment variable
NEO4J_DEBUG=true, and rerunning the import.