neo4j-etl Command Line Tool
This is the command-line tool you use to retrieve and map the metadata from your relational database and drive the export from the relational and import into Neo4j database.
With the graphical user interface you can preview the resulting graph data model and eventually adapt it by changing labels, property names, relationship-types and property types.
It supports all relational databases with a JDBC driver, like MySQL, PostgreSQL, Oracle and Microsoft SQL.
You can get the latest version of the import tool from GitHub.
Once downloaded and uncompressed the operating system specific zip / tar.gz, you also need download the proper JDBC Driver and add it to the lib folder.
You can follow the proper link in the below table in order to download the proper driver jar
| Vendor | JDBC Driver URL |
|---|---|
http://www.oracle.com/technetwork/database/features/jdbc/default-2280470.html |
|
https://www.microsoft.com/en-us/download/details.aspx?id=55539 |
| For very large databases make sure to have enough disk-space for the CSV export and the Neo4j datastore and enough RAM and CPUs to finish the import quickly. |
'generate-metadata-mapping' command
NAME
neo4j-etl generate-metadata-mapping - Create RDBMS to Neo4j metadata
mapping Json.
SYNOPSIS
neo4j-etl generate-metadata-mapping
[ {--columns | --cols} <Columns>... ]
[ --config-file <Configuration File> ]
[ {-d | --database} <RDBMS database> ] [ --debug ]
[ --delimiter <delimiter> ] [ {--driver | --jars} <--driver <PATH_TO_YOUR_JAR1> --driver <PATH_TO_YOUR_JAR2>>... ]
[ {--exclusion-mode | --exc} <Tables: exclude|include|none(default)> ]
[ {--exclusion-mode-column-type | --exctype} <exclude|include|none(default)> ]
[ {--exclusion-mode-columns | --excc} <exclude|include|none(default)> ]
[ {--exclusion-mode-tables | --exct} <exclude|include|none(default)> ]
[ --options-file <option file> ] [ --output-mapping-file <file|stdout> ]
[ {-p | --port} <RDBMS port> ] [ --quote <quote> ]
[ {--rdbms:fetch-size | --fs} <fetch-size> ]
[ {--rdbms:password | --password} <RDBMS password> ]
[ {--rdbms:schema | -s | --schema} <schema> ]
[ {--rdbms:url | --url} <RDBMS url> ]
[ {--rdbms:user | -u | --user} <RDBMS user> ]
[ {--relationship-name | --rel-name} <table(default)|column> ]
[ --schemas <Schemas>... ] [ {--tables | --tabs} <Tables>... ]
[ --tiny-int <byte(default)|boolean> ] [ --types <Types>... ] [--] [ <table1 table2 ...>... ]
OPTIONS
--columns <Columns>, --cols <Columns>
Lists all columns to include/exclude by name or pattern
Use '-r' <PATTERN> to filter by regex, ex. '-r .*\.orders\..*_id' or
'northwind\.orders\..*_id' ,
'-g' <PATTERN> for grep syntax, ex. '-g .*\.orders\..*_id' or
'northwind\.orders\..*_id' ,
or '-l' <LIST> to list all columns names ex. '-l
northwind.customers.id,northwind.purchase.id,northwind.orders.id'
--config-file <Configuration File>
Specify the path to a file containing the configuration for the
selected command
-d <RDBMS database>, --database <RDBMS database>
RDBMS database.
This option is required if any of the following options are
specified: host
--debug
Print detailed diagnostic output.
--delimiter <delimiter>
Delimiter to separate fields in CSV.
--driver <--driver <PATH_TO_YOUR_JAR1> --driver <PATH_TO_YOUR_JAR2>>, --jars <--driver <PATH_TO_YOUR_JAR1> --driver <PATH_TO_YOUR_JAR2>>
List of additional drivers as a list
--exclusion-mode <Tables: exclude|include|none(default)>, --exc <Tables: exclude|include|none(default)>
Specifies how to handle table exclusion. Options are mutually
exclusive.
exclude: Excludes specified tables from the process. All other
tables will be included.
include: Includes specified tables only. All other tables will be
excluded.
none: All tables are included in the process.
--exclusion-mode-column-type <exclude|include|none(default)>, --exctype
<exclude|include|none(default)>
Specifies how to handle column type exclusion. Options are mutually
exclusive.
exclude: Excludes specified columns types from the process. All
other columns types will be included.
include: Includes specified columns types only. All other columns
types will be excluded.
none: All columns types are included in the process.
--exclusion-mode-columns <exclude|include|none(default)>, --excc
<exclude|include|none(default)>
Specifies how to handle column exclusion. Options are mutually
exclusive.
exclude: Excludes specified columns from the process. All other
columns will be included.
include: Includes specified columns only. All other columns will be
excluded.
none: All columns are included in the process.
--exclusion-mode-tables <exclude|include|none(default)>, --exct
<exclude|include|none(default)>
Specifies how to handle table exclusion. Options are mutually
exclusive.
exclude: Excludes specified tables from the process. All other
tables will be included.
include: Includes specified tables only. All other tables will be
excluded.
none: All tables are included in the process.
--options-file <option file>
Path to file containing Neo4j import tool options.
--output-mapping-file <file|stdout>
Path to the output metadata mapping file.
-p <RDBMS port>, --port <RDBMS port>
Port number to use for connection to RDBMS.
--quote <quote>
Character to treat as quotation character for values in CSV data.
--rdbms:fetch-size <fetch-size>, --fs <fetch-size>
RDBMS Fetch size
--rdbms:password <RDBMS password>, --password <RDBMS password>
Password for login to RDBMS.
This option is required if any of the following options are
specified: --rdbms:url, --url
--rdbms:schema <schema>, -s <schema>, --schema <schema>
RDBMS schema.
--rdbms:url <RDBMS url>, --url <RDBMS url>
Url to use for connection to RDBMS.
--rdbms:user <RDBMS user>, -u <RDBMS user>, --user <RDBMS user>
User for login to RDBMS.
This option is required if any of the following options are
specified: --rdbms:url, --url
--relationship-name <table(default)|column>, --rel-name
<table(default)|column>
Specifies whether to get the name for relationships from table names
or column names.
--schemas <Schemas>
Lists all schemas to include by name or pattern.
Use '-r' <PATTERN> to filter by regex, ex. '-r .*\.north.*',
'-g' <PATTERN> for grep syntax, ex. '-g .*\.north.*' ,
or '-l' <LIST> to list all schemas names ex. '-l northwind,exc'
--tables <Tables>, --tabs <Tables>
Lists all tables to include/exclude by name or pattern.
Use '-r' <PATTERN> to filter by regex, ex. '-r .*\.purchase.*' or
'northwind.purchase.*' ,
'-g' <PATTERN> for grep syntax, ex. '-g .*\.purchase.*' or
'northwind.purchase.*' ,
or '-l' <LIST> to list all tables names ex. '-l
customers,purchase,orders'
--tiny-int <byte(default)|boolean>
Specifies whether to convert TinyInt to byte or boolean
--types <Types>
Lists all column types to include/exclude by name separated by
commas. Valid values:
unknown,
binary,
bit,
character,
id,
integer,
real,
reference,
temporal,
url,
xml,
large_object,
object;
--
This option can be used to separate command-line options from the
list of arguments (useful when arguments might be mistaken for
command-line options)
<table1 table2 ...>
Tables to be excluded/included
'export' command
NAME
neo4j-etl export - Export from RDBMS and import into NEO4J via CSV
files.
SYNOPSIS
neo4j-etl export [ {--columns | --cols} <Columns>... ]
[ --config-file <Configuration File> ]
[ --csv-directory <csv directory> ]
[ {-d | --database} <RDBMS database> ] [ --debug ]
[ --delimiter <delimiter> ] [ --destination <directory> ] [ {--driver | --jars} <--driver <PATH_TO_YOUR_JAR1> --driver <PATH_TO_YOUR_JAR2>>... ]
[ {--exclusion-mode | --exc} <Tables: exclude|include|none(default)> ]
[ {--exclusion-mode-column-type | --exctype} <exclude|include|none(default)> ]
[ {--exclusion-mode-columns | --excc} <exclude|include|none(default)> ]
[ {--exclusion-mode-tables | --exct} <exclude|include|none(default)> ]
[ --force ] [ --import-tool <import tool path> ]
[ --mapping-file <file|stdin> ] [ {--neo4j:password | --graph:password | --graph:neo4j:password} <neo4j password> ]
[ {--neo4j:url | --graph:url | --graph:neo4j:url} <neo4j url> ]
[ {--neo4j:user | --graph:user | --graph:neo4j:user} <neo4j user> ]
[ --options-file <option file> ] [ --output-mapping-file <file|stdout> ]
[ {-p | --port} <RDBMS port> ] [ --quote <quote> ]
[ {--rdbms:fetch-size | --fs} <fetch-size> ]
[ {--rdbms:password | --password} <RDBMS password> ]
[ {--rdbms:schema | -s | --schema} <schema> ]
[ {--rdbms:url | --url} <RDBMS url> ]
[ {--rdbms:user | -u | --user} <RDBMS user> ]
[ {--relationship-name | --rel-name} <table(default)|column> ]
[ --schemas <Schemas>... ] [ {--tables | --tabs} <Tables>... ]
[ --tiny-int <byte(default)|boolean> ] [ --types <Types>... ]
[ --using <import tool> ] [--] [ <table1 table2 ...>... ]
OPTIONS
--columns <Columns>, --cols <Columns>
Lists all columns to include/exclude by name or pattern
Use '-r' <PATTERN> to filter by regex, ex. '-r .*\.orders\..*_id' or
'northwind\.orders\..*_id' ,
'-g' <PATTERN> for grep syntax, ex. '-g .*\.orders\..*_id' or
'northwind\.orders\..*_id' ,
or '-l' <LIST> to list all columns names ex. '-l
northwind.customers.id,northwind.purchase.id,northwind.orders.id'
--config-file <Configuration File>
Specify the path to a file containing the configuration for the
selected command
--csv-directory <csv directory>
Path to directory for intermediate CSV files.
-d <RDBMS database>, --database <RDBMS database>
RDBMS database.
This option is required if any of the following options are
specified: host
--debug
Print detailed diagnostic output.
--delimiter <delimiter>
Delimiter to separate fields in CSV.
--destination <directory>
Path to destination store directory.
--driver <--driver <PATH_TO_YOUR_JAR1> --driver <PATH_TO_YOUR_JAR2>>, --jars <--driver <PATH_TO_YOUR_JAR1> --driver <PATH_TO_YOUR_JAR2>>
List of additional drivers as a list
--exclusion-mode <Tables: exclude|include|none(default)>, --exc <Tables: exclude|include|none(default)>
Specifies how to handle table exclusion. Options are mutually
exclusive.
exclude: Excludes specified tables from the process. All other
tables will be included.
include: Includes specified tables only. All other tables will be
excluded.
none: All tables are included in the process.
--exclusion-mode-column-type <exclude|include|none(default)>, --exctype
<exclude|include|none(default)>
Specifies how to handle column type exclusion. Options are mutually
exclusive.
exclude: Excludes specified columns types from the process. All
other columns types will be included.
include: Includes specified columns types only. All other columns
types will be excluded.
none: All columns types are included in the process.
--exclusion-mode-columns <exclude|include|none(default)>, --excc
<exclude|include|none(default)>
Specifies how to handle column exclusion. Options are mutually
exclusive.
exclude: Excludes specified columns from the process. All other
columns will be included.
include: Includes specified columns only. All other columns will be
excluded.
none: All columns are included in the process.
--exclusion-mode-tables <exclude|include|none(default)>, --exct
<exclude|include|none(default)>
Specifies how to handle table exclusion. Options are mutually
exclusive.
exclude: Excludes specified tables from the process. All other
tables will be included.
include: Includes specified tables only. All other tables will be
excluded.
none: All tables are included in the process.
--force
Force delete destination store directory if it already exists.
--import-tool <import tool path>
Path to directory containing Neo4j import tool. Mandatory only for Offline Bulk Import mode
--mapping-file <file|stdin>
Path to an existing metadata mapping file. The name 'stdin' will
cause the CSV resources definitions to be read from standard input.
--neo4j:password <neo4j password>, --graph:password <neo4j password>,
--graph:neo4j:password <neo4j password>
Password for login to Neo4j.
--neo4j:url <neo4j url>, --graph:url <neo4j url>, --graph:neo4j:url
<neo4j url>
Url to use for connection to Neo4j.
--neo4j:user <neo4j user>, --graph:user <neo4j user>, --graph:neo4j:user
<neo4j user>
User for login to Neo4j.
--options-file <option file>
Path to file containing Neo4j import tool options.
--output-mapping-file <file|stdout>
Path to the output metadata mapping file.
-p <RDBMS port>, --port <RDBMS port>
Port number to use for connection to RDBMS.
--quote <quote>
Character to treat as quotation character for values in CSV data.
--rdbms:fetch-size <fetch-size>, --fs <fetch-size>
RDBMS Fetch size
--rdbms:password <RDBMS password>, --password <RDBMS password>
Password for login to RDBMS.
This option is required if any of the following options are
specified: --rdbms:url, --url
--rdbms:schema <schema>, -s <schema>, --schema <schema>
RDBMS schema.
--rdbms:url <RDBMS url>, --url <RDBMS url>
Url to use for connection to RDBMS.
--rdbms:user <RDBMS user>, -u <RDBMS user>, --user <RDBMS user>
User for login to RDBMS.
This option is required if any of the following options are
specified: --rdbms:url, --url
--relationship-name <table(default)|column>, --rel-name
<table(default)|column>
Specifies whether to get the name for relationships from table names
or column names.
--schemas <Schemas>
Lists all schemas to include by name or pattern.
Use '-r' <PATTERN> to filter by regex, ex. '-r .*\.north.*',
'-g' <PATTERN> for grep syntax, ex. '-g .*\.north.*' ,
or '-l' <LIST> to list all schemas names ex. '-l northwind,exc'
--tables <Tables>, --tabs <Tables>
Lists all tables to include/exclude by name or pattern.
Use '-r' <PATTERN> to filter by regex, ex. '-r .*\.purchase.*' or
'northwind.purchase.*' ,
'-g' <PATTERN> for grep syntax, ex. '-g .*\.purchase.*' or
'northwind.purchase.*' ,
or '-l' <LIST> to list all tables names ex. '-l
customers,purchase,orders'
--tiny-int <byte(default)|boolean>
Specifies whether to convert TinyInt to byte or boolean
--types <Types>
Lists all column types to include/exclude by name separated by
commas. Valid values:
unknown,
binary,
bit,
character,
id,
integer,
real,
reference,
temporal,
url,
xml,
large_object,
object;
--using <import tool>
Import tool that will be used to load data into neo4j.
--
This option can be used to separate command-line options from the
list of arguments (useful when arguments might be mistaken for
command-line options)
<table1 table2 ...>
Tables to be excluded/included
Parameters Usage
There are two ways for write Etl parameters:
1) write parameters in command line:
$NEO4J_HOME/bin/neo4j-etl export|generate-metadata-mapping
--rdbms:url jdbc:oracle:thin:@localhost:49161:XE
--rdbms:user northwind --rdbms :password northwind
--rdbms:schema northwind
--using bulk:neo4j-import
--import-tool $NEO4J_HOME/bin
--csv-directory /tmp/northwind
--options-file /tmp/northwind/options.json
--quote '"' --force
...
2) use a config file:
$NEO4J_HOME/bin/neo4j-etl export|generate-metadata-mapping \
--config-file <path to .config file>
Above there is an Example of config file.
#EXAMPLE - ETL CONFIG FILE
#RDBMS
rdbms-url=url
rdbms-schema=schema
rdbms-password=neo4j
rdbms-user=neo4j
rdbms-fetch-size=10000
#NEO4J
using=cypher:direct
neo4j-url=bolt://127.0.0.1:7687
neo4j-user=neo4j
neo4j-password=neo4j
#RULES
exclusion-mode-tables=INCLUDE
tables=-l table1,table2,...
exclusion-mode-columns=INCLUDE
columns=-l column1,column2,...
exclusion-mode-column-types=EXCLUDE
column-types=type1,type2,...
#MISC
output-mapping-file=path_to_output_mapping_file
import-tool=path_to_import_tool
csv-directory=path_to_directory
mapping-file=path_to_file
debug=false
Example Session: Basic
Generate Metadata Mapping
export NEO4J_HOME=/path/to/neo4j-enterprise-3.4.0 mkdir -p /tmp/northwind $NEO4J_HOME/bin/neo4j-etl generate-metadata-mapping \ --rdbms:url jdbc:oracle:thin:@localhost:49161:XE \ --rdbms:user northwind --rdbms:password northwind \ --rdbms:schema northwind --output-mapping-file /tmp/northwind/mapping.json
Multi Database Support
Neo4j 4.0 Enterprise has multi-tenancy support, in order to support this feature we added a select options box in the import view as it follows:
If you use the command line tool you can specify the destination database by passing
the --neo4j:database-name option in this way:
echo '{ "multiline-fields" : "true" }' > /tmp/northwind/options.json
$NEO4J_HOME/bin/neo4j-etl export \
--rdbms:url jdbc:oracle:thin:@localhost:49161:XE \
--rdbms:user northwind --rdbms:password northwind \
--rdbms:schema northwind \
--using cypher:direct | cypher:batch | cypher:fromSQL \
--neo4j:url bolt://localhost:7687 \
--neo4j:user neo4j
--neo4j:password neo4j \
--neo4j:database-name myDatabase \
--import-tool $NEO4J_HOME/bin \
--csv-directory /tmp/northwind \
--options-file /tmp/northwind/options.json \
--quote '"' --force
Offline Bulk Import via neo4j-import tool for initial load (Neo4j database must be empty)
Please consider that since the neo4j-admin import removed the support for json files as source
of the options, we could have problems in case of long command lines derived from db with an high number
of entities involved into the process.
echo '{ "multiline-fields" : "true" }' > /tmp/northwind/options.json
$NEO4J_HOME/bin/neo4j-etl export \
--rdbms:url jdbc:oracle:thin:@localhost:49161:XE \
--rdbms:user northwind --rdbms :password northwind \
--rdbms:schema northwind \
--using bulk:neo4j-import \
--import-tool $NEO4J_HOME/bin \
--csv-directory /tmp/northwind \
--options-file /tmp/northwind/options.json \
--quote '"' --force
$NEO4J_HOME/bin/neo4j-shell -path $NEO4J_HOME/data/databases/graph.db/ -c 'MATCH (n) RETURN labels(n), count(*);' +--------------------------+ | labels(n) | count(*) | +--------------------------+ | ["Shipper"] | 3 | | ["Employee"] | 9 | | ["Region"] | 4 | | ["Customer"] | 93 | | ["Territory"] | 53 | | ["Product"] | 77 | | ["Supplier"] | 29 | | ["Order"] | 830 | | ["Category"] | 8 | +--------------------------+ 9 rows
Online Batch Import via java-bolt-driver for incremental load (neo4j can be already populated)
It can be done in 3 modes:
-
'cypher:direct' (LOAD CSV)
-
'cypher:batch' (which creates CSV files in the same way as cypher:direct but it translate them into UNWIND statements instead of LOAD CSV)
-
'cypher:fromSQL' (which translate tabular result data from the RDBMS, without creating CSV files)
In online mode via java-bolt-diver, in order to allow importing CSV from temporary folder is necessary:
-
enable property
dbms.security.allow_csv_import_from_file_urlsand set it to true in order to allow LOAD CSV to read CSV files from external resource -
remove property
dbms.directories.import=importor comment it out to prevent Neo4j to search for CSV file into the default import folder
For Neo4j remote instances you must use cypher:batch import mode
echo '{ "multiline-fields" : "true" }' > /tmp/northwind/options.json
$NEO4J_HOME/bin/neo4j-etl export \
--rdbms:url jdbc:oracle:thin:@localhost:49161:XE \
--rdbms:user northwind --rdbms:password northwind \
--rdbms:schema northwind \
--using cypher:direct | cypher:batch | cypher:fromSQL \
--neo4j:url bolt://localhost:7687 \
--neo4j:user neo4j --neo4j:password neo4j \
--import-tool $NEO4J_HOME/bin \
--csv-directory /tmp/northwind \
--options-file /tmp/northwind/options.json \
--quote '"' --force
Additional command line options for `cypher:batch` and `cypher:fromSQL` import modes:
--unwindBatchSize <value> (Batch size that will be used for unwind data) \
--txBatchSize <value> (Transaction Batch size that will be used for unwind commit) \
$NEO4J_HOME/bin/cypher-shell -a bolt://localhost:7687 -u neo4j -p neo4j 'MATCH (n) RETURN labels(n), count(*);' +--------------------------+ | labels(n) | count(*) | +--------------------------+ | ["Shipper"] | 3 | | ["Employee"] | 9 | | ["Region"] | 4 | | ["Customer"] | 93 | | ["Territory"] | 53 | | ["Product"] | 77 | | ["Supplier"] | 29 | | ["Order"] | 830 | | ["Category"] | 8 | +--------------------------+ 9 rows
Example Session: Docker + Northwind
This example session is based on the Northwind example dataset.
DDL scripts are available here:
MySQL
Download, start and configure the docker container with MySQL:
docker pull mysql docker run --name neo4j-etl-mysql -e MYSQL_ROOT_PASSWORD=admin -e MYSQL_DATABASE=northwind -e MYSQL_USER=neo4j -e MYSQL_PASSWORD=neo4j -d -p 3306:3306 mysql:latest docker exec -it neo4j-etl-mysql bash root@eb6f279fdb88:/# mysql -u root -p Enter password: admin Welcome to the MySQL monitor. Commands end with ; or \g. Your MySQL connection id is 3 Server version: 5.7.18 MySQL Community Server (GPL) Copyright (c) 2000, 2017, Oracle and/or its affiliates. All rights reserved. Oracle is a registered trademark of Oracle Corporation and/or its affiliates. Other names may be trademarks of their respective owners. Type 'help;' or '\h' for help. Type '\c' to clear the current input statement. mysql> grant all privileges on *.* to 'neo4j'@'%' with grant option; Query OK, 0 rows affected (0.00 sec) mysql> quit; Bye root@bf99fbc0d31c:/# exit exit
Load the database via the following sql script: https://raw.githubusercontent.com/neo4j-contrib/neo4j-etl/master/neo4j-etl-it/src/main/resources/scripts/mysql/northwind.sql
export NEO4J_HOME=/path/to/neo4j-enterprise-3.4.0
mkdir -p /tmp/northwind
echo '{ "multiline-fields" : "true" }' > /tmp/northwind/options.json
./bin/neo4j-etl export \
--rdbms:url jdbc:mysql://localhost:5433/northwind?autoReconnect=true&useSSL=false \
--rdbms:user neo4j --rdbms:password neo4j \
--import-tool $NEO4J_HOME/bin \
--options-file /tmp/northwind/options.json \
--csv-directory /tmp/northwind \
--destination $NEO4J_HOME/data/databases/graph.db/ \
--quote '"' --force
PostgreSQL
Download, start and configure the docker container with PostgreSQL 9.6.2:
docker pull postgres docker run --name neo4j-etl-postgres -e POSTGRES_USER=neo4j -e POSTGRES_PASSWORD=neo4j -d -p 5433:5432 postgres docker run -it --rm --link neo4j-etl-postgres:postgres postgres psql -h postgres -U neo4j Password for user neo4j: psql (9.6.2) Type "help" for help. neo4j=# DROP DATABASE IF EXISTS northwind; neo4j=# CREATE DATABASE northwind WITH OWNER 'neo4j' ENCODING 'UTF8' LC_COLLATE = 'en_US.utf8' LC_CTYPE = 'en_US.utf8'; neo4j=# \q
Load the database via the following sql script: northwind.sql
export NEO4J_HOME=/path/to/neo4j-enterprise-3.4.0
mkdir -p /tmp/northwind
echo '{"multiline-fields":"true"}' > /tmp/northwind/options.json
./bin/neo4j-etl export \
--rdbms:url jdbc:postgresql://localhost:5433/northwind?ssl=false \
--rdbms:user neo4j --rdbms:password neo4j \
--import-tool $NEO4J_HOME/bin \
--options-file /tmp/northwind/options.json \
--csv-directory /tmp/northwind \
--destination $NEO4J_HOME/data/databases/graph.db/ \
--quote '"' --force
Oracle
Download, start and configure the docker container with Oracle XE 11g:
docker pull wnameless/oracle-xe-11g
docker run --name neo4j-etl-oracle -d -p 49160:22 -p 49161:1521 wnameless/oracle-xe-11g
ssh root@localhost -p 49160
root@localhost's password: admin
Welcome to Ubuntu 16.04.1 LTS (GNU/Linux 4.9.13-moby x86_64)
* Documentation: https://help.ubuntu.com
* Management: https://landscape.canonical.com
* Support: https://ubuntu.com/advantage
Last login: Mon May 1 17:32:48 2017 from 172.17.0.1
root@692c446a274b:~# wget https://raw.githubusercontent.com/neo4j-contrib/neo4j-etl/master/neo4j-etl-it/src/main/resources/scripts/oracle/northwind.sql
root@692c446a274b:~# sqlplus system/oracle
SQL> CREATE USER northwind IDENTIFIED BY northwind;
SQL> GRANT DBA TO northwind;
SQL> CONN northwind/northwind;
SQL> SET sqlblanklines ON;
SQL> @northwind.sql
SQL> quit;
root@692c446a274b:~# exit
export NEO4J_HOME=/path/to/neo4j-enterprise-3.4.0
mkdir -p /tmp/northwind
echo '{"multiline-fields":"true"}' > /tmp/northwind/options.json
./bin/neo4j-etl export \
--rdbms:url jdbc:oracle:thin:@localhost:49161:XE \
--rdbms:user northwind --rdbms:password northwind \
--rdbms:schema northwind \
--import-tool $NEO4J_HOME/bin \
--options-file /tmp/northwind/options.json \
--csv-directory /tmp/northwind \
--destination $NEO4J_HOME/data/databases/graph.db/ \
--quote '"' --force
--driver /tmp/ojdbc6-11.2.0.3.jar
Microsoft SQL
Download, start and configure the docker container with Microsoft SQL Server:
docker run --name neo4j-etl-mssql -e 'ACCEPT_EULA=Y' -e 'SA_PASSWORD=Passw0rd!' -p 1433:1433 -d microsoft/mssql-server-linux
If you want to connect to Microsoft SQL client console then you can run the following command:
docker exec -it neo4j-etl-mssql /opt/mssql-tools/bin/sqlcmd -S localhost -U sa -P 'Passw0rd!' -d <DATABASE>
export NEO4J_HOME=/path/to/neo4j-enterprise-3.4.0
mkdir -p /tmp/wideworldimporters
echo '{"multiline-fields":"true"}' > /tmp/wideworldimporters/options.json
./bin/neo4j-etl export \
--rdbms:password "Passw0rd!" \
--rdbms:user sa \
--rdbms:url "jdbc:sqlserver://localhost:1433;databaseName=WideWorldImporters" \
--import-tool $NEO4J_HOME/bin \
--options-file /tmp/wideworldimporters/options.json \
--csv-directory /tmp/wideworldimporters \
--destination $NEO4J_HOME/data/databases/graph.db/ \
--driver /tmp/mssql-jdbc-6.1.0.jre8.jar \
How to import World Wide Importers database into a MS SQL server Docker instance
# Create docker instance for MS-SQL Server docker run --name mssql-etl \ -e MSSQL_COLLATION=Latin1_General_100_CI_AS \ -e 'ACCEPT_EULA=Y' \ -e 'SA_PASSWORD=<YOUR_PASSWORD>' \ -p 1433:1433 \ -v /tmp:/tmp \ -d microsoft/mssql-server-linux:2017-latest # Download World Wide Importers backup file wget https://github.com/Microsoft/sql-server-samples/releases/download/wide-world-importers-v1.0/WideWorldImporters-Full.bak # Create a backup directory sudo docker exec -it mssql-etl mkdir /var/opt/mssql/backup # Load backup file into the container sudo docker cp WideWorldImporters-Full.bak mssql-etl:/var/opt/mssql/backup # Restore Wide World Importers database sudo docker exec -it mssql-etl /opt/mssql-tools/bin/sqlcmd \ -S localhost \ -U SA \ -P '<YOUR_PASSWORD>' \ -Q 'RESTORE FILELISTONLY FROM DISK = "/var/opt/mssql/backup/WideWorldImporters-Full.bak"' \ | tr -s ' ' \ | cut -d ' ' -f 1-2 sudo docker exec -it mssql-etl /opt/mssql-tools/bin/sqlcmd \ -S localhost \ -U SA \ -P '<YOUR_PASSWORD>' \ -Q 'RESTORE DATABASE WideWorldImporters FROM DISK = "/var/opt/mssql/backup/WideWorldImporters-Full.bak" WITH MOVE "WWI_Primary" TO "/var/opt/mssql/data/WideWorldImporters.mdf", MOVE "WWI_UserData" TO "/var/opt/mssql/data/WideWorldImporters_userdata.ndf", MOVE "WWI_Log" TO "/var/opt/mssql/data/WideWorldImporters.ldf", MOVE "WWI_InMemory_Data_1" TO "/var/opt/mssql/data/WideWorldImporters_InMemory_Data_1"'