sergioromcar / pentaho-vertica-bulkloader

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

pentaho-vertica-bulkloader

This is a plugin for performing bulk data loads into a Vertica Analytic Database using a COPY FROM STDIN statement through JDBC

TODO/LIMITATIONS:

No data escaping - If there is a . the load will silently be truncated. If a field contains the delimiter or record terminator, that record will be rejected. The step should probably be enhanced with a few more options to enable checking/escaping.

Efficiency of InputStream implementation - I'm currently using a Piped Input/Output Stream pair with a worker thread to hand the data off to the JDBC driver. I will evaluate whether it would be better to instead create a custom InputStream implementation that can be fed directly. (I'm guessing "yes!")

Speed - On a dual quad xeon with a GB connection to the cluster, I'm averaging about 40k to 50k records per second for a single load stream. I think this could improve.

Error row handling - The step doesn't have any way to be notified of rejected records if it is not set to abort. If it could receive notification of rejected records, we could even enable error handling on the step and allow the ETL developer to perform additional processing on those failed records.

Statement canceling - If the transformation aborts, the step receives an exception from the JDBC driver when attempting to cancel the statement. It currently has to use stack trace parsing to determine if it should report this error or not, and Vertica logs some errors that I'd rather avoid seeing. Need a better way to cancel an in-process stream.

Update to Vertica 4.0 and 5.0 - Most important is utilizing the new bulk loading features in those versions.

TEST PLAN:

Small/Medium/Large data loads Different delimiter characters including ' as a special case Different null strings and record terminator strings including ones with the character ' as a special case Abort vs non-abort Direct vs non-direct Specifying different types of logging filenames Multiple load step copies on one machine Clustered load streams Clustered load streams with multiple step copies per machine Special character encodings?

Pre-requisites for building the project:

  • Maven, version 3+
  • Java JDK 1.8
  • This settings.xml in your /.m2 directory

Building it

Build for nightly/release

All required profiles are activated by the presence of a property named "release".

$ mvn clean install -Drelease

This will build, unit test, and package the whole project (all of the sub-modules). The artifact will be generated in: target

Build for CI/dev

The release builds will compile the source for production (meaning potential obfuscation and/or uglification). To build without that happening, just eliminate the release property.

$ mvn clean install

Running the tests

Unit tests

This will run all tests in the project (and sub-modules).

$ mvn test

If you want to remote debug a single java unit test (default port is 5005):

$ cd core
$ mvn test -Dtest=<<YourTest>> -Dmaven.surefire.debug

Integration tests In addition to the unit tests, there are integration tests in the core project.

$ mvn verify -DrunITs

To run a single integration test:

$ mvn verify -DrunITs -Dit.test=<<YourIT>>

To run a single integration test in debug mode (for remote debugging in an IDE) on the default port of 5005:

$ mvn verify -DrunITs -Dit.test=<<YourIT>> -Dmaven.failsafe.debug

IntelliJ

  • Don't use IntelliJ's built-in maven. Make it use the same one you use from the commandline.
    • Project Preferences -> Build, Execution, Deployment -> Build Tools -> Maven ==> Maven home directory

About

License:GNU Lesser General Public License v2.1


Languages

Language:Java 100.0%