Bug#675106: ITP: pgbulkload -- A high speed data loading utility for PostgreSQL
Package: wnpp
Severity: wishlist
Owner: Alexander Kuznetsov <acca@cpan.org>
* Package name : pgbulkload
Version : 3.1.1
Upstream Author :
Takahiro Itagaki itagaki.takahiro @nospam@ gmail.com
Masao Fujii masao.fujii @nospam@ gmail.com
Mitsuru Hasegawa hasegawa @nospam@ metrosystems.co.jp
Masahiko Sakamoto sakamoto_masahiko_b1 @nospam@ lab.ntt.co.jp
Toru SHIMOGAKI shimogaki.toru @nospam@ oss.ntt.co.jp
* URL : http://pgfoundry.org/projects/pgbulkload/
* License : BSD
Programming Lang: C, SQL
Description : A high speed data loading utility for PostgreSQL
pg_bulkload is designed to load huge amount of data to a database.
You can choose whether database constraints are checked and how many errors are
ignored during the loading. For example, you can skip integrity checks for
performance when you copy data from another database to PostgreSQL. On the
other hand, you can enable constraint checks when loading unclean data.
.
The original goal of pg_bulkload was an faster alternative of COPY command in
PostgreSQL, but version 3.0 or later has some ETL features like input data
validation and data transformation with filter functions.
.
In version 3.1, pg_bulkload can convert the load data into the binary file
which can be used as an input file of pg_bulkload. If you check whether
the load data is valid when converting it into the binary file, you can skip
the check when loading it from the binary file to a table. Which would reduce
the load time itself. Also in version 3.1, parallel loading works
more effectively than before.
Reply to: