[Date Prev][Date Next] [Thread Prev][Thread Next] [Date Index] [Thread Index]

Bug#675106: ITP: pgbulkload -- A high speed data loading utility for PostgreSQL

Package: wnpp
Severity: wishlist
Owner: Alexander Kuznetsov <acca@cpan.org>

* Package name    : pgbulkload
  Version         : 3.1.1
  Upstream Author :
		Takahiro Itagaki	itagaki.takahiro @nospam@ gmail.com
		Masao Fujii		masao.fujii @nospam@ gmail.com
		Mitsuru Hasegawa	hasegawa @nospam@ metrosystems.co.jp
		Masahiko Sakamoto	sakamoto_masahiko_b1 @nospam@ lab.ntt.co.jp
		Toru SHIMOGAKI		shimogaki.toru @nospam@ oss.ntt.co.jp
* URL             : http://pgfoundry.org/projects/pgbulkload/
* License         : BSD
  Programming Lang: C, SQL
  Description     : A high speed data loading utility for PostgreSQL
 pg_bulkload is designed to load huge amount of data to a database.
 You can choose whether database constraints are checked and how many errors are
 ignored during the loading. For example, you can skip integrity checks for
 performance when you copy data from another database to PostgreSQL. On the
 other hand, you can enable constraint checks when loading unclean data.
 The original goal of pg_bulkload was an faster alternative of COPY command in
 PostgreSQL, but version 3.0 or later has some ETL features like input data
 validation and data transformation with filter functions.
 In version 3.1, pg_bulkload can convert the load data into the binary file
 which can be used as an input file of pg_bulkload. If you check whether
 the load data is valid when converting it into the binary file, you can skip
 the check when loading it from the binary file to a table. Which would reduce
 the load time itself. Also in version 3.1, parallel loading works
 more effectively than before.

Reply to: