CYBERTEC PostgreSQL Logo

Bulk load performance in PostgreSQL

08.2023 / Category: / Tags: | |
A fork lift operator knows all about bulk load
© Laurenz Albe 2023

 

There are several techniques to bulk load data into PostgreSQL. I decided to compare their performance in a simple test case. I'll add some recommendations for parameter settings to improve the performance even more.

An example table to bulk load data

The table is simple enough:

It is a narrow table (only two columns), but it has a primary key index. Loading data would be much faster without the index, but in real life you cannot always drop all indexes and constraints before loading data. Moreover, dropping indexes before the bulk load and re-creating them afterwards can be slower, if the table already contains data.

For the performance test, I'm going to load 10 million rows that look like (1, '1'), (2, '2') and so on, counting up to 10000000. The test will be performed on PostgreSQL v16 with the default configuration. I am aware that the default configuration is not perfect for bulk loading, but that does not bother me, since I am only interested in comparing the different methods.

Different methods for bulk loading

I'll try the following six techniques:

Single INSERTs in autocommit mode (the fool's way)

This will be terribly slow, since each statement will run in its own transaction. For each transaction, PostgreSQL has to write the WAL (the transaction log) out to disk, which leads to 10 million I/O requests.

Naturally, this is not the correct thing to do for a bulk load. But it is the way a transactional application loads data into the database, with many clients inserting small amounts of data, and no way to bundle the individual requests into bigger transactions. There are some remedies available to boost performance in such a case:

  • Set synchronous_commit to off. That will boost performance amazingly, but an operating system crash could lead to some committed transactions getting lost.
  • Set commit_delay to a value greater than 0 and tune commit_siblings. That can reduce the number of I/O requests. The effect won't be as marked as with synchronous_commit, but you can never lose a committed transaction.

Single INSERTs in one transaction

The only difference to the previous test is that we will insert all 10 million rows in a single transaction. That is bound to boost performance quite a lot.

Single INSERTs with a prepared statement in one transaction

The difference to the previous test is that we use a prepared statement:

and perform the inserts like this:

The performance should be better, because PostgreSQL can reuse the execution plan for the INSERT statement rather than planning it 10 million times. With short statements, that can be a notable performance improvement.

Multi-line INSERTs in one transaction

You can insert several rows with a single INSERT statement:

For this test, I'll insert 1000 rows per statement, so there will be 10000 such INSERT statements. The benefit is that there are fewer client-server round trips, and PostgreSQL has to plan and execute fewer statements.

Multi-line INSERTs with a prepared statement in one transaction

This test is like the previous one, except that I'll use a prepared statement:

The idea is to combine the benefits of prepared statements and multi-line INSERTs.

Bulk load with COPY (the king's way)

It is well known that COPY is the fastest way to load data into PostgreSQL. Let's see how much better it really is! For this test, I'll load all 10 million rows with a single COPY statement.

The down side of COPY is that it is a non-standard SQL statement, so not all APIs support it, and you cannot use it in programs that are to support other database systems as well.

Note that you can use COPY (FREEZE) if you create or truncate the table in the same transaction. That won't speed up loading, but it avoids the overhead of setting hint bits on the rows by the first reader and anti-wraparound autovacuum on the table.

Performance comparison of the bulk load methods

The tests were run on my Laptop with Fedora Linux 37, untuned PostgreSQL v16 with local domain socket connections and ext4 file systems on an NVMe disk.

Bulk load method Duration
single INSERTs, many transactions 8954 s
single INSERTs 841 s
single INSERTs, prepared statement 688 s
bulk INSERTs 52 s
bulk INSERTs, prepared statement 57 s
COPY 14 s

Discussion of the bulk load performance test results

Most test results are as expected:

  • running each INSERT in its own transaction is unbearable slow
  • using prepared statements instead for the small INSERTs is a win
  • few big statements perform better than many small ones
  • nothing is as fast as COPY

What may surprise is that using prepared statements for the multi-line statements slows down processing. I am not certain what the reason is, but it could be the overhead of the extended query protocol.

Tuning PostgreSQL for bulk load

For completeness' sake, I will add some hints how you can configure PostgreSQL to speed up bulk loads. There is little you can to to speed up writes, but you can reduce the number of redundant, unnecessary writes.

  • The most important parameter to tune is max_wal_size. The default value is 1GB. If more than that amount of WAL has been written since the latest checkpoint, PostgreSQL will trigger another checkpoint. The idea is to keep crash recovery time short. But 1GB is very little when it comes to bulk loading, and you can end up having a checkpoint every couple of seconds. That is an unnecessary overhead, particularly since PostgreSQL might have to flush the same (index) pages to disk again and again.
  • For similar reasons, you should increase checkpoint_timeout. Fewer checkpoints mean fewer redundant writes. Moreover, reducing the number of checkpoints will reduce the size of the WAL, since fewer “full-page images” have to be written.

Note that tuning these parameters won't improve my special test case a lot, since the table is only extended and the index values are increasing. B-tree indexes are optimized for that usage pattern. If I had chosen UUIDs as primary keys, the index modifications would not be localized, and reducing the number of full-page images written would matter more.

Conclusion

For bulk load, COPY is the way to go.

Further Reading

Speed up PostgreSQL data loading with COPY (FREEZE) Learn how this feature works with hint bits.

If you are interested in PostgreSQL performance, you may also want to read about the performance of various methods of auto-generated primary keys or ways of storing large binary data to enhance performance.


In order to receive regular updates on important changes in PostgreSQL, subscribe to our newsletter, or follow us on Twitter, Facebook, or LinkedIn.

6 responses to “Bulk load performance in PostgreSQL”

  1. network speed between servers and disk speed is also important when moving and processing data. filesystem types, how to mount then (ext4 nodiratime, nobarrier, noatime)

    • I have added some more about the connections and the file system used, thanks.
      I don't think that the file system has a big impact, but the connections probably do.

  2. > For this test, I’ll insert 1000 rows per statement, so there will be 10000 such INSERT statements

    How does tuning that affect the performance, i.e. more rows/statement ?

    Thanks for an interesting article.

    • I didn't test it, so I don't know. My guess is that beyond a certain number of rows (less than 1000) there will be no more performance improvement, and if you keep increasing, the performance will go down again.

Leave a Reply

Your email address will not be published. Required fields are marked *

CYBERTEC Logo white
Get the newest PostgreSQL Info & Tools


    This site is protected by reCAPTCHA and the Google Privacy Policy & Terms of Service apply.

    ©
    2024
    CYBERTEC PostgreSQL International GmbH
    phone-handsetmagnifiercrosscross-circle
    linkedin facebook pinterest youtube rss twitter instagram facebook-blank rss-blank linkedin-blank pinterest youtube twitter instagram