Blocks Archive

Regular backups of the blocks data directory so you can fully sync with the EOS network.

Title Download Size MD5 Checksum
blocks_2018-12-19-07-01.tar.gz Wasabi S3 81.06 GiB 9078a7b694f0a4d19c607a3ceaa46f61
blocks_2018-12-18-07-01.tar.gz Wasabi S3 80.45 GiB abf5474defdc93b7240f59b3a21b7b38
blocks_2018-12-17-07-01.tar.gz Wasabi S3 80 GiB 22b7fa855b5f6ad8614acb9dad5ca60b
blocks_2018-12-16-07-01.tar.gz Wasabi S3 79.52 GiB 07b0de1baf0b7f9d46f8b6815ee79e03
blocks_2018-12-15-07-04.tar.gz Wasabi S3 78.75 GiB 83dcb67e717145a88682a4084d3155c3
blocks_2018-12-14-07-01.tar.gz Wasabi S3 78.02 GiB 0d448f0f7caaa4df9fec3e1a5525dfa9
blocks_2018-12-13-07-01.tar.gz Wasabi S3 77.2 GiB fc070c9fdbe5fe28a63a9b44696b1261
blocks_2018-12-12-07-01.tar.gz Wasabi S3 76.44 GiB bab5ebbb73486ad9be4cb6b4a55b53cf

The blocks archives are taken daily from our bank of API nodes. These backups can be used across all node configurations and have been tested with Ubuntu, Centos and Debian.

How To Use

Download the archive, uncompress it into your data directory and start up nodeos requesting a hard replay which deletes the state database. This will validate the blocks, rebuild your state and sync with the live chain.

The example assumes you have used our automation framework to install and configure the EOS application. It includes handy bash helpers to auto dameonise the nodeos process and capture all output into a single log file.

You can use the one-liner in the example to always download the latest backup. We also have a Blocks API which orders the archives in chronological order, newest first.

# Move to your local eos directory, removing the existing data directories (if relevant)
cd /opt/mainnet
rm -rf blocks state

# Download the latest blocks backup
wget $(wget --quiet "" -O- | jq -r '.data[0].s3') -O blocks_backup.tar.gz

# Uncompress to ./blocks
tar xvzf blocks_backup.tar.gz

# Start the chain and replay from the blocks backup
./ --hard-replay --wasm-runtime wabt

# Tail the logs to watch the sync process
tail -f log.txt
2018-08-13T09:42:10.168 initializing chain plugin
2018-08-13T09:42:10.170 Hard replay requested: deleting state database
2018-08-13T09:42:10.171 Recovering Block Log...
2018-08-13T09:42:10.171 Moved existing blocks directory to backup location: '/mnt/blocks-2018-08-13T09:42:10.171'
2018-08-13T09:42:10.172 Reconstructing '/mnt/blocks/blocks.log' from backed up block log
2018-08-13T09:44:33.490 Existing block log was undamaged. Recovered all irreversible blocks up to block number 10887835.
2018-08-13T09:44:33.493 Reversible blocks database was not corrupted. Copying from backup to blocks directory.
2018-08-13T09:44:38.833 Log is nonempty
2018-08-13T09:44:38.833 Index is empty
2018-08-13T09:44:38.833 Reconstructing Block Log Index...
2018-08-13T09:47:12.722 No head block in fork db, perhaps we need to replay
2018-08-13T09:47:12.722 Initializing new blockchain with genesis state
2018-08-13T09:47:12.755 existing block log, attempting to replay 10887835 blocks
    140700 of 10887835

How Long To Replay?

Once you kick off the hard-replay, the sync will take hours. Exactly how long is dependent on your system configuration. The replay process is mostly CPU bound, as nodeos is single threaded the important factor is your CPU clock speed, not the overall number of cores.

When you replay, you should follow the nodeos log. The code snippet on the left shows you an example of the log messages that you should see when you execute the hard-replay. After the initial validation you get a progress output to give you a better indication of the time it will take.