Bored of syncing nodes? Let EOS Infra take care of that for you 👌

Blocks Archive

Regular backups of the blocks data directory so you can fully sync with the EOS network.

Title Download Size MD5 Checksum
blocks_2019-03-19-07-02.tar.gz Wasabi S3 134.39 GiB 5e491c7de12fe655ca37c2fa275c2722
blocks_2019-03-18-07-02.tar.gz Wasabi S3 133.92 GiB d2af853e0247b5ba8b8a54ebf6b37b44
blocks_2019-03-17-07-02.tar.gz Wasabi S3 133.47 GiB 5d8ea3dd04cf283126bdc1a93c738579
blocks_2019-03-16-07-02.tar.gz Wasabi S3 133.04 GiB 36b89db3f3622f1abe8ef36e920e99ec
blocks_2019-03-15-07-02.tar.gz Wasabi S3 132.54 GiB ce4c38955ef8347b6ab53d80364d783f
blocks_2019-03-14-07-02.tar.gz Wasabi S3 132.04 GiB e6b91711466d85ab0b469f0a6b23900a
blocks_2019-03-13-07-02.tar.gz Wasabi S3 131.58 GiB d8d884c56d6e3a2818f434c46abbbd2c

The blocks archives are taken daily from our bank of API nodes. These backups can be used across all node configurations and have been tested with Ubuntu, Centos and Debian.

How To Use

Download the archive, uncompress it into your data directory and start up nodeos requesting a hard replay which deletes the state database. This will validate the blocks, rebuild your state and sync with the live chain.

The example assumes you have used our automation framework to install and configure the EOS application. It includes handy bash helpers to auto dameonise the nodeos process and capture all output into a single log file.

You can use the one-liner in the example to always download the latest backup. We also have a Blocks API which orders the archives in chronological order, newest first.

# Move to your local eos directory, removing the existing data directories (if relevant)
cd /opt/mainnet
rm -rf blocks state

# Download the latest blocks backup
wget $(wget --quiet "" -O- | jq -r '.data[0].s3') -O blocks_backup.tar.gz

# Uncompress to ./blocks
tar xvzf blocks_backup.tar.gz

# Start the chain and replay from the blocks backup
./ --hard-replay --wasm-runtime wabt

# Tail the logs to watch the sync process
tail -f log.txt
2018-08-13T09:42:10.168 initializing chain plugin
2018-08-13T09:42:10.170 Hard replay requested: deleting state database
2018-08-13T09:42:10.171 Recovering Block Log...
2018-08-13T09:42:10.171 Moved existing blocks directory to backup location: '/mnt/blocks-2018-08-13T09:42:10.171'
2018-08-13T09:42:10.172 Reconstructing '/mnt/blocks/blocks.log' from backed up block log
2018-08-13T09:44:33.490 Existing block log was undamaged. Recovered all irreversible blocks up to block number 10887835.
2018-08-13T09:44:33.493 Reversible blocks database was not corrupted. Copying from backup to blocks directory.
2018-08-13T09:44:38.833 Log is nonempty
2018-08-13T09:44:38.833 Index is empty
2018-08-13T09:44:38.833 Reconstructing Block Log Index...
2018-08-13T09:47:12.722 No head block in fork db, perhaps we need to replay
2018-08-13T09:47:12.722 Initializing new blockchain with genesis state
2018-08-13T09:47:12.755 existing block log, attempting to replay 10887835 blocks
    140700 of 10887835

How Long To Replay?

Once you kick off the hard-replay, the sync will take hours. Exactly how long is dependent on your system configuration. The replay process is mostly CPU bound, as nodeos is single threaded the important factor is your CPU clock speed, not the overall number of cores.

When you replay, you should follow the nodeos log. The code snippet on the left shows you an example of the log messages that you should see when you execute the hard-replay. After the initial validation you get a progress output to give you a better indication of the time it will take.