Steem Developer Portal
Applications that interface directly with the Steem blockchain will need to connect to a steemd
node. Developers may choose to use one of the public API nodes that are available, or run their own instance of a node.
Public Nodes
Although steemd
fully supports WebSockets (wss://
and ws://
) public nodes typically do not. All nodes listed use HTTPS (https://
). If you require WebSockets for your solutions, please consider setting up your own steemd
node or proxy WebSockets to HTTPS using lineman.
URL | Owner |
---|---|
api.steem.fans | @ety001 |
steem.61bts.com | @ety001 |
api.steemyy.com | @justyy |
steem.justyy.workers.dev | @justyy |
api.justyy.com | @justyy |
api.steemitdev.com | @steemit |
api.steemit.com | @steemit |
For a report on the latest public full nodes, check the latest posts on @fullnodeupdate by @holger80. Another excellent tool for checking real-time full node status is geo.steem.pl by @jamzed.
Private Nodes
The simplest way to get started is by deploying a pre-built dockerized container.
Dockerized p2p Node
To run a p2p node (ca. 2GB of memory is required at the moment):
Dockerized Full Node
to run a node with all the data (e.g. for supporting a content website) that uses ca. 14GB of memory and growing:
Syncing blockchain
Normally syncing blockchain starts from very first, 0
genesis block. It might take long time to catch up with live network. Because it connectes to various p2p nodes in the Steem network and requests blocks from 0 to head block. It stores blocks in block log file and builds up the current state in the shared memory file. But there is a way to bootstrap syncing by using trusted block_log
file. The block log is an external append only log of the blocks. It contains blocks that are only added to the log after they are irreversible because the log is append only.
Both block_log
files updated periodically, as of May 2018 uncompressed block_log
file size ~110 GB. Docker container on stable
branch of Steem source code has option to use USE_PUBLIC_BLOCKLOG=1
to download latest block log and start Steem node with replay.
Block log should be place in blockchain
directory below data_dir
and node should be started with --replay-blockchain
to ensure block log is valid and continue to sync from the point of snapshot. Replay uses the downloaded block log file to build up the shared memory file up to the highest block stored in that snapshot and then continues with sync up to the head block.
Replay helps to sync blockchain in much faster rate, but as blockchain grows in size replay might also take some time to verify blocks.
There is another trick which might help with faster sync/replay on smaller equipped servers:
while :
do
dd if=blockchain/block_log iflag=nocache count=0
sleep 60
done
Above bash script drops block_log
from the OS cache, leaving more memory free for backing the blockchain database. It might also help while running live, but measurement would be needed to determine this.
Few other tricks that might help:
For Linux users, virtual memory writes dirty pages of the shared file out to disk more often than is optimal which results in steemd being slowed down by redundant IO operations. These settings are recommended to optimize reindex time.
echo 75 | sudo tee /proc/sys/vm/dirty_background_ratio
echo 1000 | sudo tee /proc/sys/vm/dirty_expire_centisecs
echo 80 | sudo tee /proc/sys/vm/dirty_ratio
echo 30000 | sudo tee /proc/sys/vm/dirty_writeback_centisecs
Another settings that can be changed in config.ini
is flush
- it is to specify a target number of blocks to process before flushing the chain database to disk. This is needed on Linux machines and a value of 100000 is recommended. It is not needed on OS X, but can be used if desired.
docker run \
-d -p 2001:2001 -p 8090:8090 --name steemd-default \
steemit/steem
docker logs -f steemd-default # follow along
docker run \
--env USE_WAY_TOO_MUCH_RAM=1 \
-d -p 2001:2001 -p 8090:8090 --name steemd-full \
steemit/steem
docker logs -f steemd-full