Comment out the line sa.Column('geo_location', Geography(geometry_type='POINT', srid=4326)),
in /hive/db/schema.py, then start hive_sync
and terminate after a few seconds when the tables have been generated.
Install PostGIS and run the command CREATE EXTENSION postgis;
on your database (using psql).
You also need to create the commented-put table with ALTER TABLE "hive_posts_cache" ADD COLUMN geo_location geography(POINT,4326);
in psql.
Now comment in the previously commented out line and start hive_sync again.
Add indexes for columns queried by the API, e.g.
CREATE INDEX ON hive_posts_cache(is_travelfeed);
CREATE INDEX ON hive_posts_cache(author);
CREATE INDEX ON hive_posts_cache(permlink);
CREATE INDEX ON hive_posts(parent_id);
Hive is a "consensus interpretation" layer for the Steem blockchain, maintaining the state of social features such as post feeds, follows, and communities. Written in Python, it synchronizes an SQL database with chain state, providing developers with a more flexible/extensible alternative to the raw steemd API.
- Python 3.6 required
- Postgres 10+ recommended
Dependencies:
- OSX:
$ brew install python3 postgresql
- Ubuntu:
$ sudo apt-get install python3 python3-pip
Installation:
$ createdb hive
$ export DATABASE_URL=postgresql://user:pass@localhost:5432/hive
$ git clone https://github.com/steemit/hivemind.git
$ cd hivemind
$ pip3 install -e .[test]
Start the indexer:
$ hive sync
$ hive status
{'db_head_block': 19930833, 'db_head_time': '2018-02-16 21:37:36', 'db_head_age': 10}
Start the server:
$ hive server
$ curl --data '{"jsonrpc":"2.0","id":0,"method":"hive.db_head_state"}' http://localhost:8080
{"jsonrpc": "2.0", "result": {"db_head_block": 19930795, "db_head_time": "2018-02-16 21:35:42", "db_head_age": 10}, "id": 0}
Run tests:
$ make test
Hivemind is deployed as a Docker container.
Here is an example command that will initialize the DB schema and start the syncing process:
docker run -d --name hivemind --env DATABASE_URL=postgresql://user:pass@hostname:5432/databasename --env STEEMD_URL=https://yoursteemnode --env SYNC_SERVICE=1 -p 8080:8080 steemit/hivemind:latest
Be sure to set DATABASE_URL
to point to your postgres database and STEEMD_URL
to point to your steemd node to sync from.
Once the database is synced, Hivemind will be available for serving requests.
To follow along the logs, use this:
docker logs -f hivemind
Environment | CLI argument | Default |
---|---|---|
LOG_LEVEL |
--log-level |
INFO |
HTTP_SERVER_PORT |
--http-server-port |
8080 |
DATABASE_URL |
--database-url |
postgresql://user:pass@localhost:5432/hive |
STEEMD_URL |
--steemd-url |
https://api.steemit.com |
MAX_BATCH |
--max-batch |
50 |
MAX_WORKERS |
--max-workers |
4 |
TRAIL_BLOCKS |
--trail-blocks |
2 |
Precedence: CLI over ENV over hive.conf. Check hive --help
for details.
- Focus on Postgres performance
- 2.5GB of memory for
hive sync
process - 250GB storage for database
Build flags
LOW_MEMORY_NODE=OFF
- need post contentCLEAR_VOTES=OFF
- need all vote dataSKIP_BY_TX=ON
- tx lookup not used
Plugins
- Required:
reputation reputation_api database_api condenser_api block_api
- Not required:
follow*
,tags*
,market_history
,account_history
,witness
For a system with 16G of memory, here's a good start:
effective_cache_size = 12GB # 50-75% of avail memory
maintenance_work_mem = 2GB
random_page_cost = 1.0 # assuming SSD storage
shared_buffers = 4GB # 25% of memory
work_mem = 512MB
synchronous_commit = off
checkpoint_completion_target = 0.9
checkpoint_timeout = 30min
max_wal_size = 4GB
The minimum viable API is to remove the requirement for the follow
and tags
plugins (now rolled into condenser_api
) from the backend node while still being able to power condenser's non-wallet features. Thus, this is the core API set:
condenser_api.get_followers
condenser_api.get_following
condenser_api.get_follow_count
condenser_api.get_content
condenser_api.get_content_replies
condenser_api.get_state
condenser_api.get_trending_tags
condenser_api.get_discussions_by_trending
condenser_api.get_discussions_by_hot
condenser_api.get_discussions_by_promoted
condenser_api.get_discussions_by_created
condenser_api.get_discussions_by_blog
condenser_api.get_discussions_by_feed
condenser_api.get_discussions_by_comments
condenser_api.get_replies_by_last_update
condenser_api.get_blog
condenser_api.get_blog_entries
condenser_api.get_discussions_by_author_before_date
Initially, the steemit.com app was powered exclusively by steemd
nodes. It was purely a client-side app without any backend other than a public and permissionless API node. As powerful as this model is, there are two issues: (a) maintaining UI-specific indices/APIs becomes expensive when tightly coupled to critical consensus nodes; and (b) frontend developers must be able to iterate quickly and access data in flexible and creative ways without writing C++.
To relieve backend and frontend pressure, non-consensus and frontend-oriented concerns can be decoupled from steemd
itself. This (a) allows the consensus node to focus on scalability and reliability, and (b) allows the frontend to maintain its own state layer, allowing for flexibility not feasible otherwise.
Specifically, the goal is to completely remove the follow
and tags
plugins, as well as get_state
from the backend node itself, and re-implement them in hive
. In doing so, we form the foundational infrastructure on which to implement communities and more.
- discussions: by blog, trending, hot, created, etc
- communities: mod roles/actions, members, feeds (in 1.5; spec)
- accounts: normalized profile data, reputation
- feeds: un/follows and un/reblogs
For anything to do with wallets, orders, escrow, keys, recovery, or account history, query SBDS or steemd.
- reactions, bookmarks
- comment on reblogs
- indexing custom profile data
- reorganize old posts (categorize, filter, hide/show)
- voting/polls (democratic or burn/send to vote)
- modlists: (e.g. spammy, abuse, badtaste)
- crowdsourced metadata
- mentions indexing
- full-text search
- follow lists
- bot tracking
- mini-games
- community bots
Ingests blocks sequentially, processing operations relevant to accounts, post creations/edits/deletes, and custom_json ops for follows, reblogs, and communities. From these we build account and post lookup tables, follow/reblog state, and communities/members data. Built exclusively from raw blocks, it becomes the ground truth for internal state. Hive does not reimplement logic required for deriving payout values, reputation, and other statistics which are much more easily attained from steemd itself in the cache layer.
Synchronizes the latest state of posts and users, allowing us to serve discussions and lists of posts with all expected information (title, preview, image, payout, votes, etc) without needing steemd
. This layer is first built once the initial core indexing is complete. Incoming blocks trigger cache updates (including recalculation of trending score) for any posts referenced in comment
or vote
operations. There is a sweep to paid out posts to ensure they are updated in full with their final state.
Performs queries against the core and cache tables, merging them into a response in such a way that the frontend will not need to perform any additional calls to steemd
itself. The initial API simply mimics steemd's condenser_api
for backwards compatibility, but will be extended to leverage new opportunities and simplify application development.
Latency vs. consistency vs. complexity
The easiest way to avoid forks is to only index up to the last irreversible block, but the delay is too much where users expect quick feedback, e.g. votes and live discussions. We can apply the following approach:
- Follow the chain as closely to
head_block
as possible - Indexer trails a few blocks behind, by no more than 6s - 9s
- If missed blocks detected, back off from
head_block
- Database constraints on block linking to detect failure asap
- If a fork is encountered between
hive_head
andsteem_head
, trivial recovery - Otherwise, pop blocks until in sync. Inconsistent state possible but rare for
TRAIL_BLOCKS > 1
. - A separate service with a greater follow distance creates periodic snapshots
$ make docs && open docs/hive/index.html
MIT