Start a search engine (Docker)
Start a smart contract search engine for your own DApp
This documentation details how you can start, and host, your own smart contract search engine using Docker. If you would like to build from scratch from a fresh Ubuntu install, please refer to this document.
Prerequisite
We start from a fresh install of Ubuntu 18.04. You should first follow the instructions here to install Docker.
Next install the Python pip and AWS CLI utilities as follows. The AWS CLI is required to access AWS ElasticSearch services.
$ sudo apt update
$ sudo apt install python-pip
$ sudo apt install awscli
ElasticSearch
We use the AWS ElasticSearch services to run the search engine. You should create a new ES cluster here. For now, a single machine development cluster would suffice. In the Access Policy section, please select IAM users. You will need an IAM user already set up to access AWS ES services. Here is an example.
arn:aws:iam::522901590065:user/secondstatesearch
Once the ElasticSearch service is up and running, you should have an ES endpoint like the following.
search-smart-contract-search-engine-3paomceha6u4qzchkmbgsjdcqa.us-east-1.es.amazonaws.com
Docker
Now, go back to the Ubuntu 18.04 machine.
AWS credentials
Configure AWS CLI to access the ElasticSearch engine.
$ aws configure
It requires four pieces of information. The access keys are found in the IAM user console for the user you configured to access the ElasticSearch engine you just created.
AWS Access Key ID [None]: [IAM user console]
AWS Secret Access Key [None]: [IAM user console]
Default region name [None]: [Region for ES instance. eg us-east-1]
Default output format [None]: json
After configuration, AWS config and credentials are placed in ~/.aws/
.
Configure search engine
Next, get the source code for the search engine.
$ git clone https://github.com/second-state/smart-contract-search-engine.git
$ cd smart-contract-search-engine
Fill in the following configuration options.
ServerName
in apache configconfig/site.conf
. This could be your public IP address for now.blockchain
,elasticsearch
, and the initial ABI configs inpython/config.ini
.publicIp
injs/secondStateJS.js
. This could be your IP address for now.Check here for details about configurations.
Build Docker image
$ docker build -f docker/Dockerfile -t search-engine .
Run Docker container
$ docker run -d -it --rm -p 80:80 -v $HOME/.aws:/root/.aws search-engine
Now you can visit http://<your_host>
to check your smart contract search engine. Be patient, as it may take hours before the results show up on that page.
Upload more ABIs
Your search engine is started with a single ABI to index from the config.ini
file. You can add more ABIs to the index by executing the following script from inside the Docker instance.
You can find the container_id
for your docker instance on your host OS, by running
$ docker container ls
Next, logging into your docker container using the container_id
$ docker exec -it container_id bash
Once logged, in the /app
directory, create a file upload_abi.py
like the following.
import re
import json
import time
import requests
from harvest import Harvest
harvester = Harvest()
abiUrl1 = "http://A_raw_text_file_which_contains_only_an_abi's_text"
abiData1 = requests.get(abiUrl1).content
abiData1JSON = json.loads(abiData1)
theDeterministicHash1 = harvester.shaAnAbi(abiData1JSON)
cleanedAndOrderedAbiText1 = harvester.cleanAndConvertAbiToText(abiData1JSON)
data1 = {}
data1['indexInProgress'] = "false"
data1['epochOfLastUpdate'] = int(time.time())
data1['abi'] = cleanedAndOrderedAbiText1
harvester.es.index(index=harvester.abiIndex, id=theDeterministicHash1, body=data1)
Then run
$ python3.6 upload_abi.py
Also once all of this is done, please just exit docker and give it a reboot.
$ docker restart container_id
Last updated
Was this helpful?