a175c5c9cc | ||
---|---|---|
consul-discovery | ||
elk_netflow | ||
haproxy_web | ||
hyperledger | ||
mongo-elasticsearch | ||
mongo_cluster | ||
mongo_webui | ||
nginx_auth | ||
packetbeat_ek | ||
registry_mirror | ||
spark_cluster | ||
.gitignore | ||
README.md |
README.md
Docker Compose Files
Some typical docker compose examples.
Install Docker and Docker Compose
Take ubuntu for example
$ curl -sSL https://get.docker.com/ | sh
$ sudo pip install docker-compose
Docker-compose Usage
See Docker Compose Documentation.
Examples files
consul-discovery
Using consul to make a service-discoverable architecture.
elk_netflow
Elk cluster, with netflow support.
docker-compose scale es=3
haproxy_web
A simple haproxy and web applications cluster.
hyperledger
Quickly bootup a hyperledger cluster with several validator nodes, without vagrant or any manual configuration. By default, the cluster enables PBFT as the consensus.
Note, currently you should manually create an openblockchain/baseimage:latest
first. The
easiest way to do so is:
$ docker pull yeasy/hyperledger:latest
$ docker tag yeasy/hyperledger:latest openblockchain/baseimage:latest
$ docker pull yeasy/hyperledger-peer:pbft
$ docker pull yeasy/hyperledger-membersrvc:latest
Then you can start a 4 nodes hyperledger cluster with
$ docker-compose up
After the cluster is synced, you can validate by deploying, invoking or querying chaincode from the container or from the host. See [hyperledger-peer](https://github .com/yeasy/docker-hyperledger-peer) if you've not familiar on that.
This refers the example from the hyperledger project.
mongo_cluster
Start 3 mongo instance to make a replica set.
mongo-elasticsearch
Start mongo (as cluster) and elasticsearch, use a mongo-connector to sync the data from mongo to elasticsearch.
mongo_webui
Start 1 mongo instance and a mongo-express web tool to watch it.
The mongo instance will store data into local /opt/data/mongo_home.
The web UI will listen on local 8081 port.
nginx_auth
Use nginx as a proxy with authentication for backend application.
packetbeat_ek
Demo the packetbeat, elasticsearch and kibana.
Some kibana dashboard config files are included.
To import them, after all containers startup, go inside the kibana container, and run
$ cd /kibana/beats-dashboards-1.0.1 && ./load.sh http://elasticsearch:9200
registry_mirror
docker registry mirror, with redis as the backend cache.
spark_cluster
Spark cluster with master and worker nodes.
docker-compose scale worker=2
Try submitting a test pi application using the spark-submit command.
/urs/local/spark/bin/spark-submit --master spark://master:7077 --class org.apache.spark.examples.SparkPi /usr/local/spark/lib/spark-examples-1.4.0-hadoop2.6.0.jar 1000