Open Ethereum Classic Pool
Switch branches/tags
Nothing to show
Clone or download
Fetching latest commit…
Cannot retrieve the latest commit at this time.
Failed to load latest commit information.

Open Source Ethereum Mining Pool

PoT version

Please consider reading docs/ first.

PoT is essentially the same as PPS but with artificially created variance.


This pool is being further developed to provide an easy to use pool for Ethereum miners. This software is functional however an optimised release of the pool is expected soon. Testing and bug submissions are welcome!

  • Support EthProxy and Ethereum Stratum protocols
  • Support for parity's job notification
  • Failover geth instances: geth high availability built in
  • Modern beautiful Ember.js frontend
  • Separate stats for workers: can highlight timed-out workers so miners can perform maintenance of rigs
  • JSON-API for stats


Building on Linux


  • go >= 1.6
  • geth or parity
  • redis-server >= 2.8.0
  • nodejs >= 4 LTS
  • nginx

I highly recommend to use Ubuntu 16.04 LTS.

First install go-ethereum.

Clone & compile:

git config --global http. true
git clone -b POT
cd open-ethereum-pool

Install redis-server.

Running Pool

./build/bin/open-ethereum-pool config.json

You can use Ubuntu upstart - check for sample config in upstart.conf.

Building Frontend

Install nodejs. I suggest using LTS version >= 4.x from or from your Linux distribution or simply install nodejs on Ubuntu Xenial 16.04.

The frontend is a single-page Ember.js application that polls the pool API to render miner stats.

cd www

Change ApiUrl: '//' in www/config/environment.js to match your domain name. Also don't forget to adjust other options.

npm install -g ember-cli@2.9.1
npm install -g bower
npm install
bower install

Configure nginx to serve API on /api subdirectory. Configure nginx to serve www/dist as static website.

Serving API using nginx

Create an upstream for API:

upstream api {

and add this setting after location /:

location /api {
    proxy_pass http://api;


You can customize the layout using built-in web server with live reload:

ember server --port 8082 --environment development

Don't use built-in web server in production.

Check out www/app/templates directory and edit these templates in order to customise the frontend.


Configuration is actually simple, just read it twice and think twice before changing defaults.

Don't copy config directly from this manual. Use the example config from the package, otherwise you will get errors on start because of JSON comments.

  // Set to the number of CPU cores of your server
  "threads": 2,
  // Prefix for keys in redis store
  "coin": "eth",
  // Give unique name to each instance
  "name": "main",

  "proxy": {
    "enabled": true,

    // New work notifications endpoint (see --notify-work option of parity client)
    // Note that this endpoint should be either local or protected by firewall. Only trusted clients should be allowed to send work notifications.
    "listen": "",

    // Allow only this header and body size of HTTP request from miners
    "limitHeadersSize": 1024,
    "limitBodySize": 256,

    /* Set to true if you are behind CloudFlare (not recommended) or behind http-reverse
      proxy to enable IP detection from X-Forwarded-For header.
      Advanced users only. It's tricky to make it right and secure.
    "behindReverseProxy": false,

    // Stratum mining endpoint
    "stratum": {
      "enabled": true,
      // Bind stratum mining socket to this IP:PORT
      "listen": "",
      "protocol": "Stratum-Proxy",
      "timeout": "120s",
      "maxConn": 8192

    // Try to get new job from geth in this interval
    "blockRefreshInterval": "120ms",
    "stateUpdateInterval": "3s",
    // Require this share difficulty from miners
    "difficulty": 2000000000,
    // PPS fee applied to each share submitted
    "miningFee": 1.5,
    // Mediocre shares are getting just about 20% of average reward
    "potA": 0.8,
    // Cap share difficulty to x1.5 of network difficulty to prevent ridicuosly difficult share from bankrupting the pool
    "potCap": 1.5,

    /* Reply error to miner instead of job if redis is unavailable.
      Should save electricity to miners if pool is sick and they didn't set up failovers.
    "healthCheck": true,
    // Mark pool sick after this number of redis failures.
    "maxFails": 100,
    // TTL for workers stats, usually should be equal to large hashrate window from API section
    "hashrateExpiration": "3h",

    "policy": {
      "workers": 8,
      "resetInterval": "60m",
      "refreshInterval": "1m",

      "banning": {
        "enabled": false,
        /* Name of ipset for banning.
        Check documentation.
        "ipset": "blacklist",
        // Remove ban after this amount of time
        "timeout": 1800,
        // Percent of invalid shares from all shares to ban miner
        "invalidPercent": 30,
        // Check after after miner submitted this number of shares
        "checkThreshold": 30,
        // Bad miner after this number of malformed requests
        "malformedLimit": 5
      // Connection rate limit
      "limits": {
        "enabled": false,
        // Number of initial connections
        "limit": 30,
        "grace": "5m",
        // Increase allowed number of connections on each valid share
        "limitJump": 10

  // Provides JSON data for frontend which is static website
  "api": {
    "enabled": true,
    // Bind API endpoint to this IP:PORT
    "listen": "",
    // Collect miners stats (hashrate, ...) in this interval
    "statsCollectInterval": "5s",
    // Purge stale stats interval
    "purgeInterval": "10m",
    // Fast hashrate estimation window for each miner from it's shares
    "hashrateWindow": "30m",
    // Long and precise hashrate from shares, 3h is cool, keep it
    "hashrateLargeWindow": "3h",
    // Collect stats for shares/diff ratio for this number of blocks
    "luckWindow": [64, 128, 256],
    // Max number of payments to display in frontend
    "payments": 50,
    // Max numbers of shifts to display in frontend
    "longShifts": 30,
    "shortShifts": 24,

    /* If you are running API node on a different server where this module
      is reading data from redis writeable slave, you must run an api instance with this option enabled in order to purge hashrate stats from main redis node.
      Only redis writeable slave will work properly if you are distributing using redis slaves.
      Very advanced. Usually all modules should share same redis instance.
    "purgeOnly": false

  // Check health of each geth node in this interval
  "upstreamCheckInterval": "5s",

  /* List of geth nodes to poll for new jobs. Pool will try to get work from
    first alive one and check in background for failed to back up.
    Current block template of the pool is always cached in RAM indeed.
  "upstream": [
      "name": "main",
      "url": "",
      "timeout": "10s"
      "name": "backup",
      "url": "",
      "timeout": "10s"

  // This is standard redis connection options
  "redis": {
    // Where your redis instance is listening for commands
    "endpoint": "",
    "poolSize": 10,
    "database": 0,
    "password": ""

  // Pay out miners using this module
  "payouts": {
    "enabled": false,
    // Require minimum number of peers on node
    "requirePeers": 25,
    // Run payouts in this interval
    "interval": "12h",
    // Geth instance node rpc endpoint for payouts processing
    "daemon": "",
    // Rise error if can't reach geth in this amount of time
    "timeout": "10s",
    // Address with pool balance
    "address": "0x0",
    // Let geth to determine gas and gasPrice
    "autoGas": true,
    // Gas amount and price for payout tx (advanced users only)
    "gas": "21000",
    "gasPrice": "50000000000",
    // Normally, send payment only if miner's balance is >= 0.5 Ether
    "threshold": 500000000,
    // If user was inactive for longer than week, send payment if his balance is >= 0.05 Ether
    "threshold": 50000000,
    // Perform BGSAVE on Redis after successful payouts session
    "bgsave": false
  // Maintain daily shifts of per-user statistics
  "shifts": {
    "enabled": false,

    // Create long shifts in these intervals
    "longInterval": "24h",
    "shortInterval" : "1h",

    // Keep shifts data during these intervals
    "keepLong" : "30d",
    "keepShort": "24h"

If you are distributing your pool deployment to several servers or processes, create several configs and disable unneeded modules on each server. (Advanced users)

I recommend this deployment strategy:

  • Mining instance - 1x (it depends, you can run one node for EU, one for US, one for Asia)
  • Payouts instance - 1x (strict!)
  • Shifting instance - 1x (strict!)
  • API instance - 1x


  • Payouts are sequential, 1st tx go, 2nd waiting for 1st to confirm and so on. You can disable that in code. Carefully read docs/
  • Also, keep in mind that payouts will halt in case of backend or node RPC errors. In that case check everything and restart.
  • You must restart module if you see errors with the word suspended.
  • Don't run either payouts or shifting module as part of mining node. Create separate configs for them, launch independently and make sure you have a single instance of each module running.

Ethereum Implementations

This pool is tested to work with Parity. It is recommended to use this Ethereum client due to availability of --notify-work feature which allows you to avoid block polling.


Original code made by sammy007. Licensed under GPLv3.