/* * Copyright 2014-2015 Con Kolivas * * This program is free software; you can redistribute it and/or modify it * under the terms of the GNU General Public License as published by the Free * Software Foundation; either version 3 of the License, or (at your option) * any later version. See COPYING for more details. */ #include "config.h" #include #include #include #include #include #include #include #include #include #include #include "ckpool.h" #include "libckpool.h" #include "bitcoin.h" #include "sha2.h" #include "stratifier.h" #include "uthash.h" #include "utlist.h" /* Consistent across all pool instances */ static const char *workpadding = "000000800000000000000000000000000000000000000000000000000000000000000000000000000000000080020000"; static const char *scriptsig_header = "01000000010000000000000000000000000000000000000000000000000000000000000000ffffffff"; static uchar scriptsig_header_bin[41]; /* Add unaccounted shares when they arrive, remove them with each update of * rolling stats. */ struct pool_stats { tv_t start_time; ts_t last_update; int workers; int users; int disconnected; /* Absolute shares stats */ int64_t unaccounted_shares; int64_t accounted_shares; /* Cycle of 32 to determine which users to dump stats on */ uint8_t userstats_cycle; /* Shares per second for 1/5/15/60 minute rolling averages */ double sps1; double sps5; double sps15; double sps60; /* Diff shares stats */ int64_t unaccounted_diff_shares; int64_t accounted_diff_shares; int64_t unaccounted_rejects; int64_t accounted_rejects; /* Diff shares per second for 1/5/15... minute rolling averages */ double dsps1; double dsps5; double dsps15; double dsps60; double dsps360; double dsps1440; double dsps10080; }; typedef struct pool_stats pool_stats_t; struct workbase { /* Hash table data */ UT_hash_handle hh; int64_t id; char idstring[20]; ts_t gentime; /* GBT/shared variables */ char target[68]; double diff; double network_diff; uint32_t version; uint32_t curtime; char prevhash[68]; char ntime[12]; uint32_t ntime32; char bbversion[12]; char nbit[12]; uint64_t coinbasevalue; int height; char *flags; int transactions; char *txn_data; char *txn_hashes; int merkles; char merklehash[16][68]; char merklebin[16][32]; json_t *merkle_array; /* Template variables, lengths are binary lengths! */ char *coinb1; // coinbase1 uchar *coinb1bin; int coinb1len; // length of above char enonce1const[32]; // extranonce1 section that is constant uchar enonce1constbin[16]; int enonce1constlen; // length of above - usually zero unless proxying int enonce1varlen; // length of unique extranonce1 string for each worker - usually 8 int enonce2varlen; // length of space left for extranonce2 - usually 8 unless proxying char *coinb2; // coinbase2 uchar *coinb2bin; int coinb2len; // length of above /* Cached header binary */ char headerbin[112]; char *logdir; ckpool_t *ckp; bool proxy; /* This workbase is proxied work */ }; typedef struct workbase workbase_t; struct json_params { json_t *method; json_t *params; json_t *id_val; int64_t client_id; char address[INET6_ADDRSTRLEN]; }; typedef struct json_params json_params_t; /* Stratum json messages with their associated client id */ struct smsg { json_t *json_msg; int64_t client_id; char address[INET6_ADDRSTRLEN]; }; typedef struct smsg smsg_t; struct user_instance; struct worker_instance; struct stratum_instance; typedef struct user_instance user_instance_t; typedef struct worker_instance worker_instance_t; typedef struct stratum_instance stratum_instance_t; struct user_instance { UT_hash_handle hh; char username[128]; int64_t id; char *secondaryuserid; bool btcaddress; /* A linked list of all connected instances of this user */ stratum_instance_t *clients; /* A linked list of all connected workers of this user */ worker_instance_t *worker_instances; int workers; double best_diff; /* Best share found by this user */ double dsps1; /* Diff shares per second, 1 minute rolling average */ double dsps5; /* ... 5 minute ... */ double dsps60;/* etc */ double dsps1440; double dsps10080; tv_t last_share; tv_t last_update; bool authorised; /* Has this username ever been authorised? */ time_t auth_time; time_t failed_authtime; /* Last time this username failed to authorise */ int auth_backoff; /* How long to reject any auth attempts since last failure */ }; /* Combined data from workers with the same workername */ struct worker_instance { user_instance_t *user_instance; char *workername; worker_instance_t *next; worker_instance_t *prev; double dsps1; double dsps5; double dsps60; double dsps1440; tv_t last_share; tv_t last_update; time_t start_time; double best_diff; /* Best share found by this worker */ int mindiff; /* User chosen mindiff */ bool idle; bool notified_idle; }; typedef struct stratifier_data sdata_t; typedef struct proxy_base proxy_t; /* Per client stratum instance == workers */ struct stratum_instance { UT_hash_handle hh; int64_t id; stratum_instance_t *next; stratum_instance_t *prev; /* Reference count for when this instance is used outside of the * instance_lock */ int ref; char enonce1[32]; uchar enonce1bin[16]; char enonce1var[12]; uint64_t enonce1_64; int64_t diff; /* Current diff */ int64_t old_diff; /* Previous diff */ int64_t diff_change_job_id; /* Last job_id we changed diff */ double dsps1; /* Diff shares per second, 1 minute rolling average */ double dsps5; /* ... 5 minute ... */ double dsps60;/* etc */ double dsps1440; double dsps10080; tv_t ldc; /* Last diff change */ int ssdc; /* Shares since diff change */ tv_t first_share; tv_t last_share; time_t first_invalid; /* Time of first invalid in run of non stale rejects */ time_t start_time; char address[INET6_ADDRSTRLEN]; bool subscribed; bool authorising; /* In progress, protected by instance_lock */ bool authorised; bool dropped; bool idle; int reject; /* Indicator that this client is having a run of rejects * or other problem and should be dropped lazily if * this is set to 2 */ bool reconnect; /* Do we need to send this client a reconnect message */ time_t reconnect_request; /* The time we sent a reconnect message */ user_instance_t *user_instance; worker_instance_t *worker_instance; char *useragent; char *workername; int64_t user_id; int server; /* Which server is this instance bound to */ ckpool_t *ckp; time_t last_txns; /* Last time this worker requested txn hashes */ time_t disconnected_time; /* Time this instance disconnected */ int64_t suggest_diff; /* Stratum client suggested diff */ double best_diff; /* Best share found by this instance */ sdata_t *sdata; /* Which sdata this client is bound to */ proxy_t *proxy; /* Proxy this is bound to in proxy mode */ int64_t proxyid; /* Which proxy id */ int subproxyid; /* Which subproxy */ }; struct share { UT_hash_handle hh; uchar hash[32]; int64_t workbase_id; }; typedef struct share share_t; struct proxy_base { UT_hash_handle hh; UT_hash_handle sh; /* For subproxy hashlist */ proxy_t *next; /* For retired subproxies */ proxy_t *prev; int64_t id; int low_id; int subid; double diff; char enonce1[32]; uchar enonce1bin[16]; int enonce1constlen; int enonce1varlen; int nonce2len; int enonce2varlen; bool subscribed; bool notified; int64_t clients; /* Incrementing client count */ int64_t max_clients; /* Maximum number of clients per subproxy */ int64_t bound_clients; /* Currently actively bound clients */ int64_t combined_clients; /* Total clients of all subproxies of a parent proxy */ int64_t headroom; /* Temporary variable when calculating how many more clients can bind */ proxy_t *parent; /* Parent proxy of each subproxy */ proxy_t *subproxies; /* Hashlist of subproxies sorted by subid */ sdata_t *sdata; /* Unique stratifer data for each subproxy */ bool dead; }; struct stratifier_data { ckpool_t *ckp; char pubkeytxnbin[25]; char donkeytxnbin[25]; pool_stats_t stats; /* Protects changes to pool stats */ mutex_t stats_lock; /* Serialises sends/receives to ckdb if possible */ mutex_t ckdb_lock; bool ckdb_offline; bool verbose; uint64_t enonce1_64; /* For protecting the hashtable data */ cklock_t workbase_lock; /* For the hashtable of all workbases */ workbase_t *workbases; workbase_t *current_workbase; int workbases_generated; int64_t workbase_id; int64_t blockchange_id; char lasthash[68]; char lastswaphash[68]; ckmsgq_t *ssends; // Stratum sends ckmsgq_t *srecvs; // Stratum receives ckmsgq_t *ckdbq; // ckdb ckmsgq_t *sshareq; // Stratum share sends ckmsgq_t *sauthq; // Stratum authorisations ckmsgq_t *stxnq; // Transaction requests int64_t user_instance_id; /* Stratum_instances hashlist is stored by id, whereas disconnected_instances * is sorted by enonce1_64. */ stratum_instance_t *stratum_instances; stratum_instance_t *disconnected_instances; int stratum_generated; int disconnected_generated; user_instance_t *user_instances; /* Protects both stratum and user instances */ cklock_t instance_lock; share_t *shares; mutex_t share_lock; int64_t shares_generated; /* Linked list of block solves, added to during submission, removed on * accept/reject. It is likely we only ever have one solve on here but * you never know... */ mutex_t block_lock; ckmsg_t *block_solves; /* Generator message priority */ int gen_priority; int proxy_count; /* Total proxies generated (not necessarily still alive) */ proxy_t *proxy; /* Current proxy in use */ proxy_t *proxies; /* Hashlist of all proxies */ mutex_t proxy_lock; /* Protects all proxy data */ proxy_t *subproxy; /* Which subproxy this sdata belongs to in proxy mode */ }; typedef struct json_entry json_entry_t; struct json_entry { json_entry_t *next; json_entry_t *prev; json_t *val; }; /* Priority levels for generator messages */ #define GEN_LAX 0 #define GEN_NORMAL 1 #define GEN_PRIORITY 2 #define ID_AUTH 0 #define ID_WORKINFO 1 #define ID_AGEWORKINFO 2 #define ID_SHARES 3 #define ID_SHAREERR 4 #define ID_POOLSTATS 5 #define ID_WORKERSTATS 6 #define ID_BLOCK 7 #define ID_ADDRAUTH 8 #define ID_HEARTBEAT 9 static const char *ckdb_ids[] = { "authorise", "workinfo", "ageworkinfo", "shares", "shareerror", "poolstats", "workerstats", "block", "addrauth", "heartbeat" }; static void generate_coinbase(const ckpool_t *ckp, workbase_t *wb) { uint64_t *u64, g64, d64 = 0; sdata_t *sdata = ckp->data; char header[228]; int len, ofs = 0; ts_t now; /* Set fixed length coinb1 arrays to be more than enough */ wb->coinb1 = ckzalloc(256); wb->coinb1bin = ckzalloc(128); /* Strings in wb should have been zero memset prior. Generate binary * templates first, then convert to hex */ memcpy(wb->coinb1bin, scriptsig_header_bin, 41); ofs += 41; // Fixed header length; ofs++; // Script length is filled in at the end @wb->coinb1bin[41]; /* Put block height at start of template */ len = ser_number(wb->coinb1bin + ofs, wb->height); ofs += len; /* Followed by flag */ len = strlen(wb->flags) / 2; wb->coinb1bin[ofs++] = len; hex2bin(wb->coinb1bin + ofs, wb->flags, len); ofs += len; /* Followed by timestamp */ ts_realtime(&now); len = ser_number(wb->coinb1bin + ofs, now.tv_sec); ofs += len; /* Followed by our unique randomiser based on the nsec timestamp */ len = ser_number(wb->coinb1bin + ofs, now.tv_nsec); ofs += len; wb->enonce1varlen = ckp->nonce1length; wb->enonce2varlen = ckp->nonce2length; wb->coinb1bin[ofs++] = wb->enonce1varlen + wb->enonce2varlen; wb->coinb1len = ofs; len = wb->coinb1len - 41; len += wb->enonce1varlen; len += wb->enonce2varlen; wb->coinb2bin = ckzalloc(256); memcpy(wb->coinb2bin, "\x0a\x63\x6b\x70\x6f\x6f\x6c", 7); wb->coinb2len = 7; if (ckp->btcsig) { int siglen = strlen(ckp->btcsig); LOGDEBUG("Len %d sig %s", siglen, ckp->btcsig); if (siglen) { wb->coinb2bin[wb->coinb2len++] = siglen; memcpy(wb->coinb2bin + wb->coinb2len, ckp->btcsig, siglen); wb->coinb2len += siglen; } } len += wb->coinb2len; wb->coinb1bin[41] = len - 1; /* Set the length now */ __bin2hex(wb->coinb1, wb->coinb1bin, wb->coinb1len); LOGDEBUG("Coinb1: %s", wb->coinb1); /* Coinbase 1 complete */ memcpy(wb->coinb2bin + wb->coinb2len, "\xff\xff\xff\xff", 4); wb->coinb2len += 4; // Generation value g64 = wb->coinbasevalue; if (ckp->donvalid) { d64 = g64 / 200; // 0.5% donation g64 -= d64; // To guarantee integers add up to the original coinbasevalue wb->coinb2bin[wb->coinb2len++] = 2; // 2 transactions } else wb->coinb2bin[wb->coinb2len++] = 1; // 2 transactions u64 = (uint64_t *)&wb->coinb2bin[wb->coinb2len]; *u64 = htole64(g64); wb->coinb2len += 8; wb->coinb2bin[wb->coinb2len++] = 25; memcpy(wb->coinb2bin + wb->coinb2len, sdata->pubkeytxnbin, 25); wb->coinb2len += 25; if (ckp->donvalid) { u64 = (uint64_t *)&wb->coinb2bin[wb->coinb2len]; *u64 = htole64(d64); wb->coinb2len += 8; wb->coinb2bin[wb->coinb2len++] = 25; memcpy(wb->coinb2bin + wb->coinb2len, sdata->donkeytxnbin, 25); wb->coinb2len += 25; } wb->coinb2len += 4; // Blank lock wb->coinb2 = bin2hex(wb->coinb2bin, wb->coinb2len); LOGDEBUG("Coinb2: %s", wb->coinb2); /* Coinbase 2 complete */ snprintf(header, 225, "%s%s%s%s%s%s%s", wb->bbversion, wb->prevhash, "0000000000000000000000000000000000000000000000000000000000000000", wb->ntime, wb->nbit, "00000000", /* nonce */ workpadding); LOGDEBUG("Header: %s", header); hex2bin(wb->headerbin, header, 112); } static void stratum_broadcast_update(sdata_t *sdata, const workbase_t *wb, bool clean); static void clear_workbase(workbase_t *wb) { free(wb->flags); free(wb->txn_data); free(wb->txn_hashes); free(wb->logdir); free(wb->coinb1bin); free(wb->coinb1); free(wb->coinb2bin); free(wb->coinb2); json_decref(wb->merkle_array); free(wb); } /* Remove all shares with a workbase id less than wb_id for block changes */ static void purge_share_hashtable(sdata_t *sdata, const int64_t wb_id) { share_t *share, *tmp; int purged = 0; mutex_lock(&sdata->share_lock); HASH_ITER(hh, sdata->shares, share, tmp) { if (share->workbase_id < wb_id) { HASH_DEL(sdata->shares, share); dealloc(share); purged++; } } mutex_unlock(&sdata->share_lock); if (purged) LOGINFO("Cleared %d shares from share hashtable", purged); } /* Remove all shares with a workbase id == wb_id being discarded */ static void age_share_hashtable(sdata_t *sdata, const int64_t wb_id) { share_t *share, *tmp; int aged = 0; mutex_lock(&sdata->share_lock); HASH_ITER(hh, sdata->shares, share, tmp) { if (share->workbase_id == wb_id) { HASH_DEL(sdata->shares, share); dealloc(share); aged++; } } mutex_unlock(&sdata->share_lock); if (aged) LOGINFO("Aged %d shares from share hashtable", aged); } static char *status_chars = "|/-\\"; /* Absorbs the json and generates a ckdb json message, logs it to the ckdb * log and returns the malloced message. */ static char *ckdb_msg(ckpool_t *ckp, json_t *val, const int idtype) { char *json_msg = json_dumps(val, JSON_COMPACT); char logname[512]; char *ret = NULL; if (unlikely(!json_msg)) goto out; ASPRINTF(&ret, "%s.id.json=%s", ckdb_ids[idtype], json_msg); free(json_msg); out: json_decref(val); snprintf(logname, 511, "%s%s", ckp->logdir, ckp->ckdb_name); rotating_log(logname, ret); return ret; } static void _ckdbq_add(ckpool_t *ckp, const int idtype, json_t *val, const char *file, const char *func, const int line) { static time_t time_counter; sdata_t *sdata = ckp->data; static int counter = 0; char *json_msg; time_t now_t; char ch; if (unlikely(!val)) { LOGWARNING("Invalid json sent to ckdbq_add from %s %s:%d", file, func, line); return; } now_t = time(NULL); if (now_t != time_counter) { /* Rate limit to 1 update per second */ time_counter = now_t; ch = status_chars[(counter++) & 0x3]; fprintf(stdout, "%c\r", ch); fflush(stdout); } if (CKP_STANDALONE(ckp)) return json_decref(val); json_msg = ckdb_msg(ckp, val, idtype); if (unlikely(!json_msg)) { LOGWARNING("Failed to dump json from %s %s:%d", file, func, line); return; } ckmsgq_add(sdata->ckdbq, json_msg); } #define ckdbq_add(ckp, idtype, val) _ckdbq_add(ckp, idtype, val, __FILE__, __func__, __LINE__) static void send_workinfo(ckpool_t *ckp, const workbase_t *wb) { char cdfield[64]; json_t *val; sprintf(cdfield, "%lu,%lu", wb->gentime.tv_sec, wb->gentime.tv_nsec); JSON_CPACK(val, "{sI,ss,ss,ss,ss,ss,ss,ss,ss,sI,so,ss,ss,ss,ss}", "workinfoid", wb->id, "poolinstance", ckp->name, "transactiontree", wb->txn_hashes, "prevhash", wb->prevhash, "coinbase1", wb->coinb1, "coinbase2", wb->coinb2, "version", wb->bbversion, "ntime", wb->ntime, "bits", wb->nbit, "reward", wb->coinbasevalue, "merklehash", json_deep_copy(wb->merkle_array), "createdate", cdfield, "createby", "code", "createcode", __func__, "createinet", ckp->serverurl[0]); ckdbq_add(ckp, ID_WORKINFO, val); } static void send_ageworkinfo(ckpool_t *ckp, const int64_t id) { char cdfield[64]; ts_t ts_now; json_t *val; ts_realtime(&ts_now); sprintf(cdfield, "%lu,%lu", ts_now.tv_sec, ts_now.tv_nsec); JSON_CPACK(val, "{sI,ss,ss,ss,ss,ss}", "workinfoid", id, "poolinstance", ckp->name, "createdate", cdfield, "createby", "code", "createcode", __func__, "createinet", ckp->serverurl[0]); ckdbq_add(ckp, ID_AGEWORKINFO, val); } /* Add a new workbase to the table of workbases. Sdata is the global data in * pool mode but unique to each subproxy in proxy mode */ static void add_base(ckpool_t *ckp, sdata_t *sdata, workbase_t *wb, bool *new_block) { workbase_t *tmp, *tmpa, *aged = NULL; sdata_t *ckp_sdata = ckp->data; int len, ret; ts_realtime(&wb->gentime); wb->network_diff = diff_from_nbits(wb->headerbin + 72); len = strlen(ckp->logdir) + 8 + 1 + 16 + 1; wb->logdir = ckzalloc(len); /* In proxy mode, the wb->id is received in the notify update and * we set workbase_id from it. In server mode the stratifier is * setting the workbase_id */ ck_wlock(&sdata->workbase_lock); ckp_sdata->workbases_generated++; if (!ckp->proxy) wb->id = sdata->workbase_id++; else sdata->workbase_id = wb->id; if (strncmp(wb->prevhash, sdata->lasthash, 64)) { char bin[32], swap[32]; *new_block = true; memcpy(sdata->lasthash, wb->prevhash, 65); hex2bin(bin, sdata->lasthash, 32); swap_256(swap, bin); __bin2hex(sdata->lastswaphash, swap, 32); sdata->blockchange_id = wb->id; } if (*new_block && ckp->logshares) { sprintf(wb->logdir, "%s%08x/", ckp->logdir, wb->height); ret = mkdir(wb->logdir, 0750); if (unlikely(ret && errno != EEXIST)) LOGERR("Failed to create log directory %s", wb->logdir); } sprintf(wb->idstring, "%016lx", wb->id); if (ckp->logshares) sprintf(wb->logdir, "%s%08x/%s", ckp->logdir, wb->height, wb->idstring); /* Is this long enough to ensure we don't dereference a workbase * immediately? Should be unless clock changes 10 minutes so we use * ts_realtime */ HASH_ITER(hh, sdata->workbases, tmp, tmpa) { if (HASH_COUNT(sdata->workbases) < 3) break; if (wb == tmp) continue; /* Age old workbases older than 10 minutes old */ if (tmp->gentime.tv_sec < wb->gentime.tv_sec - 600) { HASH_DEL(sdata->workbases, tmp); aged = tmp; break; } } HASH_ADD_I64(sdata->workbases, id, wb); sdata->current_workbase = wb; ck_wunlock(&sdata->workbase_lock); if (*new_block) purge_share_hashtable(sdata, wb->id); send_workinfo(ckp, wb); /* Send the aged work message to ckdb once we have dropped the workbase lock * to prevent taking recursive locks */ if (aged) { send_ageworkinfo(ckp, aged->id); age_share_hashtable(sdata, aged->id); clear_workbase(aged); } } /* Mandatory send_recv to the generator which sets the message priority if this * message is higher priority. Races galore on gen_priority mean this might * read the wrong priority but occasional wrong values are harmless. */ static char *__send_recv_generator(ckpool_t *ckp, const char *msg, const int prio) { sdata_t *sdata = ckp->data; char *buf = NULL; bool set; if (prio > sdata->gen_priority) { sdata->gen_priority = prio; set = true; } else set = false; buf = send_recv_proc(ckp->generator, msg); if (unlikely(!buf)) buf = strdup("failed"); if (set) sdata->gen_priority = 0; return buf; } /* Conditionally send_recv a message only if it's equal or higher priority than * any currently being serviced. NULL is returned if the request is not * processed for priority reasons, "failed" for an actual failure. */ static char *send_recv_generator(ckpool_t *ckp, const char *msg, const int prio) { sdata_t *sdata = ckp->data; char *buf = NULL; if (prio >= sdata->gen_priority) buf = __send_recv_generator(ckp, msg, prio); return buf; } static void send_generator(const ckpool_t *ckp, const char *msg, const int prio) { sdata_t *sdata = ckp->data; bool set; if (prio > sdata->gen_priority) { sdata->gen_priority = prio; set = true; } else set = false; send_proc(ckp->generator, msg); if (set) sdata->gen_priority = 0; } struct update_req { pthread_t *pth; ckpool_t *ckp; int prio; }; static void broadcast_ping(sdata_t *sdata); /* This function assumes it will only receive a valid json gbt base template * since checking should have been done earlier, and creates the base template * for generating work templates. */ static void *do_update(void *arg) { struct update_req *ur = (struct update_req *)arg; ckpool_t *ckp = ur->ckp; sdata_t *sdata = ckp->data; bool new_block = false; int prio = ur->prio; bool ret = false; workbase_t *wb; json_t *val; char *buf; pthread_detach(pthread_self()); rename_proc("updater"); buf = send_recv_generator(ckp, "getbase", prio); if (unlikely(!buf)) { LOGNOTICE("Get base in update_base delayed due to higher priority request"); goto out; } if (unlikely(cmdmatch(buf, "failed"))) { LOGWARNING("Generator returned failure in update_base"); goto out; } wb = ckzalloc(sizeof(workbase_t)); wb->ckp = ckp; val = json_loads(buf, 0, NULL); json_strcpy(wb->target, val, "target"); json_dblcpy(&wb->diff, val, "diff"); json_uintcpy(&wb->version, val, "version"); json_uintcpy(&wb->curtime, val, "curtime"); json_strcpy(wb->prevhash, val, "prevhash"); json_strcpy(wb->ntime, val, "ntime"); sscanf(wb->ntime, "%x", &wb->ntime32); json_strcpy(wb->bbversion, val, "bbversion"); json_strcpy(wb->nbit, val, "nbit"); json_uint64cpy(&wb->coinbasevalue, val, "coinbasevalue"); json_intcpy(&wb->height, val, "height"); json_strdup(&wb->flags, val, "flags"); json_intcpy(&wb->transactions, val, "transactions"); if (wb->transactions) { json_strdup(&wb->txn_data, val, "txn_data"); json_strdup(&wb->txn_hashes, val, "txn_hashes"); } else wb->txn_hashes = ckzalloc(1); json_intcpy(&wb->merkles, val, "merkles"); wb->merkle_array = json_array(); if (wb->merkles) { json_t *arr; int i; arr = json_object_get(val, "merklehash"); for (i = 0; i < wb->merkles; i++) { strcpy(&wb->merklehash[i][0], json_string_value(json_array_get(arr, i))); hex2bin(&wb->merklebin[i][0], &wb->merklehash[i][0], 32); json_array_append_new(wb->merkle_array, json_string(&wb->merklehash[i][0])); } } json_decref(val); generate_coinbase(ckp, wb); add_base(ckp, sdata, wb, &new_block); if (new_block) LOGNOTICE("Block hash changed to %s", sdata->lastswaphash); stratum_broadcast_update(sdata, wb, new_block); ret = true; LOGINFO("Broadcast updated stratum base"); out: /* Send a ping to miners if we fail to get a base to keep them * connected while bitcoind recovers(?) */ if (unlikely(!ret)) { LOGINFO("Broadcast ping due to failed stratum base update"); broadcast_ping(sdata); } dealloc(buf); free(ur->pth); free(ur); return NULL; } static void update_base(ckpool_t *ckp, const int prio) { struct update_req *ur = ckalloc(sizeof(struct update_req)); pthread_t *pth = ckalloc(sizeof(pthread_t)); ur->pth = pth; ur->ckp = ckp; ur->prio = prio; create_pthread(pth, do_update, ur); } static void __kill_instance(stratum_instance_t *client) { if (client->proxy) { client->proxy->bound_clients--; client->proxy->parent->combined_clients--; } free(client->workername); free(client->useragent); free(client); } static void __del_disconnected(sdata_t *sdata, stratum_instance_t *client) { HASH_DEL(sdata->disconnected_instances, client); sdata->stats.disconnected--; __kill_instance(client); } /* Removes a client instance we know is on the stratum_instances list and from * the user client list if it's been placed on it */ static void __del_client(sdata_t *sdata, stratum_instance_t *client, user_instance_t *user) { HASH_DEL(sdata->stratum_instances, client); if (user) DL_DELETE(user->clients, client); } static void connector_drop_client(ckpool_t *ckp, const int64_t id) { char buf[256]; LOGDEBUG("Stratifier requesting connector drop client %"PRId64, id); snprintf(buf, 255, "dropclient=%"PRId64, id); send_proc(ckp->connector, buf); } static void drop_allclients(ckpool_t *ckp) { stratum_instance_t *client, *tmp; int disconnects = 0, kills = 0; sdata_t *sdata = ckp->data; ck_wlock(&sdata->instance_lock); HASH_ITER(hh, sdata->stratum_instances, client, tmp) { int64_t client_id = client->id; if (!client->ref) { __del_client(sdata, client, client->user_instance); __kill_instance(client); } else client->dropped = true; kills++; connector_drop_client(ckp, client_id); } HASH_ITER(hh, sdata->disconnected_instances, client, tmp) { disconnects++; __del_disconnected(sdata, client); } sdata->stats.users = sdata->stats.workers = 0; ck_wunlock(&sdata->instance_lock); if (disconnects) LOGNOTICE("Disconnected %d instances", disconnects); if (kills) LOGNOTICE("Dropped %d instances", kills); } /* Copy only the relevant parts of the master sdata for each subproxy */ static sdata_t *duplicate_sdata(const sdata_t *sdata) { sdata_t *dsdata = ckzalloc(sizeof(sdata_t)); dsdata->ckp = sdata->ckp; /* Copy the transaction binaries for workbase creation */ memcpy(dsdata->pubkeytxnbin, sdata->pubkeytxnbin, 25); memcpy(dsdata->donkeytxnbin, sdata->donkeytxnbin, 25); /* Use the same work queues for all subproxies */ dsdata->ssends = sdata->ssends; dsdata->srecvs = sdata->srecvs; dsdata->ckdbq = sdata->ckdbq; dsdata->sshareq = sdata->sshareq; dsdata->sauthq = sdata->sauthq; dsdata->stxnq = sdata->stxnq; /* Give the sbuproxy its own workbase list and lock */ cklock_init(&dsdata->workbase_lock); return dsdata; } static proxy_t *__generate_proxy(sdata_t *sdata, const int64_t id) { proxy_t *proxy = ckzalloc(sizeof(proxy_t)); proxy->parent = proxy; proxy->id = id; proxy->low_id = id & 0xFFFFFFFF; proxy->sdata = duplicate_sdata(sdata); proxy->sdata->subproxy = proxy; proxy->sdata->verbose = true; /* subid == 0 on parent proxy */ HASH_ADD(sh, proxy->subproxies, subid, sizeof(int), proxy); HASH_ADD_I64(sdata->proxies, id, proxy); sdata->proxy_count++; return proxy; } static proxy_t *__generate_subproxy(sdata_t *sdata, proxy_t *proxy, const int subid) { proxy_t *subproxy = ckzalloc(sizeof(proxy_t)); subproxy->parent = proxy; subproxy->id = proxy->id; subproxy->low_id = proxy->low_id; subproxy->subid = subid; HASH_ADD(sh, proxy->subproxies, subid, sizeof(int), subproxy); subproxy->sdata = duplicate_sdata(sdata); subproxy->sdata->subproxy = subproxy; return subproxy; } static proxy_t *__existing_proxy(const sdata_t *sdata, const int64_t id) { proxy_t *proxy; HASH_FIND_I64(sdata->proxies, &id, proxy); return proxy; } /* Find proxy by id number, generate one if none exist yet by that id */ static proxy_t *__proxy_by_id(sdata_t *sdata, const int64_t id) { proxy_t *proxy = __existing_proxy(sdata, id); if (unlikely(!proxy)) { proxy = __generate_proxy(sdata, id); LOGNOTICE("Stratifier added new proxy %ld", id); } return proxy; } static proxy_t *__existing_subproxy(proxy_t *proxy, const int subid) { proxy_t *subproxy; HASH_FIND(sh, proxy->subproxies, &subid, sizeof(int), subproxy); return subproxy; } static proxy_t *__subproxy_by_id(sdata_t *sdata, proxy_t *proxy, const int subid) { proxy_t *subproxy = __existing_subproxy(proxy, subid); if (!subproxy) { subproxy = __generate_subproxy(sdata, proxy, subid); LOGINFO("Stratifier added new subproxy %ld:%d", proxy->id, subid); } return subproxy; } static proxy_t *subproxy_by_id(sdata_t *sdata, const int64_t id, const int subid) { proxy_t *proxy, *subproxy; mutex_lock(&sdata->proxy_lock); proxy = __proxy_by_id(sdata, id); subproxy = __subproxy_by_id(sdata, proxy, subid); mutex_unlock(&sdata->proxy_lock); return subproxy; } static proxy_t *existing_subproxy(sdata_t *sdata, const int64_t id, const int subid) { proxy_t *proxy, *subproxy = NULL; mutex_lock(&sdata->proxy_lock); proxy = __existing_proxy(sdata, id); if (proxy) subproxy = __existing_subproxy(proxy, subid); mutex_unlock(&sdata->proxy_lock); return subproxy; } /* Set proxy to the current proxy and calculate how much headroom it has */ static int64_t current_headroom(sdata_t *sdata, proxy_t **proxy) { proxy_t *subproxy, *tmp; int64_t headroom = 0; mutex_lock(&sdata->proxy_lock); *proxy = sdata->proxy; if (!*proxy) goto out_unlock; HASH_ITER(sh, (*proxy)->subproxies, subproxy, tmp) { if (subproxy->dead) continue; headroom += subproxy->max_clients - subproxy->clients; } out_unlock: mutex_unlock(&sdata->proxy_lock); return headroom; } static void reconnect_client(sdata_t *sdata, stratum_instance_t *client); static void generator_recruit(const ckpool_t *ckp, const int recruits) { char buf[256]; sprintf(buf, "recruit=%d", recruits); LOGINFO("Stratifer requesting %d more proxies from generator", recruits); send_generator(ckp, buf, GEN_PRIORITY); } /* Find how much headroom we have and connect up to that many clients that are * not currently on this pool, setting the reconnect for the remainder to be * switched lazily. */ static void reconnect_clients(sdata_t *sdata) { stratum_instance_t *client, *tmpclient; int reconnects = 0, hard = 0; int64_t headroom; proxy_t *proxy; headroom = current_headroom(sdata, &proxy); if (!proxy) return; ck_rlock(&sdata->instance_lock); HASH_ITER(hh, sdata->stratum_instances, client, tmpclient) { if (client->proxyid == proxy->id) continue; if (headroom-- < 1) continue; reconnects++; /* Limit reconnects sent concurrently to prevent a flood of new * connections */ if (client->reconnect && hard <= SOMAXCONN / 2) { hard++; reconnect_client(sdata, client); } else client->reconnect = true; } ck_runlock(&sdata->instance_lock); if (reconnects) { LOGNOTICE("%d clients flagged for reconnect to proxy %ld", reconnects, proxy->id); } if (headroom < 0) generator_recruit(sdata->ckp, -headroom); } #if 0 static proxy_t *current_proxy(sdata_t *sdata) { proxy_t *proxy; mutex_lock(&sdata->proxy_lock); proxy = sdata->proxy; mutex_unlock(&sdata->proxy_lock); return proxy; } #endif static void dead_proxyid(sdata_t *sdata, const int64_t id, const int subid) { stratum_instance_t *client, *tmp; int reconnects = 0, hard = 0; int64_t headroom; proxy_t *proxy; proxy = existing_subproxy(sdata, id, subid); if (proxy) proxy->dead = true; LOGINFO("Stratifier dropping clients from proxy %ld:%d", id, subid); headroom = current_headroom(sdata, &proxy); ck_rlock(&sdata->instance_lock); HASH_ITER(hh, sdata->stratum_instances, client, tmp) { if (client->proxyid != id || client->subproxyid != subid) continue; reconnects++; if (headroom-- > 0 && client->reconnect && hard <= SOMAXCONN / 2) { hard++; reconnect_client(sdata, client); } else client->reconnect = true; } ck_runlock(&sdata->instance_lock); if (reconnects) { LOGNOTICE("%d clients flagged to reconnect from dead proxy %ld:%d", reconnects, id, subid); } if (headroom < 0) generator_recruit(sdata->ckp, -headroom); } static void update_subscribe(ckpool_t *ckp, const char *cmd) { sdata_t *sdata = ckp->data, *dsdata; proxy_t *proxy, *old = NULL; const char *buf; int64_t id = 0; int subid = 0; json_t *val; if (unlikely(strlen(cmd) < 11)) { LOGWARNING("Received zero length string for subscribe in update_subscribe"); return; } buf = cmd + 10; LOGDEBUG("Update subscribe: %s", buf); val = json_loads(buf, 0, NULL); if (unlikely(!val)) { LOGWARNING("Failed to json decode subscribe response in update_subscribe %s", buf); return; } if (unlikely(!json_get_int64(&id, val, "proxy"))) { LOGWARNING("Failed to json decode proxy value in update_subscribe %s", buf); return; } if (unlikely(!json_get_int(&subid, val, "subproxy"))) { LOGWARNING("Failed to json decode subproxy value in update_subscribe %s", buf); return; } if (!subid) LOGNOTICE("Got updated subscribe for proxy %ld", id); else LOGINFO("Got updated subscribe for proxy %ld:%d", id, subid); /* Is this a replacement for an existing proxy id? */ old = existing_subproxy(sdata, id, subid); if (old) { if (old->dead) dead_proxyid(sdata, id, subid); proxy = old; proxy->dead = false; } else proxy = subproxy_by_id(sdata, id, subid); dsdata = proxy->sdata; ck_wlock(&dsdata->workbase_lock); proxy->subscribed = true; proxy->diff = ckp->startdiff; /* Length is checked by generator */ strcpy(proxy->enonce1, json_string_value(json_object_get(val, "enonce1"))); proxy->enonce1constlen = strlen(proxy->enonce1) / 2; hex2bin(proxy->enonce1bin, proxy->enonce1, proxy->enonce1constlen); proxy->nonce2len = json_integer_value(json_object_get(val, "nonce2len")); if (proxy->nonce2len > 7) proxy->enonce1varlen = 4; else if (proxy->nonce2len > 5) proxy->enonce1varlen = 2; else if (proxy->nonce2len > 3) proxy->enonce1varlen = 1; else proxy->enonce1varlen = 0; proxy->enonce2varlen = proxy->nonce2len - proxy->enonce1varlen; proxy->max_clients = 1ll << (proxy->enonce1varlen * 8); proxy->clients = 0; ck_wunlock(&dsdata->workbase_lock); /* Is this a replacement proxy for the current one */ mutex_lock(&sdata->proxy_lock); if (sdata->proxy && sdata->proxy->low_id == proxy->low_id && !proxy->subid) sdata->proxy = proxy; mutex_unlock(&sdata->proxy_lock); if (subid) { LOGINFO("Upstream pool %ld:%d extranonce2 length %d, max proxy clients %"PRId64, id, subid, proxy->nonce2len, proxy->max_clients); } else { LOGNOTICE("Upstream pool %ld extranonce2 length %d, max proxy clients %"PRId64, id, proxy->nonce2len, proxy->max_clients); } json_decref(val); } static void update_notify(ckpool_t *ckp, const char *cmd) { sdata_t *sdata = ckp->data, *dsdata; bool new_block = false, clean; proxy_t *proxy, *current; int64_t current_id; int i, subid = 0; char header[228]; const char *buf; int64_t id = 0; workbase_t *wb; json_t *val; if (unlikely(strlen(cmd) < 8)) { LOGWARNING("Zero length string passed to update_notify"); return; } buf = cmd + 7; /* "notify=" */ LOGDEBUG("Update notify: %s", buf); val = json_loads(buf, 0, NULL); if (unlikely(!val)) { LOGWARNING("Failed to json decode in update_notify"); return; } json_get_int64(&id, val, "proxy"); json_get_int(&subid, val, "subproxy"); proxy = existing_subproxy(sdata, id, subid); if (unlikely(!proxy || !proxy->subscribed)) { LOGINFO("No valid proxy %ld:%d subscription to update notify yet", id, subid); goto out; } if (!subid) LOGNOTICE("Got updated notify for proxy %ld", id); else LOGINFO("Got updated notify for proxy %ld:%d", id, subid); wb = ckzalloc(sizeof(workbase_t)); wb->ckp = ckp; wb->proxy = true; json_get_int64(&wb->id, val, "jobid"); json_strcpy(wb->prevhash, val, "prevhash"); json_intcpy(&wb->coinb1len, val, "coinb1len"); wb->coinb1bin = ckalloc(wb->coinb1len); wb->coinb1 = ckalloc(wb->coinb1len * 2 + 1); json_strcpy(wb->coinb1, val, "coinbase1"); hex2bin(wb->coinb1bin, wb->coinb1, wb->coinb1len); wb->height = get_sernumber(wb->coinb1bin + 42); json_strdup(&wb->coinb2, val, "coinbase2"); wb->coinb2len = strlen(wb->coinb2) / 2; wb->coinb2bin = ckalloc(wb->coinb2len); hex2bin(wb->coinb2bin, wb->coinb2, wb->coinb2len); wb->merkle_array = json_object_dup(val, "merklehash"); wb->merkles = json_array_size(wb->merkle_array); for (i = 0; i < wb->merkles; i++) { strcpy(&wb->merklehash[i][0], json_string_value(json_array_get(wb->merkle_array, i))); hex2bin(&wb->merklebin[i][0], &wb->merklehash[i][0], 32); } json_strcpy(wb->bbversion, val, "bbversion"); json_strcpy(wb->nbit, val, "nbit"); json_strcpy(wb->ntime, val, "ntime"); sscanf(wb->ntime, "%x", &wb->ntime32); clean = json_is_true(json_object_get(val, "clean")); ts_realtime(&wb->gentime); snprintf(header, 225, "%s%s%s%s%s%s%s", wb->bbversion, wb->prevhash, "0000000000000000000000000000000000000000000000000000000000000000", wb->ntime, wb->nbit, "00000000", /* nonce */ workpadding); LOGDEBUG("Header: %s", header); hex2bin(wb->headerbin, header, 112); wb->txn_hashes = ckzalloc(1); dsdata = proxy->sdata; ck_rlock(&dsdata->workbase_lock); strcpy(wb->enonce1const, proxy->enonce1); wb->enonce1constlen = proxy->enonce1constlen; memcpy(wb->enonce1constbin, proxy->enonce1bin, wb->enonce1constlen); wb->enonce1varlen = proxy->enonce1varlen; wb->enonce2varlen = proxy->enonce2varlen; wb->diff = proxy->diff; ck_runlock(&dsdata->workbase_lock); add_base(ckp, dsdata, wb, &new_block); if (new_block) { if (subid) LOGINFO("Block hash on proxy %ld:%d changed to %s", id, subid, dsdata->lastswaphash); else LOGNOTICE("Block hash on proxy %ld changed to %s", id, dsdata->lastswaphash); } mutex_lock(&sdata->proxy_lock); current = sdata->proxy; if (unlikely(!current)) current = sdata->proxy = proxy; current_id = current->id; mutex_unlock(&sdata->proxy_lock); if (!proxy->subid && proxy->id == current_id) reconnect_clients(sdata); clean |= new_block; LOGINFO("Proxy %ld:%d broadcast updated stratum notify with%s clean", id, subid, clean ? "" : "out"); stratum_broadcast_update(dsdata, wb, clean); out: json_decref(val); } static void stratum_send_diff(sdata_t *sdata, const stratum_instance_t *client); static void update_diff(ckpool_t *ckp, const char *cmd) { sdata_t *sdata = ckp->data, *dsdata; stratum_instance_t *client, *tmp; double old_diff, diff; const char *buf; int64_t id = 0; proxy_t *proxy; int subid = 0; json_t *val; if (unlikely(strlen(cmd) < 6)) { LOGWARNING("Zero length string passed to update_diff"); return; } buf = cmd + 5; /* "diff=" */ LOGDEBUG("Update diff: %s", buf); val = json_loads(buf, 0, NULL); if (unlikely(!val)) { LOGWARNING("Failed to json decode in update_diff"); return; } json_get_int64(&id, val, "proxy"); json_get_int(&subid, val, "subproxy"); json_dblcpy(&diff, val, "diff"); json_decref(val); LOGINFO("Got updated diff for proxy %ld:%d", id, subid); proxy = existing_subproxy(sdata, id, subid); if (!proxy) { LOGINFO("No existing subproxy %ld:%d to update diff", id, subid); return; } /* We only really care about integer diffs so clamp the lower limit to * 1 or it will round down to zero. */ if (unlikely(diff < 1)) diff = 1; dsdata = proxy->sdata; if (unlikely(!dsdata->current_workbase)) { LOGINFO("No current workbase to update diff yet"); return; } ck_wlock(&dsdata->workbase_lock); old_diff = proxy->diff; dsdata->current_workbase->diff = proxy->diff = diff; ck_wunlock(&dsdata->workbase_lock); if (old_diff < diff) return; /* If the diff has dropped, iterate over all the clients and check * they're at or below the new diff, and update it if not. */ ck_rlock(&sdata->instance_lock); HASH_ITER(hh, sdata->stratum_instances, client, tmp) { if (client->proxyid != id) continue; if (client->subproxyid != subid) continue; if (client->diff > diff) { client->diff = diff; stratum_send_diff(sdata, client); } } ck_runlock(&sdata->instance_lock); } #if 0 static void generator_drop_proxy(ckpool_t *ckp, const int64_t id, const int subid) { char msg[256]; sprintf(msg, "dropproxy=%ld:%d", id, subid); send_generator(ckp, msg, GEN_LAX); } #endif static void free_proxy(proxy_t *proxy) { free(proxy->sdata); free(proxy); } /* Remove subproxies that are flagged dead. Then see if there * are any retired proxies that no longer have any other subproxies and reap * those. */ static void reap_proxies(ckpool_t *ckp, sdata_t *sdata) { proxy_t *proxy, *proxytmp, *subproxy, *subtmp; int dead = 0; if (!ckp->proxy) return; mutex_lock(&sdata->proxy_lock); HASH_ITER(hh, sdata->proxies, proxy, proxytmp) { HASH_ITER(sh, proxy->subproxies, subproxy, subtmp) { if (!subproxy->bound_clients && !subproxy->dead) { /* Reset the counter to reuse this proxy */ subproxy->clients = 0; continue; } if (proxy == subproxy) continue; if (subproxy->bound_clients) continue; if (!subproxy->dead) continue; if (unlikely(!subproxy->subid)) { LOGWARNING("Unexepectedly found proxy %ld:%d as subproxy of %ld:%d", subproxy->id, subproxy->subid, proxy->id, proxy->subid); continue; } if (unlikely(subproxy == sdata->proxy)) { LOGWARNING("Unexepectedly found proxy %ld:%d as current", subproxy->id, subproxy->subid); continue; } dead++; HASH_DELETE(sh, proxy->subproxies, subproxy); free_proxy(subproxy); } } mutex_unlock(&sdata->proxy_lock); if (dead) LOGINFO("Stratifier discarded %d dead proxies", dead); } /* Enter with instance_lock held */ static stratum_instance_t *__instance_by_id(sdata_t *sdata, const int64_t id) { stratum_instance_t *client; HASH_FIND_I64(sdata->stratum_instances, &id, client); return client; } /* Increase the reference count of instance */ static void __inc_instance_ref(stratum_instance_t *client) { client->ref++; } /* Find an __instance_by_id and increase its reference count allowing us to * use this instance outside of instance_lock without fear of it being * dereferenced. Does not return dropped clients still on the list. */ static stratum_instance_t *ref_instance_by_id(sdata_t *sdata, const int64_t id) { stratum_instance_t *client; ck_ilock(&sdata->instance_lock); client = __instance_by_id(sdata, id); if (client) { if (unlikely(client->dropped)) client = NULL; else { /* Upgrade to write lock to modify client refcount */ ck_ulock(&sdata->instance_lock); __inc_instance_ref(client); ck_dwilock(&sdata->instance_lock); } } ck_uilock(&sdata->instance_lock); return client; } /* Has this client_id already been used and is now in one of the dropped lists */ static bool __dropped_instance(sdata_t *sdata, const int64_t id) { stratum_instance_t *client, *tmp; bool ret = false; HASH_ITER(hh, sdata->disconnected_instances, client, tmp) { if (unlikely(client->id == id)) { ret = true; goto out; } } out: return ret; } /* Ret = 1 is disconnected, 2 is killed, 3 is workerless killed */ static int __drop_client(sdata_t *sdata, stratum_instance_t *client, user_instance_t *user) { stratum_instance_t *old_client = NULL; int ret; __del_client(sdata, client, user); HASH_FIND(hh, sdata->disconnected_instances, &client->enonce1_64, sizeof(uint64_t), old_client); /* Only keep around one copy of the old client in server mode */ if (!client->ckp->proxy && !old_client && client->enonce1_64 && client->authorised) { ret = 1; HASH_ADD(hh, sdata->disconnected_instances, enonce1_64, sizeof(uint64_t), client); sdata->stats.disconnected++; sdata->disconnected_generated++; client->disconnected_time = time(NULL); } else { if (client->workername) ret = 2; else ret = 3; __kill_instance(client); } return ret; } static void client_drop_message(const int64_t client_id, const int dropped, const bool lazily) { switch(dropped) { case 0: break; case 1: LOGNOTICE("Client %"PRId64" disconnected %s", client_id, lazily ? "lazily" : ""); break; case 2: LOGNOTICE("Client %"PRId64" dropped %s", client_id, lazily ? "lazily" : ""); break; case 3: LOGNOTICE("Workerless client %"PRId64" dropped %s", client_id, lazily ? "lazily" : ""); break; } } static void dec_worker(ckpool_t *ckp, user_instance_t *instance); /* Decrease the reference count of instance. */ static void _dec_instance_ref(sdata_t *sdata, stratum_instance_t *client, const char *file, const char *func, const int line) { user_instance_t *user = client->user_instance; int64_t client_id = client->id; int dropped = 0, ref; ck_wlock(&sdata->instance_lock); ref = --client->ref; /* See if there are any instances that were dropped that could not be * moved due to holding a reference and drop them now. */ if (unlikely(client->dropped && !ref)) dropped = __drop_client(sdata, client, user); ck_wunlock(&sdata->instance_lock); client_drop_message(client_id, dropped, true); /* This should never happen */ if (unlikely(ref < 0)) LOGERR("Instance ref count dropped below zero from %s %s:%d", file, func, line); if (dropped) { if (user) dec_worker(sdata->ckp, user); reap_proxies(sdata->ckp, sdata); } } #define dec_instance_ref(sdata, instance) _dec_instance_ref(sdata, instance, __FILE__, __func__, __LINE__) /* Enter with write instance_lock held */ static stratum_instance_t *__stratum_add_instance(ckpool_t *ckp, const int64_t id, const int server) { stratum_instance_t *client = ckzalloc(sizeof(stratum_instance_t)); sdata_t *sdata = ckp->data; client->start_time = time(NULL); sdata->stratum_generated++; client->id = id; client->server = server; client->diff = client->old_diff = ckp->startdiff; client->ckp = ckp; tv_time(&client->ldc); HASH_ADD_I64(sdata->stratum_instances, id, client); /* Points to ckp sdata in ckpool mode, but is changed later in proxy * mode . */ client->sdata = sdata; return client; } static uint64_t disconnected_sessionid_exists(sdata_t *sdata, const char *sessionid, const int64_t id) { stratum_instance_t *client, *tmp; uint64_t enonce1_64 = 0, ret = 0; int64_t old_id = 0; int slen; if (!sessionid) goto out; slen = strlen(sessionid) / 2; if (slen < 1 || slen > 8) goto out; if (!validhex(sessionid)) goto out; /* Number is in BE but we don't swap either of them */ hex2bin(&enonce1_64, sessionid, slen); ck_ilock(&sdata->instance_lock); HASH_ITER(hh, sdata->stratum_instances, client, tmp) { if (client->id == id) continue; if (client->enonce1_64 == enonce1_64) { /* Only allow one connected instance per enonce1 */ goto out_unlock; } } client = NULL; HASH_FIND(hh, sdata->disconnected_instances, &enonce1_64, sizeof(uint64_t), client); if (client) { /* Delete the entry once we are going to use it since there * will be a new instance with the enonce1_64 */ old_id = client->id; /* Upgrade to write lock to disconnect */ ck_ulock(&sdata->instance_lock); __del_disconnected(sdata, client); ck_dwilock(&sdata->instance_lock); ret = enonce1_64; } out_unlock: ck_uilock(&sdata->instance_lock); out: if (ret) LOGNOTICE("Reconnecting old instance %"PRId64" to instance %"PRId64, old_id, id); return ret; } static inline bool client_active(stratum_instance_t *client) { return (client->authorised && !client->dropped); } /* Ask the connector asynchronously to send us dropclient commands if this * client no longer exists. */ static void connector_test_client(ckpool_t *ckp, const int64_t id) { char buf[256]; LOGDEBUG("Stratifier requesting connector test client %"PRId64, id); snprintf(buf, 255, "testclient=%"PRId64, id); send_proc(ckp->connector, buf); } /* For creating a list of sends without locking that can then be concatenated * to the stratum_sends list. Minimises locking and avoids taking recursive * locks. Sends only to sdata bound clients (everyone in ckpool) */ static void stratum_broadcast(sdata_t *sdata, json_t *val) { ckpool_t *ckp = sdata->ckp; sdata_t *ckp_sdata = ckp->data; stratum_instance_t *client, *tmp; ckmsg_t *bulk_send = NULL; time_t now_t = time(NULL); int hard_reconnect = 0; ckmsgq_t *ssends; if (unlikely(!val)) { LOGERR("Sent null json to stratum_broadcast"); return; } /* Use this locking as an opportunity to test other clients. */ ck_rlock(&ckp_sdata->instance_lock); HASH_ITER(hh, ckp_sdata->stratum_instances, client, tmp) { ckmsg_t *client_msg; smsg_t *msg; if (sdata != ckp_sdata && client->sdata != sdata) continue; /* Look for clients that may have been dropped which the stratifer has * not been informed about and ask the connector of they still exist */ if (client->dropped) { connector_test_client(ckp, client->id); continue; } /* Test for clients that haven't authed in over a minute and drop them */ if (!client->authorised) { if (now_t > client->start_time + 60) { client->dropped = true; connector_drop_client(ckp, client->id); } continue; } if (client->reconnect) { if (hard_reconnect <= SOMAXCONN / 2) { hard_reconnect++; reconnect_client(ckp_sdata, client); } continue; } if (!client_active(client)) continue; client_msg = ckalloc(sizeof(ckmsg_t)); msg = ckzalloc(sizeof(smsg_t)); msg->json_msg = json_deep_copy(val); msg->client_id = client->id; client_msg->data = msg; DL_APPEND(bulk_send, client_msg); } ck_runlock(&ckp_sdata->instance_lock); json_decref(val); if (!bulk_send) return; ssends = sdata->ssends; mutex_lock(ssends->lock); if (ssends->msgs) DL_CONCAT(ssends->msgs, bulk_send); else ssends->msgs = bulk_send; pthread_cond_signal(ssends->cond); mutex_unlock(ssends->lock); } static void stratum_add_send(sdata_t *sdata, json_t *val, const int64_t client_id) { smsg_t *msg; msg = ckzalloc(sizeof(smsg_t)); msg->json_msg = val; msg->client_id = client_id; ckmsgq_add(sdata->ssends, msg); } static void inc_worker(ckpool_t *ckp, user_instance_t *instance) { sdata_t *sdata = ckp->data; mutex_lock(&sdata->stats_lock); sdata->stats.workers++; if (!instance->workers++) sdata->stats.users++; mutex_unlock(&sdata->stats_lock); } static void dec_worker(ckpool_t *ckp, user_instance_t *instance) { sdata_t *sdata = ckp->data; mutex_lock(&sdata->stats_lock); sdata->stats.workers--; if (!--instance->workers) sdata->stats.users--; mutex_unlock(&sdata->stats_lock); } static void drop_client(ckpool_t *ckp, sdata_t *sdata, const int64_t id) { stratum_instance_t *client, *tmp; user_instance_t *user = NULL; int dropped = 0, aged = 0; time_t now_t = time(NULL); LOGINFO("Stratifier asked to drop client %"PRId64, id); ck_ilock(&sdata->instance_lock); client = __instance_by_id(sdata, id); /* Upgrade to write lock */ ck_ulock(&sdata->instance_lock); if (client) { /* If the client is still holding a reference, don't drop them * now but wait till the reference is dropped */ if (!client->ref) { user = client->user_instance; dropped = __drop_client(sdata, client, user); } else client->dropped = true; } /* Old disconnected instances will not have any valid shares so remove * them from the disconnected instances list if they've been dead for * more than 10 minutes */ HASH_ITER(hh, sdata->disconnected_instances, client, tmp) { if (now_t - client->disconnected_time < 600) continue; aged++; __del_disconnected(sdata, client); } ck_wunlock(&sdata->instance_lock); client_drop_message(id, dropped, false); if (aged) LOGINFO("Aged %d disconnected instances to dead", aged); /* Decrease worker count outside of instance_lock to avoid recursive * locking */ if (user) dec_worker(ckp, user); if (aged || dropped) reap_proxies(ckp, sdata); } static void stratum_broadcast_message(sdata_t *sdata, const char *msg) { json_t *json_msg; JSON_CPACK(json_msg, "{sosss[s]}", "id", json_null(), "method", "client.show_message", "params", msg); stratum_broadcast(sdata, json_msg); } /* Send a generic reconnect to all clients without parameters to make them * reconnect to the same server. */ static void request_reconnect(sdata_t *sdata, const char *cmd) { char *port = strdupa(cmd), *url = NULL; stratum_instance_t *client, *tmp; json_t *json_msg; strsep(&port, ":"); if (port) url = strsep(&port, ","); if (url && port) { int port_no; port_no = strtol(port, NULL, 10); JSON_CPACK(json_msg, "{sosss[sii]}", "id", json_null(), "method", "client.reconnect", "params", url, port_no, 0); } else JSON_CPACK(json_msg, "{sosss[]}", "id", json_null(), "method", "client.reconnect", "params"); stratum_broadcast(sdata, json_msg); /* Tag all existing clients as dropped now so they can be removed * lazily */ ck_wlock(&sdata->instance_lock); HASH_ITER(hh, sdata->stratum_instances, client, tmp) { client->dropped = true; } ck_wunlock(&sdata->instance_lock); } static void reset_bestshares(sdata_t *sdata) { user_instance_t *user, *tmpuser; stratum_instance_t *client, *tmp; ck_rlock(&sdata->instance_lock); HASH_ITER(hh, sdata->stratum_instances, client, tmp) { client->best_diff = 0; } HASH_ITER(hh, sdata->user_instances, user, tmpuser) { worker_instance_t *worker; user->best_diff = 0; DL_FOREACH(user->worker_instances, worker) { worker->best_diff = 0; } } ck_runlock(&sdata->instance_lock); } static void block_solve(ckpool_t *ckp, const char *blockhash) { ckmsg_t *block, *tmp, *found = NULL; sdata_t *sdata = ckp->data; char cdfield[64]; int height = 0; ts_t ts_now; json_t *val; char *msg; update_base(ckp, GEN_PRIORITY); ts_realtime(&ts_now); sprintf(cdfield, "%lu,%lu", ts_now.tv_sec, ts_now.tv_nsec); mutex_lock(&sdata->block_lock); DL_FOREACH_SAFE(sdata->block_solves, block, tmp) { val = block->data; char *solvehash; json_get_string(&solvehash, val, "blockhash"); if (unlikely(!solvehash)) { LOGERR("Failed to find blockhash in block_solve json!"); continue; } if (!strcmp(solvehash, blockhash)) { dealloc(solvehash); found = block; DL_DELETE(sdata->block_solves, block); break; } dealloc(solvehash); } mutex_unlock(&sdata->block_lock); if (unlikely(!found)) { LOGERR("Failed to find blockhash %s in block_solve!", blockhash); return; } val = found->data; json_set_string(val, "confirmed", "1"); json_set_string(val, "createdate", cdfield); json_set_string(val, "createcode", __func__); json_get_int(&height, val, "height"); ckdbq_add(ckp, ID_BLOCK, val); free(found); ASPRINTF(&msg, "Block %d solved by %s!", height, ckp->name); stratum_broadcast_message(sdata, msg); free(msg); LOGWARNING("Solved and confirmed block %d", height); reset_bestshares(sdata); } static void block_reject(sdata_t *sdata, const char *blockhash) { ckmsg_t *block, *tmp, *found = NULL; int height = 0; json_t *val; mutex_lock(&sdata->block_lock); DL_FOREACH_SAFE(sdata->block_solves, block, tmp) { val = block->data; char *solvehash; json_get_string(&solvehash, val, "blockhash"); if (unlikely(!solvehash)) { LOGERR("Failed to find blockhash in block_reject json!"); continue; } if (!strcmp(solvehash, blockhash)) { dealloc(solvehash); found = block; DL_DELETE(sdata->block_solves, block); break; } dealloc(solvehash); } mutex_unlock(&sdata->block_lock); if (unlikely(!found)) { LOGERR("Failed to find blockhash %s in block_reject!", blockhash); return; } val = found->data; json_get_int(&height, val, "height"); json_decref(val); free(found); LOGWARNING("Submitted, but rejected block %d", height); } /* Some upstream pools (like p2pool) don't update stratum often enough and * miners disconnect if they don't receive regular communication so send them * a ping at regular intervals */ static void broadcast_ping(sdata_t *sdata) { json_t *json_msg; JSON_CPACK(json_msg, "{s:[],s:i,s:s}", "params", "id", 42, "method", "mining.ping"); stratum_broadcast(sdata, json_msg); } static void ckmsgq_stats(ckmsgq_t *ckmsgq, const int size, json_t **val) { int objects, generated; int64_t memsize; ckmsg_t *msg; mutex_lock(ckmsgq->lock); DL_COUNT(ckmsgq->msgs, msg, objects); generated = ckmsgq->messages; mutex_unlock(ckmsgq->lock); memsize = (sizeof(ckmsg_t) + size) * objects; JSON_CPACK(*val, "{si,si,si}", "count", objects, "memory", memsize, "generated", generated); } static char *stratifier_stats(ckpool_t *ckp, sdata_t *sdata) { json_t *val = json_object(), *subval; int objects, generated; int64_t memsize; char *buf; ck_rlock(&sdata->workbase_lock); objects = HASH_COUNT(sdata->workbases); memsize = SAFE_HASH_OVERHEAD(sdata->workbases) + sizeof(workbase_t) * objects; generated = sdata->workbases_generated; ck_runlock(&sdata->workbase_lock); JSON_CPACK(subval, "{si,si,si}", "count", objects, "memory", memsize, "generated", generated); json_set_object(val, "workbases", subval); ck_rlock(&sdata->instance_lock); objects = HASH_COUNT(sdata->user_instances); memsize = SAFE_HASH_OVERHEAD(sdata->user_instances) + sizeof(stratum_instance_t) * objects; JSON_CPACK(subval, "{si,si}", "count", objects, "memory", memsize); json_set_object(val, "users", subval); objects = HASH_COUNT(sdata->stratum_instances); memsize = SAFE_HASH_OVERHEAD(sdata->stratum_instances); generated = sdata->stratum_generated; JSON_CPACK(subval, "{si,si,si}", "count", objects, "memory", memsize, "generated", generated); json_set_object(val, "clients", subval); objects = sdata->stats.disconnected; generated = sdata->disconnected_generated; memsize = sizeof(stratum_instance_t) * sdata->stats.disconnected; JSON_CPACK(subval, "{si,si,si}", "count", objects, "memory", memsize, "generated", generated); json_set_object(val, "disconnected", subval); ck_runlock(&sdata->instance_lock); mutex_lock(&sdata->share_lock); generated = sdata->shares_generated; objects = HASH_COUNT(sdata->shares); memsize = SAFE_HASH_OVERHEAD(sdata->shares) + sizeof(share_t) * objects; mutex_unlock(&sdata->share_lock); JSON_CPACK(subval, "{si,si,si}", "count", objects, "memory", memsize, "generated", generated); json_set_object(val, "shares", subval); ckmsgq_stats(sdata->ssends, sizeof(smsg_t), &subval); json_set_object(val, "ssends", subval); /* Don't know exactly how big the string is so just count the pointer for now */ ckmsgq_stats(sdata->srecvs, sizeof(char *), &subval); json_set_object(val, "srecvs", subval); if (!CKP_STANDALONE(ckp)) { ckmsgq_stats(sdata->ckdbq, sizeof(char *), &subval); json_set_object(val, "ckdbq", subval); } ckmsgq_stats(sdata->stxnq, sizeof(json_params_t), &subval); json_set_object(val, "stxnq", subval); buf = json_dumps(val, JSON_NO_UTF8 | JSON_PRESERVE_ORDER); json_decref(val); LOGNOTICE("Stratifier stats: %s", buf); return buf; } /* Send a single client a reconnect request, setting the time we sent the * request so we can drop the client lazily if it hasn't reconnected on its * own one minute later */ static void reconnect_client(sdata_t *sdata, stratum_instance_t *client) { json_t *json_msg; /* Already requested? */ if (client->reconnect_request) { connector_drop_client(sdata->ckp, client->id); client->dropped = true; return; } client->reconnect_request = time(NULL); JSON_CPACK(json_msg, "{sosss[]}", "id", json_null(), "method", "client.reconnect", "params"); stratum_add_send(sdata, json_msg, client->id); } /* Sets the currently active proxy. Clients will be told to reconnect once the * first notify data comes from this proxy. Even if we are already bound to * this proxy we are only given this message if all clients must move. */ static void set_proxy(sdata_t *sdata, const char *buf) { int64_t id = 0; proxy_t *proxy; sscanf(buf, "proxy=%ld", &id); mutex_lock(&sdata->proxy_lock); proxy = __proxy_by_id(sdata, id); sdata->proxy = proxy; mutex_unlock(&sdata->proxy_lock); LOGNOTICE("Stratifier setting active proxy to %ld", id); } static void dead_proxy(sdata_t *sdata, const char *buf) { int64_t id = 0; int subid = 0; sscanf(buf, "deadproxy=%ld:%d", &id, &subid); dead_proxyid(sdata, id, subid); } /* Must hold a reference */ static void lazy_reconnect_client(sdata_t *sdata, stratum_instance_t *client) { int64_t headroom; proxy_t *proxy; headroom = current_headroom(sdata, &proxy); if (!proxy) return; if (headroom-- > 0) { LOGNOTICE("Reconnecting client %"PRId64, client->id); reconnect_client(sdata, client); } else { generator_recruit(sdata->ckp, -headroom); if (!client->reconnect) { LOGNOTICE("Flagging client %"PRId64, client->id); client->reconnect = true; } else /* Already been flagged, force the send */ reconnect_client(sdata, client); } } static void reconnect_client_id(sdata_t *sdata, const int64_t client_id) { stratum_instance_t *client; client = ref_instance_by_id(sdata, client_id); if (!client) { LOGINFO("reconnect_client_id failed to find client %"PRId64, client_id); return; } lazy_reconnect_client(sdata, client); dec_instance_ref(sdata, client); } static int stratum_loop(ckpool_t *ckp, proc_instance_t *pi) { int sockd, ret = 0, selret = 0; sdata_t *sdata = ckp->data; unixsock_t *us = &pi->us; tv_t start_tv = {0, 0}; char *buf = NULL; retry: do { double tdiff; tv_t end_tv; tv_time(&end_tv); tdiff = tvdiff(&end_tv, &start_tv); if (tdiff > ckp->update_interval) { copy_tv(&start_tv, &end_tv); if (!ckp->proxy) { LOGDEBUG("%ds elapsed in strat_loop, updating gbt base", ckp->update_interval); update_base(ckp, GEN_NORMAL); } else { LOGDEBUG("%ds elapsed in strat_loop, pinging miners", ckp->update_interval); broadcast_ping(sdata); } } selret = wait_read_select(us->sockd, 5); if (!selret && !ping_main(ckp)) { LOGEMERG("Generator failed to ping main process, exiting"); ret = 1; goto out; } } while (selret < 1); sockd = accept(us->sockd, NULL, NULL); if (sockd < 0) { LOGERR("Failed to accept on stratifier socket, exiting"); ret = 1; goto out; } dealloc(buf); buf = recv_unix_msg(sockd); if (unlikely(!buf)) { Close(sockd); LOGWARNING("Failed to get message in stratum_loop"); goto retry; } if (likely(buf[0] == '{')) { /* The bulk of the messages will be received json from the * connector so look for this first. The srecv_process frees * the buf heap ram */ Close(sockd); ckmsgq_add(sdata->srecvs, buf); buf = NULL; goto retry; } if (cmdmatch(buf, "ping")) { LOGDEBUG("Stratifier received ping request"); send_unix_msg(sockd, "pong"); Close(sockd); goto retry; } if (cmdmatch(buf, "stats")) { char *msg; LOGDEBUG("Stratifier received stats request"); msg = stratifier_stats(ckp, sdata); send_unix_msg(sockd, msg); Close(sockd); goto retry; } Close(sockd); LOGDEBUG("Stratifier received request: %s", buf); if (cmdmatch(buf, "shutdown")) { ret = 0; goto out; } else if (cmdmatch(buf, "update")) { update_base(ckp, GEN_PRIORITY); } else if (cmdmatch(buf, "subscribe")) { /* Proxifier has a new subscription */ update_subscribe(ckp, buf); } else if (cmdmatch(buf, "notify")) { /* Proxifier has a new notify ready */ update_notify(ckp, buf); } else if (cmdmatch(buf, "diff")) { update_diff(ckp, buf); } else if (cmdmatch(buf, "dropclient")) { int64_t client_id; ret = sscanf(buf, "dropclient=%"PRId64, &client_id); if (ret < 0) LOGDEBUG("Stratifier failed to parse dropclient command: %s", buf); else drop_client(ckp, sdata, client_id); } else if (cmdmatch(buf, "reconnclient")) { int64_t client_id; ret = sscanf(buf, "reconnclient=%"PRId64, &client_id); if (ret < 0) LOGDEBUG("Stratifier failed to parse reconnclient command: %s", buf); else reconnect_client_id(sdata, client_id); } else if (cmdmatch(buf, "dropall")) { drop_allclients(ckp); } else if (cmdmatch(buf, "block")) { block_solve(ckp, buf + 6); } else if (cmdmatch(buf, "noblock")) { block_reject(sdata, buf + 8); } else if (cmdmatch(buf, "reconnect")) { request_reconnect(sdata, buf); } else if (cmdmatch(buf, "proxy")) { set_proxy(sdata, buf); } else if (cmdmatch(buf, "deadproxy")) { dead_proxy(sdata, buf); } else if (cmdmatch(buf, "loglevel")) { sscanf(buf, "loglevel=%d", &ckp->loglevel); } else LOGWARNING("Unhandled stratifier message: %s", buf); goto retry; out: dealloc(buf); return ret; } static void *blockupdate(void *arg) { ckpool_t *ckp = (ckpool_t *)arg; sdata_t *sdata = ckp->data; char *buf = NULL; char request[8]; pthread_detach(pthread_self()); rename_proc("blockupdate"); buf = send_recv_proc(ckp->generator, "getbest"); if (!cmdmatch(buf, "failed")) sprintf(request, "getbest"); else sprintf(request, "getlast"); while (42) { dealloc(buf); buf = send_recv_generator(ckp, request, GEN_LAX); if (buf && cmdmatch(buf, "notify")) cksleep_ms(5000); else if (buf && strcmp(buf, sdata->lastswaphash) && !cmdmatch(buf, "failed")) update_base(ckp, GEN_PRIORITY); else cksleep_ms(ckp->blockpoll); } return NULL; } /* Enter holding workbase_lock and client a ref count. */ static void __fill_enonce1data(const workbase_t *wb, stratum_instance_t *client) { if (wb->enonce1constlen) memcpy(client->enonce1bin, wb->enonce1constbin, wb->enonce1constlen); if (wb->enonce1varlen) { memcpy(client->enonce1bin + wb->enonce1constlen, &client->enonce1_64, wb->enonce1varlen); __bin2hex(client->enonce1var, &client->enonce1_64, wb->enonce1varlen); } __bin2hex(client->enonce1, client->enonce1bin, wb->enonce1constlen + wb->enonce1varlen); } /* Create a new enonce1 from the 64 bit enonce1_64 value, using only the number * of bytes we have to work with when we are proxying with a split nonce2. * When the proxy space is less than 32 bits to work with, we look for an * unused enonce1 value and reject clients instead if there is no space left. * Needs to be entered with client holding a ref count. */ static bool new_enonce1(ckpool_t *ckp, sdata_t *ckp_sdata, sdata_t *sdata, stratum_instance_t *client) { proxy_t *proxy = NULL; uint64_t enonce1; if (ckp->proxy) { if (!ckp_sdata->proxy) return false; mutex_lock(&ckp_sdata->proxy_lock); proxy = sdata->subproxy; client->proxyid = proxy->id; client->subproxyid = proxy->subid; mutex_unlock(&ckp_sdata->proxy_lock); if (proxy->clients >= proxy->max_clients) { LOGWARNING("Proxy reached max clients %"PRId64, proxy->max_clients); return false; } } /* instance_lock protects enonce1_64. Incrementing a little endian 64bit * number ensures that no matter how many of the bits we take from the * left depending on nonce2 length, we'll always get a changing value * for every next client.*/ ck_wlock(&ckp_sdata->instance_lock); enonce1 = htole64(ckp_sdata->enonce1_64); enonce1++; client->enonce1_64 = ckp_sdata->enonce1_64 = le64toh(enonce1); if (proxy) { client->proxy = proxy; proxy->clients++; proxy->bound_clients++; proxy->parent->combined_clients++; } ck_wunlock(&ckp_sdata->instance_lock); ck_rlock(&sdata->workbase_lock); __fill_enonce1data(sdata->current_workbase, client); ck_runlock(&sdata->workbase_lock); return true; } static void stratum_send_message(sdata_t *sdata, const stratum_instance_t *client, const char *msg); /* Choose the stratifier data for a new client. Use the main ckp_sdata except * in proxy mode where we find a subproxy based on the current proxy with room * for more clients. Signal the generator to recruit more subproxies if we are * running out of room. */ static sdata_t *select_sdata(const ckpool_t *ckp, sdata_t *ckp_sdata) { proxy_t *current, *proxy, *subproxy, *best = NULL, *tmp, *tmpsub; int best_subid = 0, best_lowid; int64_t best_id = 0; if (!ckp->proxy || ckp->passthrough) return ckp_sdata; current = ckp_sdata->proxy; if (!current) { LOGWARNING("No proxy available yet to generate subscribes"); return NULL; } best_lowid = ckp_sdata->proxy_count; mutex_lock(&ckp_sdata->proxy_lock); HASH_ITER(hh, ckp_sdata->proxies, proxy, tmp) { int64_t max_headroom; best = NULL; proxy->headroom = max_headroom = 0; HASH_ITER(sh, proxy->subproxies, subproxy, tmpsub) { int64_t subproxy_headroom; if (subproxy->dead) continue; if (!subproxy->sdata->current_workbase) continue; subproxy_headroom = subproxy->max_clients - subproxy->clients; proxy->headroom += subproxy_headroom; if (subproxy_headroom > max_headroom) { best = subproxy; max_headroom = subproxy_headroom; } } if (best && (best->low_id < best_lowid || (best->low_id == best_lowid && best->id > best_id))) { best_id = best->id; best_lowid = best->low_id; best_subid = best->subid; if (best == current) break; } } mutex_unlock(&ckp_sdata->proxy_lock); if (best_id != current->id || current->headroom < 2) generator_recruit(ckp, 1); if (best_id == ckp_sdata->proxy_count) { LOGNOTICE("Temporarily insufficient subproxies to accept more clients"); return NULL; } best = subproxy_by_id(ckp_sdata, best_id, best_subid); return best->sdata; } /* Extranonce1 must be set here. Needs to be entered with client holding a ref * count. */ static json_t *parse_subscribe(stratum_instance_t *client, const int64_t client_id, const json_t *params_val) { ckpool_t *ckp = client->ckp; sdata_t *sdata, *ckp_sdata = ckp->data; bool old_match = false; int arr_size; json_t *ret; int n2len; if (unlikely(!json_is_array(params_val))) { stratum_send_message(ckp_sdata, client, "Invalid json: params not an array"); return json_string("params not an array"); } sdata = select_sdata(ckp, ckp_sdata); if (unlikely(!sdata || !sdata->current_workbase)) { LOGWARNING("Failed to provide subscription due to no %s", sdata ? "current workbase" : "sdata"); stratum_send_message(ckp_sdata, client, "Pool Initialising"); return json_string("Initialising"); } if (ckp->proxy) { LOGINFO("Current %ld, selecting proxy %ld:%d for client %"PRId64, ckp_sdata->proxy->id, sdata->subproxy->id, sdata->subproxy->subid, client->id); } client->sdata = sdata; arr_size = json_array_size(params_val); /* NOTE useragent is NULL prior to this so should not be used in code * till after this point */ if (arr_size > 0) { const char *buf; buf = json_string_value(json_array_get(params_val, 0)); if (buf && strlen(buf)) client->useragent = strdup(buf); else client->useragent = ckzalloc(1); // Set to "" if (arr_size > 1 && !ckp->proxy) { /* This would be the session id for reconnect, it will * not work for clients on a proxied connection. */ buf = json_string_value(json_array_get(params_val, 1)); LOGDEBUG("Found old session id %s", buf); /* Add matching here */ if ((client->enonce1_64 = disconnected_sessionid_exists(sdata, buf, client_id))) { sprintf(client->enonce1, "%016lx", client->enonce1_64); old_match = true; ck_rlock(&sdata->workbase_lock); __fill_enonce1data(sdata->current_workbase, client); ck_runlock(&sdata->workbase_lock); } } } else client->useragent = ckzalloc(1); if (!old_match) { /* Create a new extranonce1 based on a uint64_t pointer */ if (!new_enonce1(ckp, ckp_sdata, sdata, client)) { stratum_send_message(sdata, client, "Pool full of clients"); client->reject = 2; return json_string("proxy full"); } LOGINFO("Set new subscription %"PRId64" to new enonce1 %lx string %s", client->id, client->enonce1_64, client->enonce1); } else { LOGINFO("Set new subscription %"PRId64" to old matched enonce1 %lx string %s", client->id, client->enonce1_64, client->enonce1); } /* Workbases will exist if sdata->current_workbase is not NULL */ ck_rlock(&sdata->workbase_lock); n2len = sdata->workbases->enonce2varlen; ck_runlock(&sdata->workbase_lock); /* Send a blank sessionid in proxy mode so clients don't think we have * resumed if enonce1 ends up matching on reconnect. */ if (ckp->proxy) { JSON_CPACK(ret, "[[[s,s]],s,i]", "mining.notify", "", client->enonce1, n2len); } else { JSON_CPACK(ret, "[[[s,s]],s,i]", "mining.notify", client->enonce1, client->enonce1, n2len); } client->subscribed = true; return ret; } static bool test_address(ckpool_t *ckp, const char *address) { bool ret = false; char *buf, *msg; ASPRINTF(&msg, "checkaddr:%s", address); /* Must wait for a response here */ buf = __send_recv_generator(ckp, msg, GEN_LAX); dealloc(msg); if (!buf) return ret; ret = cmdmatch(buf, "true"); dealloc(buf); return ret; } static const double nonces = 4294967296; static double dsps_from_key(json_t *val, const char *key) { char *string, *endptr; double ret = 0; json_get_string(&string, val, key); if (!string) return ret; ret = strtod(string, &endptr) / nonces; if (endptr) { switch (endptr[0]) { case 'E': ret *= (double)1000; case 'P': ret *= (double)1000; case 'T': ret *= (double)1000; case 'G': ret *= (double)1000; case 'M': ret *= (double)1000; case 'K': ret *= (double)1000; default: break; } } free(string); return ret; } /* Enter holding a reference count */ static void read_userstats(ckpool_t *ckp, user_instance_t *user) { int tvsec_diff = 0, ret; char s[512]; json_t *val; FILE *fp; tv_t now; snprintf(s, 511, "%s/users/%s", ckp->logdir, user->username); fp = fopen(s, "re"); if (!fp) { LOGINFO("User %s does not have a logfile to read", user->username); return; } memset(s, 0, 512); ret = fread(s, 1, 511, fp); fclose(fp); if (ret < 1) { LOGINFO("Failed to read user %s logfile", user->username); return; } val = json_loads(s, 0, NULL); if (!val) { LOGINFO("Failed to json decode user %s logfile: %s", user->username, s); return; } tv_time(&now); copy_tv(&user->last_share, &now); user->dsps1 = dsps_from_key(val, "hashrate1m"); user->dsps5 = dsps_from_key(val, "hashrate5m"); user->dsps60 = dsps_from_key(val, "hashrate1hr"); user->dsps1440 = dsps_from_key(val, "hashrate1d"); user->dsps10080 = dsps_from_key(val, "hashrate7d"); json_get_int64(&user->last_update.tv_sec, val, "lastupdate"); json_get_double(&user->best_diff, val, "bestshare"); LOGINFO("Successfully read user %s stats %f %f %f %f %f %f", user->username, user->dsps1, user->dsps5, user->dsps60, user->dsps1440, user->dsps10080, user->best_diff); json_decref(val); if (user->last_update.tv_sec) tvsec_diff = now.tv_sec - user->last_update.tv_sec - 60; if (tvsec_diff > 60) { LOGINFO("Old user stats indicate not logged for %d seconds, decaying stats", tvsec_diff); decay_time(&user->dsps1, 0, tvsec_diff, 60); decay_time(&user->dsps5, 0, tvsec_diff, 300); decay_time(&user->dsps60, 0, tvsec_diff, 3600); decay_time(&user->dsps1440, 0, tvsec_diff, 86400); decay_time(&user->dsps10080, 0, tvsec_diff, 604800); } } /* Enter holding a reference count */ static void read_workerstats(ckpool_t *ckp, worker_instance_t *worker) { int tvsec_diff = 0, ret; char s[512]; json_t *val; FILE *fp; tv_t now; snprintf(s, 511, "%s/workers/%s", ckp->logdir, worker->workername); fp = fopen(s, "re"); if (!fp) { LOGINFO("Worker %s does not have a logfile to read", worker->workername); return; } memset(s, 0, 512); ret = fread(s, 1, 511, fp); fclose(fp); if (ret < 1) { LOGINFO("Failed to read worker %s logfile", worker->workername); return; } val = json_loads(s, 0, NULL); if (!val) { LOGINFO("Failed to json decode worker %s logfile: %s", worker->workername, s); return; } tv_time(&now); copy_tv(&worker->last_share, &now); worker->dsps1 = dsps_from_key(val, "hashrate1m"); worker->dsps5 = dsps_from_key(val, "hashrate5m"); worker->dsps60 = dsps_from_key(val, "hashrate1d"); worker->dsps1440 = dsps_from_key(val, "hashrate1d"); json_get_double(&worker->best_diff, val, "bestshare"); json_get_int64(&worker->last_update.tv_sec, val, "lastupdate"); LOGINFO("Successfully read worker %s stats %f %f %f %f %f", worker->workername, worker->dsps1, worker->dsps5, worker->dsps60, worker->dsps1440, worker->best_diff); json_decref(val); if (worker->last_update.tv_sec) tvsec_diff = now.tv_sec - worker->last_update.tv_sec - 60; if (tvsec_diff > 60) { LOGINFO("Old worker stats indicate not logged for %d seconds, decaying stats", tvsec_diff); decay_time(&worker->dsps1, 0, tvsec_diff, 60); decay_time(&worker->dsps5, 0, tvsec_diff, 300); decay_time(&worker->dsps60, 0, tvsec_diff, 3600); decay_time(&worker->dsps1440, 0, tvsec_diff, 86400); } } /* This simply strips off the first part of the workername and matches it to a * user or creates a new one. Needs to be entered with client holding a ref * count. */ static user_instance_t *generate_user(ckpool_t *ckp, stratum_instance_t *client, const char *workername) { char *base_username = strdupa(workername), *username; bool new_user = false, new_worker = false; sdata_t *sdata = ckp->data; worker_instance_t *tmp; user_instance_t *user; int len; username = strsep(&base_username, "._"); if (!username || !strlen(username)) username = base_username; len = strlen(username); if (unlikely(len > 127)) username[127] = '\0'; ck_ilock(&sdata->instance_lock); HASH_FIND_STR(sdata->user_instances, username, user); /* Upgrade to write lock */ ck_ulock(&sdata->instance_lock); if (!user) { /* New user instance. Secondary user id will be NULL */ user = ckzalloc(sizeof(user_instance_t)); user->auth_backoff = 3; /* Set initial backoff to 3 seconds */ strcpy(user->username, username); new_user = true; user->id = sdata->user_instance_id++; HASH_ADD_STR(sdata->user_instances, username, user); } client->user_instance = user; DL_FOREACH(user->worker_instances, tmp) { if (!safecmp(workername, tmp->workername)) { client->worker_instance = tmp; break; } } /* Create one worker instance for combined data from workers of the * same name */ if (!client->worker_instance) { worker_instance_t *worker = ckzalloc(sizeof(worker_instance_t)); worker->workername = strdup(workername); worker->user_instance = user; DL_APPEND(user->worker_instances, worker); new_worker = true; worker->start_time = time(NULL); client->worker_instance = worker; } DL_APPEND(user->clients, client); ck_wunlock(&sdata->instance_lock); if (CKP_STANDALONE(ckp) && new_user) read_userstats(ckp, user); if (CKP_STANDALONE(ckp) && new_worker) read_workerstats(ckp, client->worker_instance); if (new_user && !ckp->proxy) { /* Is this a btc address based username? */ if (len > 26 && len < 35) user->btcaddress = test_address(ckp, username); LOGNOTICE("Added new user %s%s", username, user->btcaddress ? " as address based registration" : ""); } return user; } /* Send this to the database and parse the response to authorise a user * and get SUID parameters back. We don't add these requests to the sdata->ckdbqueue * since we have to wait for the response but this is done from the authoriser * thread so it won't hold anything up but other authorisations. Needs to be * entered with client holding a ref count. */ static int send_recv_auth(stratum_instance_t *client) { user_instance_t *user = client->user_instance; ckpool_t *ckp = client->ckp; sdata_t *sdata = ckp->data; char *buf = NULL, *json_msg; bool contended = false; char cdfield[64]; int ret = 1; json_t *val; ts_t now; ts_realtime(&now); sprintf(cdfield, "%lu,%lu", now.tv_sec, now.tv_nsec); val = json_object(); json_set_string(val, "username", user->username); json_set_string(val, "workername", client->workername); json_set_string(val, "poolinstance", ckp->name); json_set_string(val, "useragent", client->useragent); json_set_int(val, "clientid", client->id); json_set_string(val,"enonce1", client->enonce1); json_set_bool(val, "preauth", false); json_set_string(val, "createdate", cdfield); json_set_string(val, "createby", "code"); json_set_string(val, "createcode", __func__); json_set_string(val, "createinet", client->address); if (user->btcaddress) json_msg = ckdb_msg(ckp, val, ID_ADDRAUTH); else json_msg = ckdb_msg(ckp, val, ID_AUTH); if (unlikely(!json_msg)) { LOGWARNING("Failed to dump json in send_recv_auth"); goto out; } /* We want responses from ckdb serialised and not interleaved with * other requests. Wait up to 3 seconds for exclusive access to ckdb * and if we don't receive it treat it as a delayed auth if possible */ if (likely(!mutex_timedlock(&sdata->ckdb_lock, 3))) { buf = ckdb_msg_call(ckp, json_msg); mutex_unlock(&sdata->ckdb_lock); } else contended = true; free(json_msg); if (likely(buf)) { worker_instance_t *worker = client->worker_instance; char *cmd = NULL, *secondaryuserid = NULL; char response[PAGESIZE] = {}; json_error_t err_val; json_t *val = NULL; LOGINFO("Got ckdb response: %s", buf); if (unlikely(sscanf(buf, "id.%*d.%s", response) < 1 || strlen(response) < 1 || !strchr(response, '='))) { if (cmdmatch(response, "failed")) goto out; LOGWARNING("Got unparseable ckdb auth response: %s", buf); goto out_fail; } cmd = response; strsep(&cmd, "="); LOGINFO("User %s Worker %s got auth response: %s cmd: %s", user->username, client->workername, response, cmd); val = json_loads(cmd, 0, &err_val); if (unlikely(!val)) LOGWARNING("AUTH JSON decode failed(%d): %s", err_val.line, err_val.text); else { json_get_string(&secondaryuserid, val, "secondaryuserid"); json_get_int(&worker->mindiff, val, "difficultydefault"); client->suggest_diff = worker->mindiff; if (!user->auth_time) user->auth_time = time(NULL); } if (secondaryuserid && (!safecmp(response, "ok.authorise") || !safecmp(response, "ok.addrauth"))) { if (!user->secondaryuserid) user->secondaryuserid = secondaryuserid; else dealloc(secondaryuserid); ret = 0; } if (likely(val)) json_decref(val); goto out; } if (contended) LOGWARNING("Prolonged lock contention for ckdb while trying to authorise"); else LOGWARNING("Got no auth response from ckdb :("); out_fail: ret = -1; out: free(buf); return ret; } /* For sending auths to ckdb after we've already decided we can authorise * these clients while ckdb is offline, based on an existing client of the * same username already having been authorised. Needs to be entered with * client holding a ref count. */ static void queue_delayed_auth(stratum_instance_t *client) { ckpool_t *ckp = client->ckp; char cdfield[64]; json_t *val; ts_t now; /* Read off any cached mindiff from previous auths */ client->suggest_diff = client->worker_instance->mindiff; ts_realtime(&now); sprintf(cdfield, "%lu,%lu", now.tv_sec, now.tv_nsec); JSON_CPACK(val, "{ss,ss,ss,ss,sI,ss,sb,ss,ss,ss,ss}", "username", client->user_instance->username, "workername", client->workername, "poolinstance", ckp->name, "useragent", client->useragent, "clientid", client->id, "enonce1", client->enonce1, "preauth", true, "createdate", cdfield, "createby", "code", "createcode", __func__, "createinet", client->address); ckdbq_add(ckp, ID_AUTH, val); } /* Needs to be entered with client holding a ref count. */ static json_t *parse_authorise(stratum_instance_t *client, const json_t *params_val, json_t **err_val, const char *address, int *errnum) { user_instance_t *user; ckpool_t *ckp = client->ckp; bool ret = false; const char *buf; int arr_size; if (unlikely(!json_is_array(params_val))) { *err_val = json_string("params not an array"); goto out; } arr_size = json_array_size(params_val); if (unlikely(arr_size < 1)) { *err_val = json_string("params missing array entries"); goto out; } if (unlikely(!client->useragent)) { *err_val = json_string("Failed subscription"); goto out; } buf = json_string_value(json_array_get(params_val, 0)); if (!buf) { *err_val = json_string("Invalid workername parameter"); goto out; } if (!strlen(buf)) { *err_val = json_string("Empty workername parameter"); goto out; } if (!memcmp(buf, ".", 1) || !memcmp(buf, "_", 1)) { *err_val = json_string("Empty username parameter"); goto out; } if (strchr(buf, '/')) { *err_val = json_string("Invalid character in username"); goto out; } user = generate_user(ckp, client, buf); client->user_id = user->id; strcpy(client->address, address); /* NOTE workername is NULL prior to this so should not be used in code * till after this point */ client->workername = strdup(buf); if (user->failed_authtime) { time_t now_t = time(NULL); if (now_t < user->failed_authtime + user->auth_backoff) { LOGNOTICE("Client %"PRId64" worker %s rate limited due to failed auth attempts", client->id, buf); client->dropped = true; goto out; } } /* NOTE worker count incremented here for any client put onto user's * list until it's dropped */ inc_worker(ckp, user); if (CKP_STANDALONE(ckp)) ret = true; else { /* Preauth workers for the first 10 minutes after the user is * first authorised by ckdb to avoid floods of worker auths. * *errnum is implied zero already so ret will be set true */ if (user->auth_time && time(NULL) - user->auth_time < 600) client->suggest_diff = client->worker_instance->mindiff; else *errnum = send_recv_auth(client); if (!*errnum) ret = true; else if (*errnum < 0 && user->secondaryuserid) { /* This user has already been authorised but ckdb is * offline so we assume they already exist but add the * auth request to the queued messages. */ queue_delayed_auth(client); ret = true; } } if (ret) { client->authorised = ret; user->authorised = ret; if (ckp->proxy) { LOGNOTICE("Authorised client %"PRId64" to proxy %ld:%d, worker %s as user %s", client->id, client->proxyid, client->subproxyid, buf, user->username); } else { LOGNOTICE("Authorised client %"PRId64" worker %s as user %s", client->id, buf, user->username); } user->auth_backoff = 3; /* Reset auth backoff time */ } else { LOGNOTICE("Client %"PRId64" worker %s failed to authorise as user %s", client->id, buf, user->username); user->failed_authtime = time(NULL); user->auth_backoff <<= 1; /* Cap backoff time to 10 mins */ if (user->auth_backoff > 600) user->auth_backoff = 600; } /* We can set this outside of lock safely */ client->authorising = false; out: return json_boolean(ret); } /* Needs to be entered with client holding a ref count. */ static void stratum_send_diff(sdata_t *sdata, const stratum_instance_t *client) { json_t *json_msg; JSON_CPACK(json_msg, "{s[I]soss}", "params", client->diff, "id", json_null(), "method", "mining.set_difficulty"); stratum_add_send(sdata, json_msg, client->id); } /* Needs to be entered with client holding a ref count. */ static void stratum_send_message(sdata_t *sdata, const stratum_instance_t *client, const char *msg) { json_t *json_msg; JSON_CPACK(json_msg, "{sosss[s]}", "id", json_null(), "method", "client.show_message", "params", msg); stratum_add_send(sdata, json_msg, client->id); } static double time_bias(const double tdiff, const double period) { double dexp = tdiff / period; /* Sanity check to prevent silly numbers for double accuracy **/ if (unlikely(dexp > 36)) dexp = 36; return 1.0 - 1.0 / exp(dexp); } /* Sanity check to prevent clock adjustments backwards from screwing up stats */ static double sane_tdiff(tv_t *end, tv_t *start) { double tdiff = tvdiff(end, start); if (unlikely(tdiff < 0.001)) tdiff = 0.001; return tdiff; } /* Needs to be entered with client holding a ref count. */ static void add_submit(ckpool_t *ckp, stratum_instance_t *client, const int diff, const bool valid, const bool submit) { sdata_t *ckp_sdata = ckp->data, *sdata = client->sdata; worker_instance_t *worker = client->worker_instance; double tdiff, bdiff, dsps, drr, network_diff, bias; user_instance_t *user = client->user_instance; int64_t next_blockid, optimal; tv_t now_t; mutex_lock(&ckp_sdata->stats_lock); if (valid) { ckp_sdata->stats.unaccounted_shares++; ckp_sdata->stats.unaccounted_diff_shares += diff; } else ckp_sdata->stats.unaccounted_rejects += diff; mutex_unlock(&ckp_sdata->stats_lock); /* Count only accepted and stale rejects in diff calculation. */ if (!valid && !submit) return; tv_time(&now_t); ck_rlock(&sdata->workbase_lock); next_blockid = sdata->workbase_id + 1; if (ckp->proxy) network_diff = sdata->current_workbase->diff; else network_diff = sdata->current_workbase->network_diff; ck_runlock(&sdata->workbase_lock); if (unlikely(!client->first_share.tv_sec)) { copy_tv(&client->first_share, &now_t); copy_tv(&client->ldc, &now_t); } tdiff = sane_tdiff(&now_t, &client->last_share); decay_time(&client->dsps1, diff, tdiff, 60); decay_time(&client->dsps5, diff, tdiff, 300); decay_time(&client->dsps60, diff, tdiff, 3600); decay_time(&client->dsps1440, diff, tdiff, 86400); decay_time(&client->dsps10080, diff, tdiff, 604800); copy_tv(&client->last_share, &now_t); tdiff = sane_tdiff(&now_t, &worker->last_share); decay_time(&worker->dsps1, diff, tdiff, 60); decay_time(&worker->dsps5, diff, tdiff, 300); decay_time(&worker->dsps60, diff, tdiff, 3600); decay_time(&worker->dsps1440, diff, tdiff, 86400); copy_tv(&worker->last_share, &now_t); worker->idle = false; tdiff = sane_tdiff(&now_t, &user->last_share); decay_time(&user->dsps1, diff, tdiff, 60); decay_time(&user->dsps5, diff, tdiff, 300); decay_time(&user->dsps60, diff, tdiff, 3600); decay_time(&user->dsps1440, diff, tdiff, 86400); decay_time(&user->dsps10080, diff, tdiff, 604800); copy_tv(&user->last_share, &now_t); client->idle = false; client->ssdc++; bdiff = sane_tdiff(&now_t, &client->first_share); bias = time_bias(bdiff, 300); tdiff = sane_tdiff(&now_t, &client->ldc); /* Check the difficulty every 240 seconds or as many shares as we * should have had in that time, whichever comes first. */ if (client->ssdc < 72 && tdiff < 240) return; if (diff != client->diff) { client->ssdc = 0; return; } /* Diff rate ratio */ dsps = client->dsps5 / bias; drr = dsps / (double)client->diff; /* Optimal rate product is 0.3, allow some hysteresis. */ if (drr > 0.15 && drr < 0.4) return; /* Allow slightly lower diffs when users choose their own mindiff */ if (worker->mindiff || client->suggest_diff) { if (drr < 0.5) return; optimal = lround(dsps * 2.4); } else optimal = lround(dsps * 3.33); /* Clamp to mindiff ~ network_diff */ if (optimal < ckp->mindiff) optimal = ckp->mindiff; /* Client suggest diff overrides worker mindiff */ if (client->suggest_diff) { if (optimal < client->suggest_diff) optimal = client->suggest_diff; } else if (optimal < worker->mindiff) optimal = worker->mindiff; if (ckp->maxdiff && optimal > ckp->maxdiff) optimal = ckp->maxdiff; if (optimal > network_diff) optimal = network_diff; if (client->diff == optimal) return; /* If this is the first share in a change, reset the last diff change * to make sure the client hasn't just fallen back after a leave of * absence */ if (optimal < client->diff && client->ssdc == 1) { copy_tv(&client->ldc, &now_t); return; } client->ssdc = 0; LOGINFO("Client %"PRId64" biased dsps %.2f dsps %.2f drr %.2f adjust diff from %"PRId64" to: %"PRId64" ", client->id, dsps, client->dsps5, drr, client->diff, optimal); copy_tv(&client->ldc, &now_t); client->diff_change_job_id = next_blockid; client->old_diff = client->diff; client->diff = optimal; stratum_send_diff(sdata, client); } /* We should already be holding the workbase_lock. Needs to be entered with * client holding a ref count. */ static void test_blocksolve(const stratum_instance_t *client, const workbase_t *wb, const uchar *data, const uchar *hash, const double diff, const char *coinbase, int cblen, const char *nonce2, const char *nonce) { int transactions = wb->transactions + 1; char hexcoinbase[1024], blockhash[68]; sdata_t *sdata = client->sdata; json_t *val = NULL, *val_copy; char *gbt_block, varint[12]; ckpool_t *ckp = wb->ckp; ckmsg_t *block_ckmsg; char cdfield[64]; uchar swap[32]; ts_t ts_now; /* Submit anything over 99% of the diff in case of rounding errors */ if (diff < sdata->current_workbase->network_diff * 0.99) return; LOGWARNING("Possible block solve diff %f !", diff); /* Can't submit a block in proxy mode without the transactions */ if (wb->proxy && wb->merkles) return; ts_realtime(&ts_now); sprintf(cdfield, "%lu,%lu", ts_now.tv_sec, ts_now.tv_nsec); gbt_block = ckalloc(1024); flip_32(swap, hash); __bin2hex(blockhash, swap, 32); /* Message format: "submitblock:hash,data" */ sprintf(gbt_block, "submitblock:%s,", blockhash); __bin2hex(gbt_block + 12 + 64 + 1, data, 80); if (transactions < 0xfd) { uint8_t val8 = transactions; __bin2hex(varint, (const unsigned char *)&val8, 1); } else if (transactions <= 0xffff) { uint16_t val16 = htole16(transactions); strcat(gbt_block, "fd"); __bin2hex(varint, (const unsigned char *)&val16, 2); } else { uint32_t val32 = htole32(transactions); strcat(gbt_block, "fe"); __bin2hex(varint, (const unsigned char *)&val32, 4); } strcat(gbt_block, varint); __bin2hex(hexcoinbase, coinbase, cblen); strcat(gbt_block, hexcoinbase); if (wb->transactions) realloc_strcat(&gbt_block, wb->txn_data); send_generator(ckp, gbt_block, GEN_PRIORITY); free(gbt_block); JSON_CPACK(val, "{si,ss,ss,sI,ss,ss,sI,ss,ss,ss,sI,ss,ss,ss,ss}", "height", wb->height, "blockhash", blockhash, "confirmed", "n", "workinfoid", wb->id, "username", client->user_instance->username, "workername", client->workername, "clientid", client->id, "enonce1", client->enonce1, "nonce2", nonce2, "nonce", nonce, "reward", wb->coinbasevalue, "createdate", cdfield, "createby", "code", "createcode", __func__, "createinet", ckp->serverurl[client->server]); val_copy = json_deep_copy(val); block_ckmsg = ckalloc(sizeof(ckmsg_t)); block_ckmsg->data = val_copy; mutex_lock(&sdata->block_lock); DL_APPEND(sdata->block_solves, block_ckmsg); mutex_unlock(&sdata->block_lock); ckdbq_add(ckp, ID_BLOCK, val); } /* Needs to be entered with client holding a ref count. */ static double submission_diff(const stratum_instance_t *client, const workbase_t *wb, const char *nonce2, const uint32_t ntime32, const char *nonce, uchar *hash) { unsigned char merkle_root[32], merkle_sha[64]; uint32_t *data32, *swap32, benonce32; char *coinbase, data[80]; uchar swap[80], hash1[32]; int cblen, i; double ret; coinbase = alloca(wb->coinb1len + wb->enonce1constlen + wb->enonce1varlen + wb->enonce2varlen + wb->coinb2len); memcpy(coinbase, wb->coinb1bin, wb->coinb1len); cblen = wb->coinb1len; memcpy(coinbase + cblen, &client->enonce1bin, wb->enonce1constlen + wb->enonce1varlen); cblen += wb->enonce1constlen + wb->enonce1varlen; hex2bin(coinbase + cblen, nonce2, wb->enonce2varlen); cblen += wb->enonce2varlen; memcpy(coinbase + cblen, wb->coinb2bin, wb->coinb2len); cblen += wb->coinb2len; gen_hash((uchar *)coinbase, merkle_root, cblen); memcpy(merkle_sha, merkle_root, 32); for (i = 0; i < wb->merkles; i++) { memcpy(merkle_sha + 32, &wb->merklebin[i], 32); gen_hash(merkle_sha, merkle_root, 64); memcpy(merkle_sha, merkle_root, 32); } data32 = (uint32_t *)merkle_sha; swap32 = (uint32_t *)merkle_root; flip_32(swap32, data32); /* Copy the cached header binary and insert the merkle root */ memcpy(data, wb->headerbin, 80); memcpy(data + 36, merkle_root, 32); /* Insert the nonce value into the data */ hex2bin(&benonce32, nonce, 4); data32 = (uint32_t *)(data + 64 + 12); *data32 = benonce32; /* Insert the ntime value into the data */ data32 = (uint32_t *)(data + 68); *data32 = htobe32(ntime32); /* Hash the share */ data32 = (uint32_t *)data; swap32 = (uint32_t *)swap; flip_80(swap32, data32); sha256(swap, 80, hash1); sha256(hash1, 32, hash); /* Calculate the diff of the share here */ ret = diff_from_target(hash); /* Test we haven't solved a block regardless of share status */ test_blocksolve(client, wb, swap, hash, ret, coinbase, cblen, nonce2, nonce); return ret; } /* Optimised for the common case where shares are new */ static bool new_share(sdata_t *sdata, const uchar *hash, const int64_t wb_id) { share_t *share = ckzalloc(sizeof(share_t)), *match = NULL; bool ret = true; memcpy(share->hash, hash, 32); share->workbase_id = wb_id; mutex_lock(&sdata->share_lock); sdata->shares_generated++; HASH_FIND(hh, sdata->shares, hash, 32, match); if (likely(!match)) HASH_ADD(hh, sdata->shares, hash, 32, share); mutex_unlock(&sdata->share_lock); if (unlikely(match)) { dealloc(share); ret = false; } return ret; } static void update_client(const stratum_instance_t *client, const int64_t client_id); /* Submit a share in proxy mode to the parent pool. workbase_lock is held. * Needs to be entered with client holding a ref count. */ static void submit_share(stratum_instance_t *client, const int64_t jobid, const char *nonce2, const char *ntime, const char *nonce) { ckpool_t *ckp = client->ckp; json_t *json_msg; char enonce2[32]; char *msg; sprintf(enonce2, "%s%s", client->enonce1var, nonce2); JSON_CPACK(json_msg, "{sIsssssssIsIsi}", "jobid", jobid, "nonce2", enonce2, "ntime", ntime, "nonce", nonce, "client_id", client->id, "proxy", client->proxyid, "subproxy", client->subproxyid); msg = json_dumps(json_msg, 0); json_decref(json_msg); send_generator(ckp, msg, GEN_LAX); free(msg); } #define JSON_ERR(err) json_string(SHARE_ERR(err)) /* Needs to be entered with client holding a ref count. */ static json_t *parse_submit(stratum_instance_t *client, json_t *json_msg, const json_t *params_val, json_t **err_val) { bool share = false, result = false, invalid = true, submit = false; user_instance_t *user = client->user_instance; double diff = client->diff, wdiff = 0, sdiff = -1; char hexhash[68] = {}, sharehash[32], cdfield[64]; const char *workername, *job_id, *ntime, *nonce; char *fname = NULL, *s, *nonce2; sdata_t *sdata = client->sdata; enum share_err err = SE_NONE; ckpool_t *ckp = client->ckp; char idstring[20] = {}; workbase_t *wb = NULL; uint32_t ntime32; uchar hash[32]; int nlen, len; time_t now_t; json_t *val; int64_t id; ts_t now; FILE *fp; ts_realtime(&now); now_t = now.tv_sec; sprintf(cdfield, "%lu,%lu", now.tv_sec, now.tv_nsec); if (unlikely(!json_is_array(params_val))) { err = SE_NOT_ARRAY; *err_val = JSON_ERR(err); goto out; } if (unlikely(json_array_size(params_val) != 5)) { err = SE_INVALID_SIZE; *err_val = JSON_ERR(err); goto out; } workername = json_string_value(json_array_get(params_val, 0)); if (unlikely(!workername || !strlen(workername))) { err = SE_NO_USERNAME; *err_val = JSON_ERR(err); goto out; } job_id = json_string_value(json_array_get(params_val, 1)); if (unlikely(!job_id || !strlen(job_id))) { err = SE_NO_JOBID; *err_val = JSON_ERR(err); goto out; } nonce2 = (char *)json_string_value(json_array_get(params_val, 2)); if (unlikely(!nonce2 || !strlen(nonce2))) { err = SE_NO_NONCE2; *err_val = JSON_ERR(err); goto out; } ntime = json_string_value(json_array_get(params_val, 3)); if (unlikely(!ntime || !strlen(ntime))) { err = SE_NO_NTIME; *err_val = JSON_ERR(err); goto out; } nonce = json_string_value(json_array_get(params_val, 4)); if (unlikely(!nonce || !strlen(nonce))) { err = SE_NO_NONCE; *err_val = JSON_ERR(err); goto out; } if (safecmp(workername, client->workername)) { err = SE_WORKER_MISMATCH; *err_val = JSON_ERR(err); goto out; } sscanf(job_id, "%lx", &id); sscanf(ntime, "%x", &ntime32); share = true; ck_rlock(&sdata->workbase_lock); HASH_FIND_I64(sdata->workbases, &id, wb); if (unlikely(!wb)) { if (!sdata->current_workbase) return json_boolean(false); id = sdata->current_workbase->id; err = SE_INVALID_JOBID; json_set_string(json_msg, "reject-reason", SHARE_ERR(err)); strncpy(idstring, job_id, 19); ASPRINTF(&fname, "%s.sharelog", sdata->current_workbase->logdir); goto out_unlock; } wdiff = wb->diff; strncpy(idstring, wb->idstring, 19); ASPRINTF(&fname, "%s.sharelog", wb->logdir); /* Fix broken clients sending too many chars. Nonce2 is part of the * read only json so use a temporary variable and modify it. */ len = wb->enonce2varlen * 2; nlen = strlen(nonce2); if (nlen > len) { nonce2 = strdupa(nonce2); nonce2[len] = '\0'; } else if (nlen < len) { char *tmp = nonce2; nonce2 = strdupa("0000000000000000"); memcpy(nonce2, tmp, nlen); nonce2[len] = '\0'; } sdiff = submission_diff(client, wb, nonce2, ntime32, nonce, hash); if (sdiff > client->best_diff) { worker_instance_t *worker = client->worker_instance; client->best_diff = sdiff; LOGINFO("User %s worker %s client %"PRId64" new best diff %lf", user->username, worker->workername, client->id, sdiff); if (sdiff > worker->best_diff) worker->best_diff = sdiff; if (sdiff > user->best_diff) user->best_diff = sdiff; } bswap_256(sharehash, hash); __bin2hex(hexhash, sharehash, 32); if (id < sdata->blockchange_id) { err = SE_STALE; json_set_string(json_msg, "reject-reason", SHARE_ERR(err)); goto out_submit; } /* Ntime cannot be less, but allow forward ntime rolling up to max */ if (ntime32 < wb->ntime32 || ntime32 > wb->ntime32 + 7000) { err = SE_NTIME_INVALID; json_set_string(json_msg, "reject-reason", SHARE_ERR(err)); goto out_unlock; } invalid = false; out_submit: if (sdiff >= wdiff) submit = true; out_unlock: ck_runlock(&sdata->workbase_lock); /* Accept the lower of new and old diffs until the next update */ if (id < client->diff_change_job_id && client->old_diff < client->diff) diff = client->old_diff; if (!invalid) { char wdiffsuffix[16]; suffix_string(wdiff, wdiffsuffix, 16, 0); if (sdiff >= diff) { if (new_share(sdata, hash, id)) { LOGINFO("Accepted client %"PRId64" share diff %.1f/%.0f/%s: %s", client->id, sdiff, diff, wdiffsuffix, hexhash); result = true; } else { err = SE_DUPE; json_set_string(json_msg, "reject-reason", SHARE_ERR(err)); LOGINFO("Rejected client %"PRId64" dupe diff %.1f/%.0f/%s: %s", client->id, sdiff, diff, wdiffsuffix, hexhash); submit = false; } } else { err = SE_HIGH_DIFF; LOGINFO("Rejected client %"PRId64" high diff %.1f/%.0f/%s: %s", client->id, sdiff, diff, wdiffsuffix, hexhash); json_set_string(json_msg, "reject-reason", SHARE_ERR(err)); submit = false; } } else LOGINFO("Rejected client %"PRId64" invalid share", client->id); /* Submit share to upstream pool in proxy mode. We submit valid and * stale shares and filter out the rest. */ if (wb && wb->proxy && submit) { LOGINFO("Submitting share upstream: %s", hexhash); submit_share(client, id, nonce2, ntime, nonce); } add_submit(ckp, client, diff, result, submit); /* Now write to the pool's sharelog. */ val = json_object(); json_set_int(val, "workinfoid", id); json_set_int(val, "clientid", client->id); json_set_string(val, "enonce1", client->enonce1); if (!CKP_STANDALONE(ckp)) json_set_string(val, "secondaryuserid", user->secondaryuserid); json_set_string(val, "nonce2", nonce2); json_set_string(val, "nonce", nonce); json_set_string(val, "ntime", ntime); json_set_double(val, "diff", diff); json_set_double(val, "sdiff", sdiff); json_set_string(val, "hash", hexhash); json_set_bool(val, "result", result); json_object_set(val, "reject-reason", json_object_get(json_msg, "reject-reason")); json_object_set(val, "error", *err_val); json_set_int(val, "errn", err); json_set_string(val, "createdate", cdfield); json_set_string(val, "createby", "code"); json_set_string(val, "createcode", __func__); json_set_string(val, "createinet", ckp->serverurl[client->server]); json_set_string(val, "workername", client->workername); json_set_string(val, "username", user->username); if (ckp->logshares) { fp = fopen(fname, "ae"); if (likely(fp)) { s = json_dumps(val, JSON_EOL); len = strlen(s); len = fprintf(fp, "%s", s); free(s); fclose(fp); if (unlikely(len < 0)) LOGERR("Failed to fwrite to %s", fname); } else LOGERR("Failed to fopen %s", fname); } ckdbq_add(ckp, ID_SHARES, val); out: if ((!result && !submit) || !share) { /* Is this the first in a run of invalids? */ if (client->first_invalid < client->last_share.tv_sec || !client->first_invalid) client->first_invalid = now_t; else if (client->first_invalid && client->first_invalid < now_t - 120) { LOGNOTICE("Client %"PRId64" rejecting for 120s, disconnecting", client->id); stratum_send_message(sdata, client, "Disconnecting for continuous invalid shares"); client->reject = 2; } else if (client->first_invalid && client->first_invalid < now_t - 60) { if (!client->reject) { LOGINFO("Client %"PRId64" rejecting for 60s, sending update", client->id); update_client(client, client->id); client->reject = 1; } } } else { client->first_invalid = 0; client->reject = 0; } if (!share) { val = json_object(); json_set_int(val, "clientid", client->id); if (!CKP_STANDALONE(ckp)) json_set_string(val, "secondaryuserid", user->secondaryuserid); json_set_string(val, "enonce1", client->enonce1); json_set_int(val, "workinfoid", sdata->current_workbase->id); json_set_string(val, "workername", client->workername); json_set_string(val, "username", user->username); json_object_set(val, "error", *err_val); json_set_int(val, "errn", err); json_set_string(val, "createdate", cdfield); json_set_string(val, "createby", "code"); json_set_string(val, "createcode", __func__); json_set_string(val, "createinet", ckp->serverurl[client->server]); ckdbq_add(ckp, ID_SHAREERR, val); LOGINFO("Invalid share from client %"PRId64": %s", client->id, client->workername); } free(fname); return json_boolean(result); } /* Must enter with workbase_lock held */ static json_t *__stratum_notify(const workbase_t *wb, const bool clean) { json_t *val; JSON_CPACK(val, "{s:[ssssosssb],s:o,s:s}", "params", wb->idstring, wb->prevhash, wb->coinb1, wb->coinb2, json_deep_copy(wb->merkle_array), wb->bbversion, wb->nbit, wb->ntime, clean, "id", json_null(), "method", "mining.notify"); return val; } static void stratum_broadcast_update(sdata_t *sdata, const workbase_t *wb, const bool clean) { json_t *json_msg; ck_rlock(&sdata->workbase_lock); json_msg = __stratum_notify(wb, clean); ck_runlock(&sdata->workbase_lock); stratum_broadcast(sdata, json_msg); } /* For sending a single stratum template update */ static void stratum_send_update(sdata_t *sdata, const int64_t client_id, const bool clean) { json_t *json_msg; if (unlikely(!sdata->current_workbase)) { LOGWARNING("No current workbase to send stratum update"); return; } ck_rlock(&sdata->workbase_lock); json_msg = __stratum_notify(sdata->current_workbase, clean); ck_runlock(&sdata->workbase_lock); stratum_add_send(sdata, json_msg, client_id); } static void send_json_err(sdata_t *sdata, const int64_t client_id, json_t *id_val, const char *err_msg) { json_t *val; JSON_CPACK(val, "{soss}", "id", json_copy(id_val), "error", err_msg); stratum_add_send(sdata, val, client_id); } /* Needs to be entered with client holding a ref count. */ static void update_client(const stratum_instance_t *client, const int64_t client_id) { sdata_t *sdata = client->sdata; stratum_send_update(sdata, client_id, true); stratum_send_diff(sdata, client); } static json_params_t *create_json_params(const int64_t client_id, const json_t *method, const json_t *params, const json_t *id_val, const char *address) { json_params_t *jp = ckalloc(sizeof(json_params_t)); jp->method = json_deep_copy(method); jp->params = json_deep_copy(params); jp->id_val = json_deep_copy(id_val); jp->client_id = client_id; strcpy(jp->address, address); return jp; } static void set_worker_mindiff(ckpool_t *ckp, const char *workername, int mindiff) { worker_instance_t *worker = NULL, *tmp; char *username = strdupa(workername), *ignore; user_instance_t *user = NULL; stratum_instance_t *client; sdata_t *sdata = ckp->data; ignore = username; strsep(&ignore, "._"); /* Find the user first */ ck_rlock(&sdata->instance_lock); HASH_FIND_STR(sdata->user_instances, username, user); ck_runlock(&sdata->instance_lock); /* They may just have not connected yet */ if (!user) { LOGINFO("Failed to find user %s in set_worker_mindiff", username); return; } /* Then find the matching worker user */ ck_rlock(&sdata->instance_lock); DL_FOREACH(user->worker_instances, tmp) { if (!safecmp(workername, tmp->workername)) { worker = tmp; break; } } ck_runlock(&sdata->instance_lock); /* They may just not be connected at the moment */ if (!worker) { LOGINFO("Failed to find worker %s in set_worker_mindiff", workername); return; } if (mindiff < 1) { LOGINFO("Worker %s requested invalid diff %d", worker->workername, mindiff); return; } if (mindiff < ckp->mindiff) mindiff = ckp->mindiff; if (mindiff == worker->mindiff) return; worker->mindiff = mindiff; /* Iterate over all the workers from this user to find any with the * matching worker that are currently live and send them a new diff * if we can. Otherwise it will only act as a clamp on next share * submission. */ ck_rlock(&sdata->instance_lock); DL_FOREACH(user->clients, client) { if (client->worker_instance != worker) continue; /* Per connection suggest diff overrides worker mindiff ugh */ if (mindiff < client->suggest_diff) continue; if (mindiff == client->diff) continue; client->diff = mindiff; stratum_send_diff(sdata, client); } ck_runlock(&sdata->instance_lock); } /* Implement support for the diff in the params as well as the originally * documented form of placing diff within the method. Needs to be entered with * client holding a ref count. */ static void suggest_diff(stratum_instance_t *client, const char *method, const json_t *params_val) { json_t *arr_val = json_array_get(params_val, 0); sdata_t *sdata = client->ckp->data; int64_t sdiff; if (unlikely(!client_active(client))) { LOGWARNING("Attempted to suggest diff on unauthorised client %"PRId64, client->id); return; } if (arr_val && json_is_integer(arr_val)) sdiff = json_integer_value(arr_val); else if (sscanf(method, "mining.suggest_difficulty(%"PRId64, &sdiff) != 1) { LOGINFO("Failed to parse suggest_difficulty for client %"PRId64, client->id); return; } if (sdiff == client->suggest_diff) return; client->suggest_diff = sdiff; if (client->diff == sdiff) return; if (sdiff < client->ckp->mindiff) client->diff = client->ckp->mindiff; else client->diff = sdiff; stratum_send_diff(sdata, client); } /* Enter with client holding ref count */ static void parse_method(sdata_t *sdata, stratum_instance_t *client, const int64_t client_id, json_t *id_val, json_t *method_val, json_t *params_val, const char *address) { const char *method; /* Random broken clients send something not an integer as the id so we * copy the json item for id_val as is for the response. By far the * most common messages will be shares so look for those first */ method = json_string_value(method_val); if (likely(cmdmatch(method, "mining.submit") && client->authorised)) { json_params_t *jp = create_json_params(client_id, method_val, params_val, id_val, address); ckmsgq_add(sdata->sshareq, jp); return; } if (cmdmatch(method, "mining.subscribe")) { json_t *val, *result_val; if (unlikely(client->subscribed)) { LOGNOTICE("Client %"PRId64" trying to subscribe twice", client_id); return; } result_val = parse_subscribe(client, client_id, params_val); /* Shouldn't happen, sanity check */ if (unlikely(!result_val)) { LOGWARNING("parse_subscribe returned NULL result_val"); return; } val = json_object(); json_object_set_new_nocheck(val, "result", result_val); json_object_set_nocheck(val, "id", id_val); json_object_set_new_nocheck(val, "error", json_null()); stratum_add_send(sdata, val, client_id); if (likely(client->subscribed)) update_client(client, client_id); return; } if (unlikely(cmdmatch(method, "mining.passthrough"))) { char buf[256]; LOGNOTICE("Adding passthrough client %"PRId64, client_id); /* We need to inform the connector process that this client * is a passthrough and to manage its messages accordingly. * The client_id stays on the list but we won't send anything * to it since it's unauthorised. Set the flag just in case. */ client->authorised = false; snprintf(buf, 255, "passthrough=%"PRId64, client_id); send_proc(client->ckp->connector, buf); return; } /* We should only accept subscribed requests from here on */ if (!client->subscribed) { LOGINFO("Dropping unsubscribed client %"PRId64, client_id); connector_drop_client(client->ckp, client_id); return; } if (cmdmatch(method, "mining.auth")) { json_params_t *jp; if (unlikely(client->authorised)) { LOGNOTICE("Client %"PRId64" trying to authorise twice", client_id); return; } jp = create_json_params(client_id, method_val, params_val, id_val, address); ckmsgq_add(sdata->sauthq, jp); return; } /* We should only accept authorised requests from here on */ if (!client->authorised) { /* Dropping unauthorised clients here also allows the * stratifier process to restart since it will have lost all * the stratum instance data. Clients will just reconnect. */ LOGINFO("Dropping unauthorised client %"PRId64, client_id); connector_drop_client(client->ckp, client_id); return; } if (cmdmatch(method, "mining.suggest")) { suggest_diff(client, method, params_val); return; } /* Covers both get_transactions and get_txnhashes */ if (cmdmatch(method, "mining.get")) { json_params_t *jp = create_json_params(client_id, method_val, params_val, id_val, address); ckmsgq_add(sdata->stxnq, jp); return; } /* Unhandled message here */ LOGINFO("Unhandled client %"PRId64" method %s", client_id, method); return; } static void free_smsg(smsg_t *msg) { json_decref(msg->json_msg); free(msg); } /* Entered with client holding ref count */ static void parse_instance_msg(ckpool_t *ckp, sdata_t *sdata, smsg_t *msg, stratum_instance_t *client) { json_t *val = msg->json_msg, *id_val, *method, *params; int64_t client_id = msg->client_id; if (client->reject == 2 || (client->reconnect_request && time(NULL) - client->reconnect_request > 60)) { LOGINFO("Dropping client %"PRId64" tagged for lazy invalidation", client_id); connector_drop_client(ckp, client_id); goto out; } /* Return back the same id_val even if it's null or not existent. */ id_val = json_object_get(val, "id"); method = json_object_get(val, "method"); if (unlikely(!method)) { json_t *res_val = json_object_get(val, "result"); /* Is this a spurious result or ping response? */ if (res_val) { const char *result = json_string_value(res_val); if (!safecmp(result, "pong")) LOGDEBUG("Received pong from client %"PRId64, client_id); else LOGDEBUG("Received spurious response %s from client %"PRId64, result ? result : "", client_id); goto out; } send_json_err(sdata, client_id, id_val, "-3:method not found"); goto out; } if (unlikely(!json_is_string(method))) { send_json_err(sdata, client_id, id_val, "-1:method is not string"); goto out; } params = json_object_get(val, "params"); if (unlikely(!params)) { send_json_err(sdata, client_id, id_val, "-1:params not found"); goto out; } parse_method(sdata, client, client_id, id_val, method, params, msg->address); out: free_smsg(msg); } static void srecv_process(ckpool_t *ckp, char *buf) { bool noid = false, dropped = false; sdata_t *sdata = ckp->data; stratum_instance_t *client; smsg_t *msg; json_t *val; int server; val = json_loads(buf, 0, NULL); if (unlikely(!val)) { LOGWARNING("Received unrecognised non-json message: %s", buf); goto out; } msg = ckzalloc(sizeof(smsg_t)); msg->json_msg = val; val = json_object_get(msg->json_msg, "client_id"); if (unlikely(!val)) { LOGWARNING("Failed to extract client_id from connector json smsg %s", buf); json_decref(msg->json_msg); free(msg); goto out; } msg->client_id = json_integer_value(val); json_object_clear(val); val = json_object_get(msg->json_msg, "address"); if (unlikely(!val)) { LOGWARNING("Failed to extract address from connector json smsg %s", buf); json_decref(msg->json_msg); free(msg); goto out; } strcpy(msg->address, json_string_value(val)); json_object_clear(val); val = json_object_get(msg->json_msg, "server"); if (unlikely(!val)) { LOGWARNING("Failed to extract server from connector json smsg %s", buf); json_decref(msg->json_msg); free(msg); goto out; } server = json_integer_value(val); json_object_clear(val); /* Parse the message here */ ck_ilock(&sdata->instance_lock); client = __instance_by_id(sdata, msg->client_id); /* Upgrade to write lock */ ck_ulock(&sdata->instance_lock); /* If client_id instance doesn't exist yet, create one */ if (unlikely(!client)) { if (likely(!__dropped_instance(sdata, msg->client_id))) { noid = true; client = __stratum_add_instance(ckp, msg->client_id, server); } else dropped = true; } else if (unlikely(client->dropped)) dropped = true; if (likely(!dropped)) __inc_instance_ref(client); ck_wunlock(&sdata->instance_lock); if (unlikely(dropped)) { /* Client may be NULL here */ LOGNOTICE("Stratifier skipped dropped instance %"PRId64" message from server %d", msg->client_id, server); connector_drop_client(ckp, msg->client_id); free_smsg(msg); goto out; } if (unlikely(noid)) LOGINFO("Stratifier added instance %"PRId64" server %d", client->id, server); parse_instance_msg(ckp, sdata, msg, client); /* The client is still active but has been issued a reconnect request * so use this opportunity to send it a reconnect message */ if (unlikely(client->reconnect)) lazy_reconnect_client(sdata, client); dec_instance_ref(sdata, client); out: free(buf); } static void ssend_process(ckpool_t *ckp, smsg_t *msg) { char *s; if (unlikely(!msg->json_msg)) { LOGERR("Sent null json msg to stratum_sender"); free(msg); return; } /* Add client_id to the json message and send it to the * connector process to be delivered */ json_object_set_new_nocheck(msg->json_msg, "client_id", json_integer(msg->client_id)); s = json_dumps(msg->json_msg, 0); send_proc(ckp->connector, s); free(s); free_smsg(msg); } static void discard_json_params(json_params_t *jp) { json_decref(jp->method); json_decref(jp->params); json_decref(jp->id_val); free(jp); } static void sshare_process(ckpool_t *ckp, json_params_t *jp) { json_t *result_val, *json_msg, *err_val = NULL; stratum_instance_t *client; sdata_t *sdata = ckp->data; int64_t client_id; client_id = jp->client_id; client = ref_instance_by_id(sdata, client_id); if (unlikely(!client)) { LOGINFO("Share processor failed to find client id %"PRId64" in hashtable!", client_id); goto out; } if (unlikely(!client->authorised)) { LOGDEBUG("Client %"PRId64" no longer authorised to submit shares", client_id); goto out_decref; } json_msg = json_object(); result_val = parse_submit(client, json_msg, jp->params, &err_val); json_object_set_new_nocheck(json_msg, "result", result_val); json_object_set_new_nocheck(json_msg, "error", err_val ? err_val : json_null()); json_object_set_nocheck(json_msg, "id", jp->id_val); stratum_add_send(sdata, json_msg, client_id); out_decref: dec_instance_ref(sdata, client); out: discard_json_params(jp); } /* As ref_instance_by_id but only returns clients not authorising or authorised, * and sets the authorising flag */ static stratum_instance_t *preauth_ref_instance_by_id(sdata_t *sdata, const int64_t id) { stratum_instance_t *client; ck_ilock(&sdata->instance_lock); client = __instance_by_id(sdata, id); if (client) { if (client->dropped || client->authorising || client->authorised) client = NULL; else { /* Upgrade to write lock to modify client data */ ck_ulock(&sdata->instance_lock); __inc_instance_ref(client); client->authorising = true; ck_dwilock(&sdata->instance_lock); } } ck_uilock(&sdata->instance_lock); return client; } static void sauth_process(ckpool_t *ckp, json_params_t *jp) { json_t *result_val, *json_msg, *err_val = NULL; stratum_instance_t *client; sdata_t *sdata = ckp->data; int mindiff, errnum = 0; int64_t client_id; client_id = jp->client_id; client = preauth_ref_instance_by_id(sdata, client_id); if (unlikely(!client)) { LOGINFO("Authoriser failed to find client id %"PRId64" in hashtable!", client_id); goto out; } result_val = parse_authorise(client, jp->params, &err_val, jp->address, &errnum); if (json_is_true(result_val)) { char *buf; ASPRINTF(&buf, "Authorised, welcome to %s %s!", ckp->name, client->user_instance->username); stratum_send_message(sdata, client, buf); free(buf); } else { if (errnum < 0) stratum_send_message(sdata, client, "Authorisations temporarily offline :("); else stratum_send_message(sdata, client, "Failed authorisation :("); } json_msg = json_object(); json_object_set_new_nocheck(json_msg, "result", result_val); json_object_set_new_nocheck(json_msg, "error", err_val ? err_val : json_null()); json_object_set_nocheck(json_msg, "id", jp->id_val); stratum_add_send(sdata, json_msg, client_id); if (!json_is_true(result_val) || !client->suggest_diff) goto out; /* Update the client now if they have set a valid mindiff different * from the startdiff */ mindiff = MAX(ckp->mindiff, client->suggest_diff); if (mindiff != client->diff) { client->diff = mindiff; stratum_send_diff(sdata, client); } out: if (client) dec_instance_ref(sdata, client); discard_json_params(jp); } static void parse_ckdb_cmd(ckpool_t *ckp, const char *cmd) { json_t *val, *res_val, *arr_val; json_error_t err_val; size_t index; val = json_loads(cmd, 0, &err_val); if (unlikely(!val)) { LOGWARNING("CKDB MSG %s JSON decode failed(%d): %s", cmd, err_val.line, err_val.text); return; } res_val = json_object_get(val, "diffchange"); json_array_foreach(res_val, index, arr_val) { char *workername; int mindiff; json_get_string(&workername, arr_val, "workername"); if (!workername) continue; json_get_int(&mindiff, arr_val, "difficultydefault"); set_worker_mindiff(ckp, workername, mindiff); dealloc(workername); } json_decref(val); } /* Test a value under lock and set it, returning the original value */ static bool test_and_set(bool *val, mutex_t *lock) { bool ret; mutex_lock(lock); ret = *val; *val = true; mutex_unlock(lock); return ret; } static bool test_and_clear(bool *val, mutex_t *lock) { bool ret; mutex_lock(lock); ret = *val; *val = false; mutex_unlock(lock); return ret; } static void ckdbq_process(ckpool_t *ckp, char *msg) { sdata_t *sdata = ckp->data; char *buf = NULL; while (!buf) { mutex_lock(&sdata->ckdb_lock); buf = ckdb_msg_call(ckp, msg); mutex_unlock(&sdata->ckdb_lock); if (unlikely(!buf)) { if (!test_and_set(&sdata->ckdb_offline, &sdata->ckdb_lock)) LOGWARNING("Failed to talk to ckdb, queueing messages"); sleep(5); } } free(msg); if (test_and_clear(&sdata->ckdb_offline, &sdata->ckdb_lock)) LOGWARNING("Successfully resumed talking to ckdb"); /* Process any requests from ckdb that are heartbeat responses with * specific requests. */ if (likely(buf)) { char response[PAGESIZE] = {}; sscanf(buf, "id.%*d.%s", response); if (safecmp(response, "ok")) { char *cmd; cmd = response; strsep(&cmd, "."); LOGDEBUG("Got ckdb response: %s cmd %s", response, cmd); if (cmdmatch(cmd, "heartbeat=")) { strsep(&cmd, "="); parse_ckdb_cmd(ckp, cmd); } } else LOGWARNING("Got failed ckdb response: %s", buf); free(buf); } } static int transactions_by_jobid(sdata_t *sdata, const int64_t id) { workbase_t *wb; int ret = -1; ck_rlock(&sdata->workbase_lock); HASH_FIND_I64(sdata->workbases, &id, wb); if (wb) ret = wb->transactions; ck_runlock(&sdata->workbase_lock); return ret; } static json_t *txnhashes_by_jobid(sdata_t *sdata, const int64_t id) { json_t *ret = NULL; workbase_t *wb; ck_rlock(&sdata->workbase_lock); HASH_FIND_I64(sdata->workbases, &id, wb); if (wb) ret = json_string(wb->txn_hashes); ck_runlock(&sdata->workbase_lock); return ret; } static void send_transactions(ckpool_t *ckp, json_params_t *jp) { const char *msg = json_string_value(jp->method), *params = json_string_value(json_array_get(jp->params, 0)); stratum_instance_t *client = NULL; sdata_t *sdata = ckp->data; json_t *val, *hashes; int64_t job_id = 0; time_t now_t; if (unlikely(!msg || !strlen(msg))) { LOGWARNING("send_transactions received null method"); goto out; } val = json_object(); json_object_set_nocheck(val, "id", jp->id_val); if (cmdmatch(msg, "mining.get_transactions")) { int txns; /* We don't actually send the transactions as that would use * up huge bandwidth, so we just return the number of * transactions :) . Support both forms of encoding the * request in method name and as a parameter. */ if (params && strlen(params) > 0) sscanf(params, "%lx", &job_id); else sscanf(msg, "mining.get_transactions(%lx", &job_id); txns = transactions_by_jobid(sdata, job_id); if (txns != -1) { json_set_int(val, "result", txns); json_object_set_new_nocheck(val, "error", json_null()); } else json_set_string(val, "error", "Invalid job_id"); goto out_send; } if (!cmdmatch(msg, "mining.get_txnhashes")) { LOGDEBUG("Unhandled mining get request: %s", msg); json_set_string(val, "error", "Unhandled"); goto out_send; } client = ref_instance_by_id(sdata, jp->client_id); if (unlikely(!client)) { LOGINFO("send_transactions failed to find client id %"PRId64" in hashtable!", jp->client_id); goto out; } now_t = time(NULL); if (now_t - client->last_txns < ckp->update_interval) { LOGNOTICE("Rate limiting get_txnhashes on client %"PRId64"!", jp->client_id); json_set_string(val, "error", "Ratelimit"); goto out_send; } client->last_txns = now_t; if (!params || !strlen(params)) { json_set_string(val, "error", "Invalid params"); goto out_send; } sscanf(params, "%lx", &job_id); hashes = txnhashes_by_jobid(sdata, job_id); if (hashes) { json_object_set_new_nocheck(val, "result", hashes); json_object_set_new_nocheck(val, "error", json_null()); } else json_set_string(val, "error", "Invalid job_id"); out_send: stratum_add_send(sdata, val, jp->client_id); out: if (client) dec_instance_ref(sdata, client); discard_json_params(jp); } /* Called 32 times per min, we send the updated stats to ckdb of those users * who have gone 1 minute between updates. This ends up staggering stats to * avoid floods of stat data coming at once. */ static void update_workerstats(ckpool_t *ckp, sdata_t *sdata) { json_entry_t *json_list = NULL, *entry, *tmpentry; user_instance_t *user, *tmp; char cdfield[64]; time_t now_t; ts_t ts_now; if (sdata->ckdb_offline) { LOGDEBUG("Not queueing workerstats due to ckdb offline"); return; } if (++sdata->stats.userstats_cycle > 0x1f) sdata->stats.userstats_cycle = 0; ts_realtime(&ts_now); sprintf(cdfield, "%lu,%lu", ts_now.tv_sec, ts_now.tv_nsec); now_t = ts_now.tv_sec; ck_rlock(&sdata->instance_lock); HASH_ITER(hh, sdata->user_instances, user, tmp) { worker_instance_t *worker; uint8_t cycle_mask; if (!user->authorised) continue; /* Select users using a mask to return each user's stats once * every ~10 minutes */ cycle_mask = user->id & 0x1f; if (cycle_mask != sdata->stats.userstats_cycle) continue; DL_FOREACH(user->worker_instances, worker) { double ghs1, ghs5, ghs60, ghs1440; json_t *val; int elapsed; /* Send one lot of stats once the worker is idle if * they have submitted no shares in the last 10 minutes * with the idle bool set. */ if (worker->idle && worker->notified_idle) continue; elapsed = now_t - worker->start_time; ghs1 = worker->dsps1 * nonces; ghs5 = worker->dsps5 * nonces; ghs60 = worker->dsps60 * nonces; ghs1440 = worker->dsps1440 * nonces; JSON_CPACK(val, "{ss,si,ss,ss,sf,sf,sf,sf,sb,ss,ss,ss,ss}", "poolinstance", ckp->name, "elapsed", elapsed, "username", user->username, "workername", worker->workername, "hashrate", ghs1, "hashrate5m", ghs5, "hashrate1hr", ghs60, "hashrate24hr", ghs1440, "idle", worker->idle, "createdate", cdfield, "createby", "code", "createcode", __func__, "createinet", ckp->serverurl[0]); worker->notified_idle = worker->idle; entry = ckalloc(sizeof(json_entry_t)); entry->val = val; DL_APPEND(json_list, entry); } } ck_runlock(&sdata->instance_lock); /* Add all entries outside of the instance lock */ DL_FOREACH_SAFE(json_list, entry, tmpentry) { ckdbq_add(ckp, ID_WORKERSTATS, entry->val); DL_DELETE(json_list, entry); free(entry); } } static void *statsupdate(void *arg) { ckpool_t *ckp = (ckpool_t *)arg; sdata_t *sdata = ckp->data; pool_stats_t *stats = &sdata->stats; pthread_detach(pthread_self()); rename_proc("statsupdate"); tv_time(&stats->start_time); cksleep_prepare_r(&stats->last_update); sleep(1); while (42) { double ghs, ghs1, ghs5, ghs15, ghs60, ghs360, ghs1440, ghs10080, per_tdiff; char suffix1[16], suffix5[16], suffix15[16], suffix60[16], cdfield[64]; char suffix360[16], suffix1440[16], suffix10080[16]; char_entry_t *char_list = NULL, *char_t, *chartmp_t; stratum_instance_t *client, *tmp; user_instance_t *user, *tmpuser; int idle_workers = 0; char fname[512] = {}; tv_t now, diff; ts_t ts_now; json_t *val; FILE *fp; char *s; int i; tv_time(&now); timersub(&now, &stats->start_time, &diff); ck_rlock(&sdata->instance_lock); HASH_ITER(hh, sdata->stratum_instances, client, tmp) { if (!client_active(client)) continue; per_tdiff = tvdiff(&now, &client->last_share); /* Decay times per connected instance */ if (per_tdiff > 60) { /* No shares for over a minute, decay to 0 */ decay_time(&client->dsps1, 0, per_tdiff, 60); decay_time(&client->dsps5, 0, per_tdiff, 300); decay_time(&client->dsps60, 0, per_tdiff, 3600); decay_time(&client->dsps1440, 0, per_tdiff, 86400); decay_time(&client->dsps10080, 0, per_tdiff, 604800); idle_workers++; if (per_tdiff > 600) client->idle = true; continue; } } HASH_ITER(hh, sdata->user_instances, user, tmpuser) { worker_instance_t *worker; bool idle = false; if (!user->authorised) continue; /* Decay times per worker */ DL_FOREACH(user->worker_instances, worker) { per_tdiff = tvdiff(&now, &worker->last_share); if (per_tdiff > 60) { decay_time(&worker->dsps1, 0, per_tdiff, 60); decay_time(&worker->dsps5, 0, per_tdiff, 300); decay_time(&worker->dsps60, 0, per_tdiff, 3600); decay_time(&worker->dsps1440, 0, per_tdiff, 86400); worker->idle = true; } ghs = worker->dsps1 * nonces; suffix_string(ghs, suffix1, 16, 0); ghs = worker->dsps5 * nonces; suffix_string(ghs, suffix5, 16, 0); ghs = worker->dsps60 * nonces; suffix_string(ghs, suffix60, 16, 0); ghs = worker->dsps1440 * nonces; suffix_string(ghs, suffix1440, 16, 0); copy_tv(&worker->last_update, &now); JSON_CPACK(val, "{ss,ss,ss,ss,si,sf}", "hashrate1m", suffix1, "hashrate5m", suffix5, "hashrate1hr", suffix60, "hashrate1d", suffix1440, "lastupdate", now.tv_sec, "bestshare", worker->best_diff); snprintf(fname, 511, "%s/workers/%s", ckp->logdir, worker->workername); fp = fopen(fname, "we"); if (unlikely(!fp)) { LOGERR("Failed to fopen %s", fname); continue; } s = json_dumps(val, JSON_NO_UTF8 | JSON_PRESERVE_ORDER | JSON_EOL); fprintf(fp, "%s", s); dealloc(s); json_decref(val); fclose(fp); } /* Decay times per user */ per_tdiff = tvdiff(&now, &user->last_share); if (per_tdiff > 60) { decay_time(&user->dsps1, 0, per_tdiff, 60); decay_time(&user->dsps5, 0, per_tdiff, 300); decay_time(&user->dsps60, 0, per_tdiff, 3600); decay_time(&user->dsps1440, 0, per_tdiff, 86400); decay_time(&user->dsps10080, 0, per_tdiff, 604800); idle = true; } ghs = user->dsps1 * nonces; suffix_string(ghs, suffix1, 16, 0); ghs = user->dsps5 * nonces; suffix_string(ghs, suffix5, 16, 0); ghs = user->dsps60 * nonces; suffix_string(ghs, suffix60, 16, 0); ghs = user->dsps1440 * nonces; suffix_string(ghs, suffix1440, 16, 0); ghs = user->dsps10080 * nonces; suffix_string(ghs, suffix10080, 16, 0); copy_tv(&user->last_update, &now); JSON_CPACK(val, "{ss,ss,ss,ss,ss,si,si,sf}", "hashrate1m", suffix1, "hashrate5m", suffix5, "hashrate1hr", suffix60, "hashrate1d", suffix1440, "hashrate7d", suffix10080, "lastupdate", now.tv_sec, "workers", user->workers, "bestshare", user->best_diff); snprintf(fname, 511, "%s/users/%s", ckp->logdir, user->username); fp = fopen(fname, "we"); if (unlikely(!fp)) { LOGERR("Failed to fopen %s", fname); continue; } s = json_dumps(val, JSON_NO_UTF8 | JSON_PRESERVE_ORDER); fprintf(fp, "%s\n", s); if (!idle) { char_t = ckalloc(sizeof(char_entry_t)); ASPRINTF(&char_t->buf, "User %s:%s", user->username, s); DL_APPEND(char_list, char_t); } dealloc(s); json_decref(val); fclose(fp); } ck_runlock(&sdata->instance_lock); DL_FOREACH_SAFE(char_list, char_t, chartmp_t) { LOGNOTICE("%s", char_t->buf); DL_DELETE(char_list, char_t); free(char_t->buf); dealloc(char_t); } ghs1 = stats->dsps1 * nonces; suffix_string(ghs1, suffix1, 16, 0); ghs5 = stats->dsps5 * nonces; suffix_string(ghs5, suffix5, 16, 0); ghs15 = stats->dsps15 * nonces; suffix_string(ghs15, suffix15, 16, 0); ghs60 = stats->dsps60 * nonces; suffix_string(ghs60, suffix60, 16, 0); ghs360 = stats->dsps360 * nonces; suffix_string(ghs360, suffix360, 16, 0); ghs1440 = stats->dsps1440 * nonces; suffix_string(ghs1440, suffix1440, 16, 0); ghs10080 = stats->dsps10080 * nonces; suffix_string(ghs10080, suffix10080, 16, 0); snprintf(fname, 511, "%s/pool/pool.status", ckp->logdir); fp = fopen(fname, "we"); if (unlikely(!fp)) LOGERR("Failed to fopen %s", fname); JSON_CPACK(val, "{si,si,si,si,si,si}", "runtime", diff.tv_sec, "lastupdate", now.tv_sec, "Users", stats->users, "Workers", stats->workers, "Idle", idle_workers, "Disconnected", stats->disconnected); s = json_dumps(val, JSON_NO_UTF8 | JSON_PRESERVE_ORDER); json_decref(val); LOGNOTICE("Pool:%s", s); fprintf(fp, "%s\n", s); dealloc(s); JSON_CPACK(val, "{ss,ss,ss,ss,ss,ss,ss}", "hashrate1m", suffix1, "hashrate5m", suffix5, "hashrate15m", suffix15, "hashrate1hr", suffix60, "hashrate6hr", suffix360, "hashrate1d", suffix1440, "hashrate7d", suffix10080); s = json_dumps(val, JSON_NO_UTF8 | JSON_PRESERVE_ORDER); json_decref(val); LOGNOTICE("Pool:%s", s); fprintf(fp, "%s\n", s); dealloc(s); JSON_CPACK(val, "{sf,sf,sf,sf}", "SPS1m", stats->sps1, "SPS5m", stats->sps5, "SPS15m", stats->sps15, "SPS1h", stats->sps60); s = json_dumps(val, JSON_NO_UTF8 | JSON_PRESERVE_ORDER); json_decref(val); LOGNOTICE("Pool:%s", s); fprintf(fp, "%s\n", s); dealloc(s); fclose(fp); ts_realtime(&ts_now); sprintf(cdfield, "%lu,%lu", ts_now.tv_sec, ts_now.tv_nsec); JSON_CPACK(val, "{ss,si,si,si,sf,sf,sf,sf,ss,ss,ss,ss}", "poolinstance", ckp->name, "elapsed", diff.tv_sec, "users", stats->users, "workers", stats->workers, "hashrate", ghs1, "hashrate5m", ghs5, "hashrate1hr", ghs60, "hashrate24hr", ghs1440, "createdate", cdfield, "createby", "code", "createcode", __func__, "createinet", ckp->serverurl[0]); ckdbq_add(ckp, ID_POOLSTATS, val); /* Update stats 32 times per minute to divide up userstats for * ckdb, displaying status every minute. */ for (i = 0; i < 32; i++) { cksleep_ms_r(&stats->last_update, 1875); cksleep_prepare_r(&stats->last_update); update_workerstats(ckp, sdata); mutex_lock(&sdata->stats_lock); stats->accounted_shares += stats->unaccounted_shares; stats->accounted_diff_shares += stats->unaccounted_diff_shares; stats->accounted_rejects += stats->unaccounted_rejects; decay_time(&stats->sps1, stats->unaccounted_shares, 1.875, 60); decay_time(&stats->sps5, stats->unaccounted_shares, 1.875, 300); decay_time(&stats->sps15, stats->unaccounted_shares, 1.875, 900); decay_time(&stats->sps60, stats->unaccounted_shares, 1.875, 3600); decay_time(&stats->dsps1, stats->unaccounted_diff_shares, 1.875, 60); decay_time(&stats->dsps5, stats->unaccounted_diff_shares, 1.875, 300); decay_time(&stats->dsps15, stats->unaccounted_diff_shares, 1.875, 900); decay_time(&stats->dsps60, stats->unaccounted_diff_shares, 1.875, 3600); decay_time(&stats->dsps360, stats->unaccounted_diff_shares, 1.875, 21600); decay_time(&stats->dsps1440, stats->unaccounted_diff_shares, 1.875, 86400); decay_time(&stats->dsps10080, stats->unaccounted_diff_shares, 1.875, 604800); stats->unaccounted_shares = stats->unaccounted_diff_shares = stats->unaccounted_rejects = 0; mutex_unlock(&sdata->stats_lock); } } return NULL; } /* Sends a heartbeat to ckdb every second to maintain the relationship of * ckpool always initiating a request -> getting a ckdb response, but allows * ckdb to provide specific commands to ckpool. */ static void *ckdb_heartbeat(void *arg) { ckpool_t *ckp = (ckpool_t *)arg; sdata_t *sdata = ckp->data; pthread_detach(pthread_self()); rename_proc("heartbeat"); while (42) { char cdfield[64]; ts_t ts_now; json_t *val; cksleep_ms(1000); if (unlikely(!ckmsgq_empty(sdata->ckdbq))) { LOGDEBUG("Witholding heartbeat due to ckdb messages being queued"); continue; } ts_realtime(&ts_now); sprintf(cdfield, "%lu,%lu", ts_now.tv_sec, ts_now.tv_nsec); JSON_CPACK(val, "{ss,ss,ss,ss}", "createdate", cdfield, "createby", "code", "createcode", __func__, "createinet", ckp->serverurl[0]); ckdbq_add(ckp, ID_HEARTBEAT, val); } return NULL; } static void read_poolstats(ckpool_t *ckp) { char *s = alloca(4096), *pstats, *dsps, *sps; sdata_t *sdata = ckp->data; pool_stats_t *stats = &sdata->stats; int tvsec_diff = 0, ret; tv_t now, last; json_t *val; FILE *fp; snprintf(s, 4095, "%s/pool/pool.status", ckp->logdir); fp = fopen(s, "re"); if (!fp) { LOGINFO("Pool does not have a logfile to read"); return; } memset(s, 0, 4096); ret = fread(s, 1, 4095, fp); fclose(fp); if (ret < 1 || !strlen(s)) { LOGDEBUG("No string to read in pool logfile"); return; } /* Strip out end of line terminators */ pstats = strsep(&s, "\n"); dsps = strsep(&s, "\n"); sps = strsep(&s, "\n"); if (!s) { LOGINFO("Failed to find EOL in pool logfile"); return; } val = json_loads(pstats, 0, NULL); if (!val) { LOGINFO("Failed to json decode pstats line from pool logfile: %s", pstats); return; } tv_time(&now); last.tv_sec = 0; json_get_int64(&last.tv_sec, val, "lastupdate"); json_decref(val); LOGINFO("Successfully read pool pstats: %s", pstats); val = json_loads(dsps, 0, NULL); if (!val) { LOGINFO("Failed to json decode dsps line from pool logfile: %s", sps); return; } stats->dsps1 = dsps_from_key(val, "hashrate1m"); stats->dsps5 = dsps_from_key(val, "hashrate5m"); stats->dsps15 = dsps_from_key(val, "hashrate15m"); stats->dsps60 = dsps_from_key(val, "hashrate1hr"); stats->dsps360 = dsps_from_key(val, "hashrate6hr"); stats->dsps1440 = dsps_from_key(val, "hashrate1d"); stats->dsps10080 = dsps_from_key(val, "hashrate7d"); json_decref(val); LOGINFO("Successfully read pool dsps: %s", dsps); val = json_loads(sps, 0, NULL); if (!val) { LOGINFO("Failed to json decode sps line from pool logfile: %s", dsps); return; } json_get_double(&stats->sps1, val, "SPS1m"); json_get_double(&stats->sps5, val, "SPS5m"); json_get_double(&stats->sps15, val, "SPS15m"); json_get_double(&stats->sps60, val, "SPS1h"); json_decref(val); LOGINFO("Successfully read pool sps: %s", sps); if (last.tv_sec) tvsec_diff = now.tv_sec - last.tv_sec - 60; if (tvsec_diff > 60) { LOGNOTICE("Old pool stats indicate pool down for %d seconds, decaying stats", tvsec_diff); decay_time(&stats->sps1, 0, tvsec_diff, 60); decay_time(&stats->sps5, 0, tvsec_diff, 300); decay_time(&stats->sps15, 0, tvsec_diff, 900); decay_time(&stats->sps60, 0, tvsec_diff, 3600); decay_time(&stats->dsps1, 0, tvsec_diff, 60); decay_time(&stats->dsps5, 0, tvsec_diff, 300); decay_time(&stats->dsps15, 0, tvsec_diff, 900); decay_time(&stats->dsps60, 0, tvsec_diff, 3600); decay_time(&stats->dsps360, 0, tvsec_diff, 21600); decay_time(&stats->dsps1440, 0, tvsec_diff, 86400); decay_time(&stats->dsps10080, 0, tvsec_diff, 604800); } } int stratifier(proc_instance_t *pi) { pthread_t pth_blockupdate, pth_statsupdate, pth_heartbeat; ckpool_t *ckp = pi->ckp; int ret = 1, threads; int64_t randomiser; char *buf = NULL; sdata_t *sdata; LOGWARNING("%s stratifier starting", ckp->name); sdata = ckzalloc(sizeof(sdata_t)); ckp->data = sdata; sdata->ckp = ckp; sdata->verbose = true; /* Wait for the generator to have something for us */ do { if (!ping_main(ckp)) { ret = 1; goto out; } if (ckp->proxy) break; buf = send_recv_proc(ckp->generator, "ping"); } while (!buf); dealloc(buf); if (!ckp->proxy) { if (!test_address(ckp, ckp->btcaddress)) { LOGEMERG("Fatal: btcaddress invalid according to bitcoind"); goto out; } /* Store this for use elsewhere */ hex2bin(scriptsig_header_bin, scriptsig_header, 41); address_to_pubkeytxn(sdata->pubkeytxnbin, ckp->btcaddress); if (test_address(ckp, ckp->donaddress)) { ckp->donvalid = true; address_to_pubkeytxn(sdata->donkeytxnbin, ckp->donaddress); } } randomiser = time(NULL); sdata->enonce1_64 = htole64(randomiser); /* Set the initial id to time as high bits so as to not send the same * id on restarts */ randomiser <<= 32; if (!ckp->proxy) sdata->blockchange_id = sdata->workbase_id = randomiser; if (!ckp->serverurls) { ckp->serverurl[0] = "127.0.0.1"; ckp->serverurls = 1; } cklock_init(&sdata->instance_lock); mutex_init(&sdata->ckdb_lock); sdata->ssends = create_ckmsgq(ckp, "ssender", &ssend_process); /* Create half as many share processing threads as there are CPUs */ threads = sysconf(_SC_NPROCESSORS_ONLN) / 2 ? : 1; sdata->sshareq = create_ckmsgqs(ckp, "sprocessor", &sshare_process, threads); /* Create 1/4 as many stratum processing threads as there are CPUs */ threads = threads / 2 ? : 1; sdata->srecvs = create_ckmsgqs(ckp, "sreceiver", &srecv_process, threads); sdata->sauthq = create_ckmsgq(ckp, "authoriser", &sauth_process); sdata->stxnq = create_ckmsgq(ckp, "stxnq", &send_transactions); if (!CKP_STANDALONE(ckp)) { sdata->ckdbq = create_ckmsgq(ckp, "ckdbqueue", &ckdbq_process); create_pthread(&pth_heartbeat, ckdb_heartbeat, ckp); } read_poolstats(ckp); cklock_init(&sdata->workbase_lock); if (!ckp->proxy) create_pthread(&pth_blockupdate, blockupdate, ckp); else { mutex_init(&sdata->proxy_lock); } mutex_init(&sdata->stats_lock); create_pthread(&pth_statsupdate, statsupdate, ckp); mutex_init(&sdata->share_lock); mutex_init(&sdata->block_lock); LOGWARNING("%s stratifier ready", ckp->name); ret = stratum_loop(ckp, pi); out: if (ckp->proxy) { proxy_t *proxy, *tmpproxy; mutex_lock(&sdata->proxy_lock); HASH_ITER(hh, sdata->proxies, proxy, tmpproxy) { HASH_DEL(sdata->proxies, proxy); dealloc(proxy); } mutex_unlock(&sdata->proxy_lock); } dealloc(ckp->data); return process_exit(ckp, pi, ret); }