Browse Source

Set and get the current pool under lock.

nfactor-troky
Con Kolivas 14 years ago
parent
commit
ecda75946b
  1. 73
      main.c

73
main.c

@ -181,7 +181,7 @@ static unsigned int local_work; @@ -181,7 +181,7 @@ static unsigned int local_work;
static unsigned int total_lo, total_ro;
static struct pool *pools = NULL;
static struct pool *cp; /* Current pool */
static struct pool *currentpool;
static int total_pools;
static bool curses_active = false;
@ -247,6 +247,23 @@ static bool pool_tclear(struct pool *pool, bool *var) @@ -247,6 +247,23 @@ static bool pool_tclear(struct pool *pool, bool *var)
return ret;
}
static struct pool *current_pool(void)
{
struct pool *pool;
pthread_mutex_lock(&control_lock);
pool = currentpool;
pthread_mutex_unlock(&control_lock);
return pool;
}
static void set_current_pool(struct pool *pool)
{
pthread_mutex_lock(&control_lock);
currentpool = pool;
pthread_mutex_unlock(&control_lock);
}
/* FIXME: Use asprintf for better errors. */
static char *set_algo(const char *arg, enum sha256_algos *algo)
{
@ -651,7 +668,7 @@ static void text_print_status(int thr_id) @@ -651,7 +668,7 @@ static void text_print_status(int thr_id)
/* Must be called with curses mutex lock held and curses_active */
static void curses_print_status(int thr_id)
{
struct pool *pool = cp;
struct pool *pool = current_pool();
wmove(statuswin, 0, 0);
wattron(statuswin, A_BOLD);
@ -732,7 +749,7 @@ static bool submit_upstream_work(const struct work *work) @@ -732,7 +749,7 @@ static bool submit_upstream_work(const struct work *work)
int thr_id = work->thr_id;
struct cgpu_info *cgpu = thr_info[thr_id].cgpu;
CURL *curl = curl_easy_init();
struct pool *pool = cp;
struct pool *pool = work->pool;
if (unlikely(!curl)) {
applog(LOG_ERR, "CURL initialisation failed");
@ -816,7 +833,7 @@ static const char *rpc_req = @@ -816,7 +833,7 @@ static const char *rpc_req =
static bool get_upstream_work(struct work *work)
{
struct pool *pool = cp;
struct pool *pool = current_pool();
json_t *val;
bool rc = false;
CURL *curl = curl_easy_init();
@ -834,6 +851,7 @@ static bool get_upstream_work(struct work *work) @@ -834,6 +851,7 @@ static bool get_upstream_work(struct work *work)
}
rc = work_decode(json_object_get(val, "result"), work);
work->pool = pool;
json_decref(val);
out:
@ -997,12 +1015,13 @@ static bool stale_work(struct work *work) @@ -997,12 +1015,13 @@ static bool stale_work(struct work *work)
static void *submit_work_thread(void *userdata)
{
struct workio_cmd *wc = (struct workio_cmd *)userdata;
struct pool *pool = cp;
struct work *work = wc->u.work;
struct pool *pool = work->pool;
int failures = 0;
pthread_detach(pthread_self());
if (stale_work(wc->u.work)) {
if (stale_work(work)) {
applog(LOG_WARNING, "Stale share detected, discarding");
total_stale++;
pool->stale_shares++;
@ -1010,8 +1029,8 @@ static void *submit_work_thread(void *userdata) @@ -1010,8 +1029,8 @@ static void *submit_work_thread(void *userdata)
}
/* submit solution to bitcoin via JSON-RPC */
while (!submit_upstream_work(wc->u.work)) {
if (stale_work(wc->u.work)) {
while (!submit_upstream_work(work)) {
if (stale_work(work)) {
applog(LOG_WARNING, "Stale share detected, discarding");
total_stale++;
pool->stale_shares++;
@ -1044,10 +1063,8 @@ static bool workio_submit_work(struct workio_cmd *wc) @@ -1044,10 +1063,8 @@ static bool workio_submit_work(struct workio_cmd *wc)
return true;
}
static void inc_staged(int inc, bool lp)
static void inc_staged(struct pool *pool, int inc, bool lp)
{
struct pool *pool = cp;
pthread_mutex_lock(&stgd_lock);
if (lp) {
lp_staged += inc;
@ -1161,7 +1178,7 @@ static void *stage_thread(void *userdata) @@ -1161,7 +1178,7 @@ static void *stage_thread(void *userdata)
ok = false;
break;
}
inc_staged(1, false);
inc_staged(work->pool, 1, false);
}
tq_freeze(mythr->q);
@ -1305,13 +1322,15 @@ static bool queue_request(void) @@ -1305,13 +1322,15 @@ static bool queue_request(void)
{
int maxq = opt_queue + mining_threads;
struct workio_cmd *wc;
struct pool *pool = cp;
struct pool *pool;
/* If we've been generating lots of local work we may already have
* enough in the queue */
if (requests_queued() >= maxq || real_staged() >= maxq)
return true;
pool = current_pool();
/* fill out work request message */
wc = calloc(1, sizeof(*wc));
if (unlikely(!wc)) {
@ -1338,7 +1357,7 @@ static bool queue_request(void) @@ -1338,7 +1357,7 @@ static bool queue_request(void)
static void discard_staged(void)
{
struct work *work_heap;
struct pool *pool = cp;
struct pool *pool;
/* Just in case we fell in a hole and missed a queue filling */
if (unlikely(!requests_staged()))
@ -1348,6 +1367,7 @@ static void discard_staged(void) @@ -1348,6 +1367,7 @@ static void discard_staged(void)
if (unlikely(!work_heap))
return;
pool = work_heap->pool;
free(work_heap);
dec_queued();
pool->discarded_work++;
@ -1356,6 +1376,7 @@ static void discard_staged(void) @@ -1356,6 +1376,7 @@ static void discard_staged(void)
static void flush_requests(bool longpoll)
{
struct pool *pool = current_pool();
int i, stale;
/* We should have one fresh work item staged from the block change. */
@ -1365,7 +1386,7 @@ static void flush_requests(bool longpoll) @@ -1365,7 +1386,7 @@ static void flush_requests(bool longpoll)
/* Temporarily increase the staged count so that get_work thinks there
* is work available instead of making threads reuse existing work */
inc_staged(mining_threads, true);
inc_staged(pool, mining_threads, true);
for (i = 0; i < stale; i++) {
/* Queue a whole batch of new requests */
@ -1382,7 +1403,7 @@ static void flush_requests(bool longpoll) @@ -1382,7 +1403,7 @@ static void flush_requests(bool longpoll)
static bool get_work(struct work *work, bool queued)
{
struct pool *pool = cp;
struct pool *pool = current_pool();
struct work *work_heap;
bool ret = false;
int failures = 0;
@ -1412,7 +1433,7 @@ retry: @@ -1412,7 +1433,7 @@ retry:
/* A new block appears on average every 10 mins */
applog(LOG_WARNING, "Prolonged outage. Going idle till network recovers.");
/* Force every thread to wait for new work */
inc_staged(mining_threads, true);
inc_staged(pool, mining_threads, true);
goto retry;
}
}
@ -1960,7 +1981,7 @@ static void *longpoll_thread(void *userdata) @@ -1960,7 +1981,7 @@ static void *longpoll_thread(void *userdata)
char *copy_start, *hdr_path, *lp_url = NULL;
bool need_slash = false;
int failures = 0;
struct pool *pool = cp;
struct pool *pool = current_pool();
pthread_setcanceltype(PTHREAD_CANCEL_ASYNCHRONOUS, NULL);
@ -2047,6 +2068,7 @@ out: @@ -2047,6 +2068,7 @@ out:
return NULL;
}
#if 0
static void reinit_cputhread(int thr_id)
{
struct thr_info *thr = &thr_info[thr_id];
@ -2114,12 +2136,13 @@ static void reinit_thread(int thr_id) @@ -2114,12 +2136,13 @@ static void reinit_thread(int thr_id)
else
reinit_cputhread(thr_id);
}
#else
#else /* HAVE_OPENCL */
static void reinit_thread(int thr_id)
{
reinit_cputhread(thr_id);
}
#endif
#endif /* 0 */
/* Determine which are the first threads belonging to a device and if they're
* active */
@ -2142,7 +2165,6 @@ static void *watchdog_thread(void *userdata) @@ -2142,7 +2165,6 @@ static void *watchdog_thread(void *userdata)
{
const unsigned int interval = opt_log_interval / 2 ? : 1;
struct timeval zero_tv;
struct pool *pool = cp;
pthread_setcanceltype(PTHREAD_CANCEL_ASYNCHRONOUS, NULL);
@ -2182,10 +2204,8 @@ static void *watchdog_thread(void *userdata) @@ -2182,10 +2204,8 @@ static void *watchdog_thread(void *userdata)
gettimeofday(&now, NULL);
#if 0
for (i = 0; i < mining_threads; i++) {
#else
//for (i = 0; i < mining_threads; i++) {
for (i = 0; i < gpu_threads; i++) {
#endif
struct thr_info *thr = &thr_info[i];
/* Do not kill threads waiting on longpoll staged work
@ -2203,6 +2223,7 @@ static void *watchdog_thread(void *userdata) @@ -2203,6 +2223,7 @@ static void *watchdog_thread(void *userdata)
applog(LOG_WARNING, "Thread %d restarted", i);
}
}
#endif
}
return NULL;
@ -2339,8 +2360,8 @@ int main (int argc, char *argv[]) @@ -2339,8 +2360,8 @@ int main (int argc, char *argv[])
return 1;
}
cp = &pools[0];
pool = cp;
set_current_pool(pools);
pool = current_pool();
if (total_devices) {
if (total_devices > nDevs) {
@ -2466,7 +2487,7 @@ int main (int argc, char *argv[]) @@ -2466,7 +2487,7 @@ int main (int argc, char *argv[])
/* Flag the work as ready forcing the mining threads to wait till we
* actually put something into the queue */
inc_staged(mining_threads, true);
inc_staged(current_pool(), mining_threads, true);
/* Create a unique get work queue */
getq = tq_new();

Loading…
Cancel
Save