Browse Source

Going back to e68ecf5eb2

nfactor-troky
Con Kolivas 12 years ago
parent
commit
f27bcb8ee5
  1. 289
      cgminer.c
  2. 4
      driver-bitforce.c

289
cgminer.c

@ -166,6 +166,7 @@ static int api_thr_id;
static int total_threads; static int total_threads;
static pthread_mutex_t hash_lock; static pthread_mutex_t hash_lock;
static pthread_mutex_t qd_lock;
static pthread_mutex_t *stgd_lock; static pthread_mutex_t *stgd_lock;
pthread_mutex_t console_lock; pthread_mutex_t console_lock;
static pthread_mutex_t ch_lock; static pthread_mutex_t ch_lock;
@ -1333,38 +1334,33 @@ void decay_time(double *f, double fadd)
*f = (fadd + *f * 0.58) / 1.58; *f = (fadd + *f * 0.58) / 1.58;
} }
static int __total_staged(void)
{
return HASH_COUNT(staged_work);
}
static int total_staged(void) static int total_staged(void)
{ {
int ret; int ret;
mutex_lock(stgd_lock); mutex_lock(stgd_lock);
ret = __total_staged(); ret = HASH_COUNT(staged_work);
mutex_unlock(stgd_lock); mutex_unlock(stgd_lock);
return ret; return ret;
} }
/* We count the total_queued as pending staged as these are requests in flight static int pool_staged(struct pool *pool)
* one way or another which have not yet staged a work item but will */
static int __pending_staged(void)
{
return HASH_COUNT(staged_work) + total_queued;
}
static int pending_staged(void)
{ {
int ret; int ret;
mutex_lock(stgd_lock); mutex_lock(stgd_lock);
ret = __pending_staged(); ret = pool->staged;
mutex_unlock(stgd_lock); mutex_unlock(stgd_lock);
return ret; return ret;
} }
static int current_staged(void)
{
struct pool *pool = current_pool();
return pool_staged(pool);
}
#ifdef HAVE_CURSES #ifdef HAVE_CURSES
WINDOW *mainwin, *statuswin, *logwin; WINDOW *mainwin, *statuswin, *logwin;
#endif #endif
@ -2233,56 +2229,55 @@ static void push_curl_entry(struct curl_ent *ce, struct pool *pool)
/* This is overkill, but at least we'll know accurately how much work is /* This is overkill, but at least we'll know accurately how much work is
* queued to prevent ever being left without work */ * queued to prevent ever being left without work */
static void __inc_queued(void) static void inc_queued(struct pool *pool)
{ {
if (unlikely(!pool))
return;
mutex_lock(&qd_lock);
pool->queued++;
total_queued++; total_queued++;
mutex_unlock(&qd_lock);
} }
static int __pool_pending_staged(struct pool *pool) static void dec_queued(struct pool *pool)
{ {
return pool->queued + pool->staged; if (unlikely(!pool))
} return;
static void inc_pool_queued(struct pool *pool) mutex_lock(&qd_lock);
{ pool->queued--;
mutex_lock(stgd_lock); total_queued--;
pool->queued++; mutex_unlock(&qd_lock);
mutex_unlock(stgd_lock);
} }
static void __dec_queued(struct pool *pool) static int current_queued(void)
{ {
if (total_queued) struct pool *pool = current_pool();
total_queued--; int ret;
if (pool && pool->queued)
pool->queued--;
}
static void dec_queued(void) mutex_lock(&qd_lock);
{ ret = pool->queued;
mutex_lock(stgd_lock); mutex_unlock(&qd_lock);
__dec_queued(NULL); return ret;
mutex_unlock(stgd_lock);
} }
static int global_queued(void) static int global_queued(void)
{ {
int ret; int ret;
mutex_lock(stgd_lock); mutex_lock(&qd_lock);
ret = total_queued; ret = total_queued;
mutex_unlock(stgd_lock); mutex_unlock(&qd_lock);
return ret; return ret;
} }
static bool clone_available(void);
/* ce and pool may appear uninitialised at push_curl_entry, but they're always /* ce and pool may appear uninitialised at push_curl_entry, but they're always
* set when we don't have opt_benchmark enabled */ * set when we don't have opt_benchmark enabled */
static void *get_work_thread(void *userdata) static void *get_work_thread(void *userdata)
{ {
struct workio_cmd *wc = (struct workio_cmd *)userdata; struct workio_cmd *wc = (struct workio_cmd *)userdata;
struct curl_ent *ce = NULL; struct curl_ent * uninitialised_var(ce);
struct pool * uninitialised_var(pool); struct pool * uninitialised_var(pool);
struct work *ret_work = make_work(); struct work *ret_work = make_work();
int failures = 0; int failures = 0;
@ -2291,9 +2286,6 @@ static void *get_work_thread(void *userdata)
applog(LOG_DEBUG, "Creating extra get work thread"); applog(LOG_DEBUG, "Creating extra get work thread");
if (clone_available())
goto out;
if (wc->thr) if (wc->thr)
ret_work->thr = wc->thr; ret_work->thr = wc->thr;
else else
@ -2303,8 +2295,9 @@ static void *get_work_thread(void *userdata)
get_benchmark_work(ret_work); get_benchmark_work(ret_work);
else { else {
pool = ret_work->pool = select_pool(wc->lagging); pool = ret_work->pool = select_pool(wc->lagging);
inc_queued(pool);
ce = pop_curl_entry(pool); ce = pop_curl_entry(pool);
inc_pool_queued(pool);
/* obtain new work from bitcoin via JSON-RPC */ /* obtain new work from bitcoin via JSON-RPC */
while (!get_upstream_work(ret_work, ce->curl)) { while (!get_upstream_work(ret_work, ce->curl)) {
@ -2322,6 +2315,8 @@ static void *get_work_thread(void *userdata)
fail_pause += opt_fail_pause; fail_pause += opt_fail_pause;
} }
fail_pause = opt_fail_pause; fail_pause = opt_fail_pause;
dec_queued(pool);
} }
applog(LOG_DEBUG, "Pushing work to requesting thread"); applog(LOG_DEBUG, "Pushing work to requesting thread");
@ -2335,7 +2330,7 @@ static void *get_work_thread(void *userdata)
out: out:
workio_cmd_free(wc); workio_cmd_free(wc);
if (ce) if (!opt_benchmark)
push_curl_entry(ce, pool); push_curl_entry(ce, pool);
return NULL; return NULL;
} }
@ -2606,6 +2601,7 @@ static void discard_stale(void)
HASH_ITER(hh, staged_work, work, tmp) { HASH_ITER(hh, staged_work, work, tmp) {
if (stale_work(work, false)) { if (stale_work(work, false)) {
HASH_DEL(staged_work, work); HASH_DEL(staged_work, work);
work->pool->staged--;
discard_work(work); discard_work(work);
stale++; stale++;
} }
@ -2778,7 +2774,6 @@ static bool hash_push(struct work *work)
if (likely(!getq->frozen)) { if (likely(!getq->frozen)) {
HASH_ADD_INT(staged_work, id, work); HASH_ADD_INT(staged_work, id, work);
work->pool->staged++; work->pool->staged++;
__dec_queued(work->pool);
HASH_SORT(staged_work, tv_sort); HASH_SORT(staged_work, tv_sort);
} else } else
rc = false; rc = false;
@ -3736,114 +3731,35 @@ static void pool_resus(struct pool *pool)
switch_pools(NULL); switch_pools(NULL);
} }
static struct work *make_clone(struct work *work)
{
struct work *work_clone = make_work();
memcpy(work_clone, work, sizeof(struct work));
work_clone->clone = true;
work_clone->longpoll = false;
work_clone->mandatory = false;
return work_clone;
}
static inline bool should_roll(struct work *work)
{
if (work->pool == current_pool() || pool_strategy == POOL_LOADBALANCE)
return true;
return false;
}
/* Limit rolls to 7000 to not beyond 2 hours in the future where bitcoind will
* reject blocks as invalid. */
static inline bool can_roll(struct work *work)
{
return (work->pool && work->rolltime && !work->clone &&
work->rolls < 7000 && !stale_work(work, false));
}
static void roll_work(struct work *work)
{
uint32_t *work_ntime;
uint32_t ntime;
work_ntime = (uint32_t *)(work->data + 68);
ntime = be32toh(*work_ntime);
ntime++;
*work_ntime = htobe32(ntime);
local_work++;
work->rolls++;
work->blk.nonce = 0;
applog(LOG_DEBUG, "Successfully rolled work");
/* This is now a different work item so it needs a different ID for the
* hashtable */
work->id = total_work++;
}
static bool clone_available(void)
{
struct work *work, *tmp;
bool cloned = false;
mutex_lock(stgd_lock);
HASH_ITER(hh, staged_work, work, tmp) {
if (can_roll(work) && should_roll(work)) {
struct work *work_clone;
roll_work(work);
work_clone = make_clone(work);
roll_work(work);
applog(LOG_DEBUG, "Pushing cloned available work to stage thread");
if (unlikely(!stage_work(work_clone))) {
free(work_clone);
break;
}
cloned = true;
break;
}
}
mutex_unlock(stgd_lock);
return cloned;
}
bool queue_request(struct thr_info *thr, bool needed) bool queue_request(struct thr_info *thr, bool needed)
{ {
struct pool *cp = current_pool(); int cq, cs, ts, tq, maxq = opt_queue + mining_threads;
struct workio_cmd *wc; struct workio_cmd *wc;
int ps, ts, maxq, pps; bool lag = false;
bool lag, ret, qing;
maxq = opt_queue + mining_threads;
lag = ret = qing = false;
mutex_lock(stgd_lock); cq = current_queued();
__inc_queued(); cs = current_staged();
ps = __pending_staged(); ts = total_staged();
ts = __total_staged(); tq = global_queued();
pps = __pool_pending_staged(cp);
mutex_unlock(stgd_lock);
if (opt_fail_only) {
if (pps >= maxq && ps) {
ret = true;
goto out;
}
} else if (pps && ps >= maxq) {
ret = true;
goto out;
}
if (needed && !ts && !opt_fail_only) if (needed && cq >= maxq && !ts && !opt_fail_only) {
/* If we're queueing work faster than we can stage it, consider
* the system lagging and allow work to be gathered from
* another pool if possible */
lag = true; lag = true;
} else {
/* Test to make sure we have enough work for pools without rolltime
* and enough original work for pools with rolltime */
if (((cs || cq >= opt_queue) && ts >= maxq) ||
((cs || cq) && tq >= maxq))
return true;
}
/* fill out work request message */ /* fill out work request message */
wc = calloc(1, sizeof(*wc)); wc = calloc(1, sizeof(*wc));
if (unlikely(!wc)) { if (unlikely(!wc)) {
applog(LOG_ERR, "Failed to calloc wc in queue_request"); applog(LOG_ERR, "Failed to calloc wc in queue_request");
goto out; return false;
} }
wc->cmd = WC_GET_WORK; wc->cmd = WC_GET_WORK;
@ -3856,41 +3772,26 @@ bool queue_request(struct thr_info *thr, bool needed)
if (unlikely(!tq_push(thr_info[work_thr_id].q, wc))) { if (unlikely(!tq_push(thr_info[work_thr_id].q, wc))) {
applog(LOG_ERR, "Failed to tq_push in queue_request"); applog(LOG_ERR, "Failed to tq_push in queue_request");
workio_cmd_free(wc); workio_cmd_free(wc);
goto out; return false;
} }
qing = ret = true;
out:
if (!qing)
dec_queued();
return true; return true;
} }
static struct work *hash_pop(const struct timespec *abstime) static struct work *hash_pop(const struct timespec *abstime)
{ {
struct work *work = NULL, *worka, *workb; struct work *work = NULL;
int rc = 0; int rc = 0;
mutex_lock(stgd_lock); mutex_lock(stgd_lock);
while (!getq->frozen && !HASH_COUNT(staged_work) && !rc) while (!getq->frozen && !HASH_COUNT(staged_work) && !rc)
rc = pthread_cond_timedwait(&getq->cond, stgd_lock, abstime); rc = pthread_cond_timedwait(&getq->cond, stgd_lock, abstime);
if (unlikely(!HASH_COUNT(staged_work))) if (HASH_COUNT(staged_work)) {
goto out_unlock; work = staged_work;
HASH_DEL(staged_work, work);
/* Look for cloned work first since original work can be used to work->pool->staged--;
* generate further clones */
HASH_ITER(hh, staged_work, worka, workb) {
if (worka->clone) {
HASH_DEL(staged_work, worka);
work = worka;
work->pool->staged--;
goto out_unlock;
}
} }
work = staged_work;
HASH_DEL(staged_work, work);
out_unlock:
mutex_unlock(stgd_lock); mutex_unlock(stgd_lock);
queue_request(NULL, false); queue_request(NULL, false);
@ -3898,6 +3799,40 @@ out_unlock:
return work; return work;
} }
static inline bool should_roll(struct work *work)
{
if (work->pool == current_pool() || pool_strategy == POOL_LOADBALANCE)
return true;
return false;
}
/* Limit rolls to 7000 to not beyond 2 hours in the future where bitcoind will
* reject blocks as invalid. */
static inline bool can_roll(struct work *work)
{
return (work->pool && work->rolltime && !work->clone &&
work->rolls < 7000 && !stale_work(work, false));
}
static void roll_work(struct work *work)
{
uint32_t *work_ntime;
uint32_t ntime;
work_ntime = (uint32_t *)(work->data + 68);
ntime = be32toh(*work_ntime);
ntime++;
*work_ntime = htobe32(ntime);
local_work++;
work->rolls++;
work->blk.nonce = 0;
applog(LOG_DEBUG, "Successfully rolled work");
/* This is now a different work item so it needs a different ID for the
* hashtable */
work->id = total_work++;
}
static bool reuse_work(struct work *work) static bool reuse_work(struct work *work)
{ {
if (can_roll(work) && should_roll(work)) { if (can_roll(work) && should_roll(work)) {
@ -3907,6 +3842,21 @@ static bool reuse_work(struct work *work)
return false; return false;
} }
static struct work *make_clone(struct work *work)
{
struct work *work_clone = make_work();
memcpy(work_clone, work, sizeof(struct work));
work_clone->clone = true;
work_clone->longpoll = false;
work_clone->mandatory = false;
/* Make cloned work appear slightly older to bias towards keeping the
* master work item which can be further rolled */
work_clone->tv_staged.tv_sec -= 1;
return work_clone;
}
/* Clones work by rolling it if possible, and returning a clone instead of the /* Clones work by rolling it if possible, and returning a clone instead of the
* original work item which gets staged again to possibly be rolled again in * original work item which gets staged again to possibly be rolled again in
* the future */ * the future */
@ -3952,7 +3902,7 @@ static bool get_work(struct work *work, bool requested, struct thr_info *thr,
struct timespec abstime = {0, 0}; struct timespec abstime = {0, 0};
struct timeval now; struct timeval now;
struct work *work_heap; struct work *work_heap;
int failures = 0, tq; int failures = 0, cq;
struct pool *pool; struct pool *pool;
/* Tell the watchdog thread this thread is waiting on getwork and /* Tell the watchdog thread this thread is waiting on getwork and
@ -3965,10 +3915,10 @@ static bool get_work(struct work *work, bool requested, struct thr_info *thr,
return true; return true;
} }
tq = global_queued(); cq = current_queued();
retry: retry:
pool = current_pool(); pool = current_pool();
if (!requested || tq < opt_queue) { if (!requested || cq < opt_queue) {
if (unlikely(!queue_request(thr, true))) { if (unlikely(!queue_request(thr, true))) {
applog(LOG_WARNING, "Failed to queue_request in get_work"); applog(LOG_WARNING, "Failed to queue_request in get_work");
goto out; goto out;
@ -3981,7 +3931,7 @@ retry:
goto out; goto out;
} }
if (!pool->lagging && requested && !newreq && !total_staged() && pending_staged() >= mining_threads + opt_queue) { if (!pool->lagging && requested && !newreq && !pool_staged(pool) && cq >= mining_threads + opt_queue) {
struct cgpu_info *cgpu = thr->cgpu; struct cgpu_info *cgpu = thr->cgpu;
bool stalled = true; bool stalled = true;
int i; int i;
@ -5219,6 +5169,7 @@ int main(int argc, char *argv[])
#endif #endif
mutex_init(&hash_lock); mutex_init(&hash_lock);
mutex_init(&qd_lock);
mutex_init(&console_lock); mutex_init(&console_lock);
mutex_init(&control_lock); mutex_init(&control_lock);
mutex_init(&sharelog_lock); mutex_init(&sharelog_lock);

4
driver-bitforce.c

@ -421,6 +421,8 @@ re_send:
goto re_send; goto re_send;
} }
applog(LOG_ERR, "BFL%i: Error: Send work reports: %s", bitforce->device_id, pdevbuf); applog(LOG_ERR, "BFL%i: Error: Send work reports: %s", bitforce->device_id, pdevbuf);
bitforce->hw_errors++;
bitforce_clear_buffer(bitforce);
return false; return false;
} }
@ -461,6 +463,8 @@ re_send:
if (unlikely(strncasecmp(pdevbuf, "OK", 2))) { if (unlikely(strncasecmp(pdevbuf, "OK", 2))) {
applog(LOG_ERR, "BFL%i: Error: Send block data reports: %s", bitforce->device_id, pdevbuf); applog(LOG_ERR, "BFL%i: Error: Send block data reports: %s", bitforce->device_id, pdevbuf);
bitforce->hw_errors++;
bitforce_clear_buffer(bitforce);
return false; return false;
} }

Loading…
Cancel
Save