Browse Source

Simplify the total_queued count to those staged not cloned and remove the locking since it's no longer a critical value.

Clone only anticipated difference sicne there will be a lag from the value returned by requests_staged().
Keep 1/3 buffer of extra work items when ageing them.
nfactor-troky
Con Kolivas 12 years ago
parent
commit
9f811c528f
  1. 54
      cgminer.c

54
cgminer.c

@ -161,7 +161,6 @@ static int total_threads;
struct work_restart *work_restart = NULL; struct work_restart *work_restart = NULL;
static pthread_mutex_t hash_lock; static pthread_mutex_t hash_lock;
static pthread_mutex_t qd_lock;
static pthread_mutex_t *stgd_lock; static pthread_mutex_t *stgd_lock;
#ifdef HAVE_CURSES #ifdef HAVE_CURSES
static pthread_mutex_t curses_lock; static pthread_mutex_t curses_lock;
@ -2364,11 +2363,6 @@ void switch_pools(struct pool *selected)
if (pool != last_pool) if (pool != last_pool)
applog(LOG_WARNING, "Switching to %s", pool->rpc_url); applog(LOG_WARNING, "Switching to %s", pool->rpc_url);
/* Reset the queued amount to allow more to be queued for the new pool */
mutex_lock(&qd_lock);
total_queued = 0;
mutex_unlock(&qd_lock);
mutex_lock(&lp_lock); mutex_lock(&lp_lock);
pthread_cond_broadcast(&lp_cond); pthread_cond_broadcast(&lp_cond);
mutex_unlock(&lp_lock); mutex_unlock(&lp_lock);
@ -2386,31 +2380,21 @@ static void discard_work(struct work *work)
free_work(work); free_work(work);
} }
/* This is overkill, but at least we'll know accurately how much work is /* Done lockless since this is not a critical value */
* queued to prevent ever being left without work */ static inline void inc_queued(void)
static void inc_queued(void)
{ {
mutex_lock(&qd_lock);
total_queued++; total_queued++;
mutex_unlock(&qd_lock);
} }
static void dec_queued(void) static inline void dec_queued(void)
{ {
mutex_lock(&qd_lock); if (likely(total_queued > 0))
if (total_queued > 0)
total_queued--; total_queued--;
mutex_unlock(&qd_lock);
} }
static int requests_queued(void) static int requests_queued(void)
{ {
int ret; return requests_staged() - staged_extras;
mutex_lock(&qd_lock);
ret = total_queued;
mutex_unlock(&qd_lock);
return ret;
} }
static int discard_stale(void) static int discard_stale(void)
@ -3509,20 +3493,12 @@ static void pool_resus(struct pool *pool)
switch_pools(NULL); switch_pools(NULL);
} }
static long requested_tv_sec;
static bool queue_request(struct thr_info *thr, bool needed) static bool queue_request(struct thr_info *thr, bool needed)
{ {
int rq = requests_queued(); int rs = requests_staged(), rq = requests_queued();
struct workio_cmd *wc; struct workio_cmd *wc;
struct timeval now;
gettimeofday(&now, NULL);
/* Space out retrieval of extra work according to the number of mining if (rq >= mining_threads || (rq >= opt_queue && rs >= mining_threads))
* threads */
if (rq >= mining_threads + staged_extras &&
(now.tv_sec - requested_tv_sec) < opt_scantime / (mining_threads + 1))
return true; return true;
/* fill out work request message */ /* fill out work request message */
@ -3553,7 +3529,6 @@ static bool queue_request(struct thr_info *thr, bool needed)
return false; return false;
} }
requested_tv_sec = now.tv_sec;
inc_queued(); inc_queued();
return true; return true;
} }
@ -3637,11 +3612,17 @@ static struct work *make_clone(struct work *work)
* the future */ * the future */
static struct work *clone_work(struct work *work) static struct work *clone_work(struct work *work)
{ {
int mrs = mining_threads - requests_staged();
struct work *work_clone; struct work *work_clone;
bool cloned = false; bool cloned;
if (mrs < 1)
return work;
cloned = false;
work_clone = make_clone(work); work_clone = make_clone(work);
while (requests_staged() < mining_threads && can_roll(work) && should_roll(work)) { while (mrs-- > 0 && can_roll(work) && should_roll(work)) {
applog(LOG_DEBUG, "Pushing rolled converted work to stage thread"); applog(LOG_DEBUG, "Pushing rolled converted work to stage thread");
if (unlikely(!stage_work(work_clone))) { if (unlikely(!stage_work(work_clone))) {
cloned = false; cloned = false;
@ -3699,7 +3680,7 @@ retry:
goto out; goto out;
} }
if (!pool->lagging && requested && !newreq && !requests_staged() && requests_queued() >= mining_threads) { if (!pool->lagging && requested && !newreq && !requests_staged()) {
struct cgpu_info *cgpu = thr->cgpu; struct cgpu_info *cgpu = thr->cgpu;
bool stalled = true; bool stalled = true;
int i; int i;
@ -4321,7 +4302,7 @@ static void age_work(void)
{ {
int discarded = 0; int discarded = 0;
while (requests_staged() > mining_threads) { while (requests_staged() > mining_threads * 4 / 3) {
struct work *work = hash_pop(NULL); struct work *work = hash_pop(NULL);
if (unlikely(!work)) if (unlikely(!work))
@ -4905,7 +4886,6 @@ int main(int argc, char *argv[])
#endif #endif
mutex_init(&hash_lock); mutex_init(&hash_lock);
mutex_init(&qd_lock);
#ifdef HAVE_CURSES #ifdef HAVE_CURSES
mutex_init(&curses_lock); mutex_init(&curses_lock);
#endif #endif

Loading…
Cancel
Save