Browse Source

Simplify queued hashtable by storing unqueued work separately in a single pointer.

nfactor-troky
ckolivas 11 years ago
parent
commit
680f014c85
  1. 70
      cgminer.c
  2. 2
      driver-bflsc.c
  3. 4
      driver-klondike.c
  4. 2
      miner.h

70
cgminer.c

@ -6341,55 +6341,39 @@ static void hash_sole_work(struct thr_info *mythr)
cgpu->deven = DEV_DISABLED; cgpu->deven = DEV_DISABLED;
} }
/* Create a hashtable of work items for devices with a queue. The device /* Put a new unqueued work item in cgpu->unqueued_work under cgpu->qlock till
* driver must have a custom queue_full function or it will default to true * the driver tells us it's full so that it may extract the work item using
* and put only one work item in the queue. Work items should not be removed * the get_queued() function which adds it to the hashtable on
* from this hashtable until they are no longer in use anywhere. Once a work * cgpu->queued_work. */
* item is physically queued on the device itself, the work->queued flag
* should be set under cgpu->qlock write lock to prevent it being dereferenced
* while still in use. */
static void fill_queue(struct thr_info *mythr, struct cgpu_info *cgpu, struct device_drv *drv, const int thr_id) static void fill_queue(struct thr_info *mythr, struct cgpu_info *cgpu, struct device_drv *drv, const int thr_id)
{ {
do { do {
bool need_work;
rd_lock(&cgpu->qlock);
need_work = (HASH_COUNT(cgpu->queued_work) == cgpu->queued_count);
rd_unlock(&cgpu->qlock);
if (need_work) {
struct work *work = get_work(mythr, thr_id);
wr_lock(&cgpu->qlock); wr_lock(&cgpu->qlock);
HASH_ADD_INT(cgpu->queued_work, id, work); if (!cgpu->unqueued_work)
cgpu->unqueued_work = get_work(mythr, thr_id);
wr_unlock(&cgpu->qlock); wr_unlock(&cgpu->qlock);
}
/* The queue_full function should be used by the driver to /* The queue_full function should be used by the driver to
* actually place work items on the physical device if it * actually place work items on the physical device if it
* does have a queue. */ * does have a queue. */
} while (!drv->queue_full(cgpu)); } while (!drv->queue_full(cgpu));
} }
/* This function is for retrieving one work item from the queued hashtable of /* This function is for retrieving one work item from the unqueued pointer and
* available work items that are not yet physically on a device (which is * adding it to the hashtable of queued work. Code using this function must be
* flagged with the work->queued bool). Code using this function must be able * able to handle NULL as a return which implies there is no work available. */
* to handle NULL as a return which implies there is no work available. */
struct work *get_queued(struct cgpu_info *cgpu) struct work *get_queued(struct cgpu_info *cgpu)
{ {
struct work *work, *tmp, *ret = NULL; struct work *work = NULL;
wr_lock(&cgpu->qlock); wr_lock(&cgpu->qlock);
HASH_ITER(hh, cgpu->queued_work, work, tmp) { if (cgpu->unqueued_work) {
if (!work->queued) { work = cgpu->unqueued_work;
work->queued = true; HASH_ADD_INT(cgpu->queued_work, id, work);
cgpu->queued_count++; cgpu->unqueued_work = NULL;
ret = work;
break;
}
} }
wr_unlock(&cgpu->qlock); wr_unlock(&cgpu->qlock);
return ret; return work;
} }
/* This function is for finding an already queued work item in the /* This function is for finding an already queued work item in the
@ -6402,8 +6386,7 @@ struct work *__find_work_bymidstate(struct work *que, char *midstate, size_t mid
struct work *work, *tmp, *ret = NULL; struct work *work, *tmp, *ret = NULL;
HASH_ITER(hh, que, work, tmp) { HASH_ITER(hh, que, work, tmp) {
if (work->queued && if (memcmp(work->midstate, midstate, midstatelen) == 0 &&
memcmp(work->midstate, midstate, midstatelen) == 0 &&
memcmp(work->data + offset, data, datalen) == 0) { memcmp(work->data + offset, data, datalen) == 0) {
ret = work; ret = work;
break; break;
@ -6443,7 +6426,6 @@ struct work *clone_queued_work_bymidstate(struct cgpu_info *cgpu, char *midstate
void __work_completed(struct cgpu_info *cgpu, struct work *work) void __work_completed(struct cgpu_info *cgpu, struct work *work)
{ {
if (work->queued)
cgpu->queued_count--; cgpu->queued_count--;
HASH_DEL(cgpu->queued_work, work); HASH_DEL(cgpu->queued_work, work);
} }
@ -6475,23 +6457,17 @@ struct work *take_queued_work_bymidstate(struct cgpu_info *cgpu, char *midstate,
static void flush_queue(struct cgpu_info *cgpu) static void flush_queue(struct cgpu_info *cgpu)
{ {
struct work *work, *tmp; struct work *work = NULL;
int discarded = 0;
wr_lock(&cgpu->qlock); wr_lock(&cgpu->qlock);
HASH_ITER(hh, cgpu->queued_work, work, tmp) { work = cgpu->unqueued_work;
/* Can only discard the work items if they're not physically cgpu->unqueued_work = NULL;
* queued on the device. */
if (!work->queued) {
HASH_DEL(cgpu->queued_work, work);
discard_work(work);
discarded++;
}
}
wr_unlock(&cgpu->qlock); wr_unlock(&cgpu->qlock);
if (discarded) if (work) {
applog(LOG_DEBUG, "Discarded %d queued work items", discarded); free_work(work);
applog(LOG_DEBUG, "Discarded queued work item");
}
} }
/* This version of hash work is for devices that are fast enough to always /* This version of hash work is for devices that are fast enough to always

2
driver-bflsc.c

@ -948,7 +948,7 @@ static void flush_one_dev(struct cgpu_info *bflsc, int dev)
rd_lock(&bflsc->qlock); rd_lock(&bflsc->qlock);
HASH_ITER(hh, bflsc->queued_work, work, tmp) { HASH_ITER(hh, bflsc->queued_work, work, tmp) {
if (work->queued && work->subid == dev) { if (work->subid == dev) {
// devflag is used to flag stale work // devflag is used to flag stale work
work->devflag = true; work->devflag = true;
did = true; did = true;

4
driver-klondike.c

@ -720,7 +720,7 @@ static void klondike_check_nonce(struct cgpu_info *klncgpu, KLIST *kitem)
cgtime(&tv_now); cgtime(&tv_now);
rd_lock(&(klncgpu->qlock)); rd_lock(&(klncgpu->qlock));
HASH_ITER(hh, klncgpu->queued_work, look, tmp) { HASH_ITER(hh, klncgpu->queued_work, look, tmp) {
if (look->queued && ms_tdiff(&tv_now, &(look->tv_stamp)) < OLD_WORK_MS && if (ms_tdiff(&tv_now, &(look->tv_stamp)) < OLD_WORK_MS &&
(look->subid == (kline->wr.dev*256 + kline->wr.workid))) { (look->subid == (kline->wr.dev*256 + kline->wr.workid))) {
work = look; work = look;
break; break;
@ -1026,14 +1026,12 @@ static bool klondike_send_work(struct cgpu_info *klncgpu, int dev, struct work *
cgtime(&tv_old); cgtime(&tv_old);
wr_lock(&klncgpu->qlock); wr_lock(&klncgpu->qlock);
HASH_ITER(hh, klncgpu->queued_work, look, tmp) { HASH_ITER(hh, klncgpu->queued_work, look, tmp) {
if (look->queued) {
if (ms_tdiff(&tv_old, &(look->tv_stamp)) > OLD_WORK_MS) { if (ms_tdiff(&tv_old, &(look->tv_stamp)) > OLD_WORK_MS) {
__work_completed(klncgpu, look); __work_completed(klncgpu, look);
free_work(look); free_work(look);
} else } else
wque_size++; wque_size++;
} }
}
wr_unlock(&klncgpu->qlock); wr_unlock(&klncgpu->qlock);
wr_lock(&(klninfo->stat_lock)); wr_lock(&(klninfo->stat_lock));

2
miner.h

@ -574,6 +574,7 @@ struct cgpu_info {
pthread_rwlock_t qlock; pthread_rwlock_t qlock;
struct work *queued_work; struct work *queued_work;
struct work *unqueued_work;
unsigned int queued_count; unsigned int queued_count;
bool shutdown; bool shutdown;
@ -1381,7 +1382,6 @@ struct work {
bool stale; bool stale;
bool mandatory; bool mandatory;
bool block; bool block;
bool queued;
bool stratum; bool stratum;
char *job_id; char *job_id;

Loading…
Cancel
Save