Browse Source

Assume we need a full allotment of work after lp and make clone_work clone_lpwork only.

nfactor-troky
Con Kolivas 12 years ago
parent
commit
c1886c9c99
  1. 6
      cgminer.c

6
cgminer.c

@ -3970,9 +3970,9 @@ static bool reuse_work(struct work *work)
/* Clones work by rolling it if possible, and returning a clone instead of the /* Clones work by rolling it if possible, and returning a clone instead of the
* original work item which gets staged again to possibly be rolled again in * original work item which gets staged again to possibly be rolled again in
* the future */ * the future */
static struct work *clone_work(struct work *work) static struct work *clone_lpwork(struct work *work)
{ {
int oq = opt_queue * mining_threads, mrs = mining_threads + oq - total_staged(); int oq = opt_queue * mining_threads, mrs = mining_threads + oq;
struct work *work_clone; struct work *work_clone;
bool cloned; bool cloned;
@ -4413,7 +4413,7 @@ static void convert_to_work(json_t *val, int rolltime, struct pool *pool)
return; return;
} }
work = clone_work(work); work = clone_lpwork(work);
applog(LOG_DEBUG, "Pushing converted work to stage thread"); applog(LOG_DEBUG, "Pushing converted work to stage thread");

Loading…
Cancel
Save