mirror of
https://github.com/GOSTSec/sgminer
synced 2025-01-11 15:27:53 +00:00
Further simplify the queue request mechanism.
This commit is contained in:
parent
f83863a996
commit
8085ae6854
86
cgminer.c
86
cgminer.c
@ -1355,23 +1355,6 @@ static int pending_staged(void)
|
|||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int pool_staged(struct pool *pool)
|
|
||||||
{
|
|
||||||
int ret;
|
|
||||||
|
|
||||||
mutex_lock(stgd_lock);
|
|
||||||
ret = pool->staged;
|
|
||||||
mutex_unlock(stgd_lock);
|
|
||||||
return ret;
|
|
||||||
}
|
|
||||||
|
|
||||||
static int current_staged(void)
|
|
||||||
{
|
|
||||||
struct pool *pool = current_pool();
|
|
||||||
|
|
||||||
return pool_staged(pool);
|
|
||||||
}
|
|
||||||
|
|
||||||
#ifdef HAVE_CURSES
|
#ifdef HAVE_CURSES
|
||||||
WINDOW *mainwin, *statuswin, *logwin;
|
WINDOW *mainwin, *statuswin, *logwin;
|
||||||
#endif
|
#endif
|
||||||
@ -2240,43 +2223,26 @@ static void push_curl_entry(struct curl_ent *ce, struct pool *pool)
|
|||||||
|
|
||||||
/* This is overkill, but at least we'll know accurately how much work is
|
/* This is overkill, but at least we'll know accurately how much work is
|
||||||
* queued to prevent ever being left without work */
|
* queued to prevent ever being left without work */
|
||||||
static void inc_queued(struct pool *pool)
|
static void inc_queued(void)
|
||||||
{
|
{
|
||||||
mutex_lock(stgd_lock);
|
mutex_lock(stgd_lock);
|
||||||
if (likely(pool))
|
|
||||||
pool->queued++;
|
|
||||||
total_queued++;
|
total_queued++;
|
||||||
mutex_unlock(stgd_lock);
|
mutex_unlock(stgd_lock);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void __dec_queued(struct pool *pool)
|
static void __dec_queued(void)
|
||||||
{
|
{
|
||||||
if (!total_queued)
|
if (total_queued)
|
||||||
return;
|
|
||||||
|
|
||||||
if (likely(pool))
|
|
||||||
pool->queued--;
|
|
||||||
total_queued--;
|
total_queued--;
|
||||||
}
|
}
|
||||||
|
|
||||||
static void dec_queued(struct pool *pool)
|
static void dec_queued(void)
|
||||||
{
|
{
|
||||||
mutex_lock(stgd_lock);
|
mutex_lock(stgd_lock);
|
||||||
__dec_queued(pool);
|
__dec_queued();
|
||||||
mutex_unlock(stgd_lock);
|
mutex_unlock(stgd_lock);
|
||||||
}
|
}
|
||||||
|
|
||||||
static int current_queued(void)
|
|
||||||
{
|
|
||||||
struct pool *pool = current_pool();
|
|
||||||
int ret;
|
|
||||||
|
|
||||||
mutex_lock(stgd_lock);
|
|
||||||
ret = pool->queued;
|
|
||||||
mutex_unlock(stgd_lock);
|
|
||||||
return ret;
|
|
||||||
}
|
|
||||||
|
|
||||||
static int global_queued(void)
|
static int global_queued(void)
|
||||||
{
|
{
|
||||||
int ret;
|
int ret;
|
||||||
@ -2617,7 +2583,6 @@ static void discard_stale(void)
|
|||||||
HASH_ITER(hh, staged_work, work, tmp) {
|
HASH_ITER(hh, staged_work, work, tmp) {
|
||||||
if (stale_work(work, false)) {
|
if (stale_work(work, false)) {
|
||||||
HASH_DEL(staged_work, work);
|
HASH_DEL(staged_work, work);
|
||||||
work->pool->staged--;
|
|
||||||
discard_work(work);
|
discard_work(work);
|
||||||
stale++;
|
stale++;
|
||||||
}
|
}
|
||||||
@ -2789,8 +2754,7 @@ static bool hash_push(struct work *work)
|
|||||||
mutex_lock(stgd_lock);
|
mutex_lock(stgd_lock);
|
||||||
if (likely(!getq->frozen)) {
|
if (likely(!getq->frozen)) {
|
||||||
HASH_ADD_INT(staged_work, id, work);
|
HASH_ADD_INT(staged_work, id, work);
|
||||||
work->pool->staged++;
|
__dec_queued();
|
||||||
__dec_queued(work->pool);
|
|
||||||
HASH_SORT(staged_work, tv_sort);
|
HASH_SORT(staged_work, tv_sort);
|
||||||
} else
|
} else
|
||||||
rc = false;
|
rc = false;
|
||||||
@ -3823,33 +3787,24 @@ static bool clone_available(void)
|
|||||||
|
|
||||||
bool queue_request(struct thr_info *thr, bool needed)
|
bool queue_request(struct thr_info *thr, bool needed)
|
||||||
{
|
{
|
||||||
int cq, cs, ts, tq, maxq;
|
|
||||||
bool lag, ret, qing;
|
|
||||||
struct workio_cmd *wc;
|
struct workio_cmd *wc;
|
||||||
|
bool lag, ret, qing;
|
||||||
|
int ps, ts, maxq;
|
||||||
|
|
||||||
inc_queued(NULL);
|
inc_queued();
|
||||||
|
|
||||||
maxq = opt_queue + mining_threads;
|
maxq = opt_queue + mining_threads;
|
||||||
lag = ret = qing = false;
|
lag = ret = qing = false;
|
||||||
cq = current_queued();
|
ps = pending_staged();
|
||||||
cs = current_staged();
|
ts = total_staged();
|
||||||
ts = pending_staged();
|
|
||||||
tq = global_queued();
|
|
||||||
|
|
||||||
if (needed && cq >= maxq && !ts && !opt_fail_only) {
|
if (ps >= maxq) {
|
||||||
/* If we're queueing work faster than we can stage it, consider
|
|
||||||
* the system lagging and allow work to be gathered from
|
|
||||||
* another pool if possible */
|
|
||||||
lag = true;
|
|
||||||
} else {
|
|
||||||
/* Test to make sure we have enough work for pools without rolltime
|
|
||||||
* and enough original work for pools with rolltime */
|
|
||||||
if (((cs || cq >= opt_queue) && ts >= maxq) ||
|
|
||||||
((cs || cq) && tq >= maxq)) {
|
|
||||||
ret = true;
|
ret = true;
|
||||||
goto out;
|
goto out;
|
||||||
}
|
}
|
||||||
}
|
|
||||||
|
if (needed && !ts && !opt_fail_only)
|
||||||
|
lag = true;
|
||||||
|
|
||||||
/* fill out work request message */
|
/* fill out work request message */
|
||||||
wc = calloc(1, sizeof(*wc));
|
wc = calloc(1, sizeof(*wc));
|
||||||
@ -3873,7 +3828,7 @@ bool queue_request(struct thr_info *thr, bool needed)
|
|||||||
qing = ret = true;
|
qing = ret = true;
|
||||||
out:
|
out:
|
||||||
if (!qing)
|
if (!qing)
|
||||||
dec_queued(NULL);
|
dec_queued();
|
||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -3889,7 +3844,6 @@ static struct work *hash_pop(const struct timespec *abstime)
|
|||||||
if (HASH_COUNT(staged_work)) {
|
if (HASH_COUNT(staged_work)) {
|
||||||
work = staged_work;
|
work = staged_work;
|
||||||
HASH_DEL(staged_work, work);
|
HASH_DEL(staged_work, work);
|
||||||
work->pool->staged--;
|
|
||||||
}
|
}
|
||||||
mutex_unlock(stgd_lock);
|
mutex_unlock(stgd_lock);
|
||||||
|
|
||||||
@ -3952,7 +3906,7 @@ static bool get_work(struct work *work, bool requested, struct thr_info *thr,
|
|||||||
struct timespec abstime = {0, 0};
|
struct timespec abstime = {0, 0};
|
||||||
struct timeval now;
|
struct timeval now;
|
||||||
struct work *work_heap;
|
struct work *work_heap;
|
||||||
int failures = 0, cq;
|
int failures = 0, tq;
|
||||||
struct pool *pool;
|
struct pool *pool;
|
||||||
|
|
||||||
/* Tell the watchdog thread this thread is waiting on getwork and
|
/* Tell the watchdog thread this thread is waiting on getwork and
|
||||||
@ -3965,10 +3919,10 @@ static bool get_work(struct work *work, bool requested, struct thr_info *thr,
|
|||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
|
|
||||||
cq = current_queued();
|
tq = global_queued();
|
||||||
retry:
|
retry:
|
||||||
pool = current_pool();
|
pool = current_pool();
|
||||||
if (!requested || cq < opt_queue) {
|
if (!requested || tq < opt_queue) {
|
||||||
if (unlikely(!queue_request(thr, true))) {
|
if (unlikely(!queue_request(thr, true))) {
|
||||||
applog(LOG_WARNING, "Failed to queue_request in get_work");
|
applog(LOG_WARNING, "Failed to queue_request in get_work");
|
||||||
goto out;
|
goto out;
|
||||||
@ -3981,7 +3935,7 @@ retry:
|
|||||||
goto out;
|
goto out;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (!pool->lagging && requested && !newreq && !pool_staged(pool) && cq >= mining_threads + opt_queue) {
|
if (!pool->lagging && requested && !newreq && !total_staged() && pending_staged() >= mining_threads + opt_queue) {
|
||||||
struct cgpu_info *cgpu = thr->cgpu;
|
struct cgpu_info *cgpu = thr->cgpu;
|
||||||
bool stalled = true;
|
bool stalled = true;
|
||||||
int i;
|
int i;
|
||||||
|
Loading…
Reference in New Issue
Block a user