mirror of
https://github.com/GOSTSec/sgminer
synced 2025-01-11 07:17:58 +00:00
Add mandatory and block fields to the work struct. Flag any shares that are detected as blocks as mandatory to submit, along with longpoll work from a previously rejecting pool.
This commit is contained in:
parent
1e3888336b
commit
dbe9056df4
31
cgminer.c
31
cgminer.c
@ -1625,7 +1625,6 @@ static bool submit_upstream_work(const struct work *work, CURL *curl)
|
||||
bool rolltime;
|
||||
uint32_t *hash32;
|
||||
char hashshow[64+1] = "";
|
||||
bool isblock;
|
||||
|
||||
#ifdef __BIG_ENDIAN__
|
||||
int swapcounter = 0;
|
||||
@ -1666,17 +1665,9 @@ static bool submit_upstream_work(const struct work *work, CURL *curl)
|
||||
res = json_object_get(val, "result");
|
||||
|
||||
if (!QUIET) {
|
||||
#ifndef MIPSEB
|
||||
// This one segfaults on my router for some reason
|
||||
isblock = regeneratehash(work);
|
||||
if (unlikely(isblock)) {
|
||||
pool->solved++;
|
||||
found_blocks++;
|
||||
}
|
||||
hash32 = (uint32_t *)(work->hash);
|
||||
sprintf(hashshow, "%08lx.%08lx%s", (unsigned long)(hash32[6]), (unsigned long)(hash32[5]),
|
||||
isblock ? " BLOCK!" : "");
|
||||
#endif
|
||||
work->block? " BLOCK!" : "");
|
||||
}
|
||||
|
||||
/* Theoretically threads could race when modifying accepted and
|
||||
@ -1829,6 +1820,7 @@ static void get_benchmark_work(struct work *work)
|
||||
size_t min_size = (work_size < bench_size ? work_size : bench_size);
|
||||
memset(work, 0, sizeof(work));
|
||||
memcpy(work, &bench_block, min_size);
|
||||
work->mandatory = true;
|
||||
}
|
||||
|
||||
static bool get_upstream_work(struct work *work, CURL *curl)
|
||||
@ -2164,7 +2156,7 @@ static bool stale_work(struct work *work, bool share)
|
||||
struct timeval now;
|
||||
struct pool *pool;
|
||||
|
||||
if (opt_benchmark)
|
||||
if (work->mandatory)
|
||||
return false;
|
||||
|
||||
gettimeofday(&now, NULL);
|
||||
@ -2184,6 +2176,16 @@ static bool stale_work(struct work *work, bool share)
|
||||
return false;
|
||||
}
|
||||
|
||||
static void check_solve(struct work *work)
|
||||
{
|
||||
work->block = regeneratehash(work);
|
||||
if (unlikely(work->block)) {
|
||||
work->pool->solved++;
|
||||
found_blocks++;
|
||||
work->mandatory = true;
|
||||
applog(LOG_NOTICE, "Found block for pool %d!", work->pool);
|
||||
}
|
||||
}
|
||||
|
||||
static void *submit_work_thread(void *userdata)
|
||||
{
|
||||
@ -2197,6 +2199,8 @@ static void *submit_work_thread(void *userdata)
|
||||
|
||||
applog(LOG_DEBUG, "Creating extra submit work thread");
|
||||
|
||||
check_solve(work);
|
||||
|
||||
if (stale_work(work, true)) {
|
||||
if (opt_submit_stale)
|
||||
applog(LOG_NOTICE, "Stale share detected, submitting as user requested");
|
||||
@ -2479,7 +2483,7 @@ static void test_work_current(struct work *work)
|
||||
{
|
||||
char *hexstr;
|
||||
|
||||
if (opt_benchmark)
|
||||
if (work->mandatory)
|
||||
return;
|
||||
|
||||
hexstr = bin2hex(work->data, 18);
|
||||
@ -4009,6 +4013,9 @@ static void convert_to_work(json_t *val, bool rolltime, struct pool *pool)
|
||||
work->rolltime = rolltime;
|
||||
work->longpoll = true;
|
||||
|
||||
if (pool->enabled == POOL_REJECTING)
|
||||
work->mandatory = true;
|
||||
|
||||
/* We'll be checking this work item twice, but we already know it's
|
||||
* from a new block so explicitly force the new block detection now
|
||||
* rather than waiting for it to hit the stage thread. This also
|
||||
|
Loading…
Reference in New Issue
Block a user