|
|
|
@ -1485,16 +1485,16 @@ static void bflsc_thread_enable(struct thr_info *thr)
@@ -1485,16 +1485,16 @@ static void bflsc_thread_enable(struct thr_info *thr)
|
|
|
|
|
bflsc_initialise(bflsc); |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
static bool bflsc_send_work(struct cgpu_info *bflsc, int dev, struct work *work, |
|
|
|
|
bool mandatory) |
|
|
|
|
static bool bflsc_send_work(struct cgpu_info *bflsc, int dev, bool mandatory) |
|
|
|
|
{ |
|
|
|
|
struct bflsc_info *sc_info = (struct bflsc_info *)(bflsc->device_data); |
|
|
|
|
struct FullNonceRangeJob data; |
|
|
|
|
char buf[BFLSC_BUFSIZ+1]; |
|
|
|
|
bool sent, ret = false; |
|
|
|
|
struct work *work; |
|
|
|
|
int err, amount; |
|
|
|
|
int len, try; |
|
|
|
|
int stage; |
|
|
|
|
bool sent; |
|
|
|
|
|
|
|
|
|
// Device is gone
|
|
|
|
|
if (bflsc->usbinfo.nodev) |
|
|
|
@ -1506,8 +1506,6 @@ static bool bflsc_send_work(struct cgpu_info *bflsc, int dev, struct work *work,
@@ -1506,8 +1506,6 @@ static bool bflsc_send_work(struct cgpu_info *bflsc, int dev, struct work *work,
|
|
|
|
|
|
|
|
|
|
// Initially code only deals with sending one work item
|
|
|
|
|
data.payloadSize = BFLSC_JOBSIZ; |
|
|
|
|
memcpy(data.midState, work->midstate, MIDSTATE_BYTES); |
|
|
|
|
memcpy(data.blockData, work->data + MERKLE_OFFSET, MERKLE_BYTES); |
|
|
|
|
data.endOfBlock = BFLSC_EOB; |
|
|
|
|
|
|
|
|
|
len = sizeof(struct FullNonceRangeJob); |
|
|
|
@ -1519,9 +1517,16 @@ static bool bflsc_send_work(struct cgpu_info *bflsc, int dev, struct work *work,
@@ -1519,9 +1517,16 @@ static bool bflsc_send_work(struct cgpu_info *bflsc, int dev, struct work *work,
|
|
|
|
|
mutex_lock(&(bflsc->device_mutex)); |
|
|
|
|
else { |
|
|
|
|
if (mutex_trylock(&bflsc->device_mutex)) |
|
|
|
|
return false; |
|
|
|
|
return ret; |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
work = get_queued(bflsc); |
|
|
|
|
if (unlikely(!work)) { |
|
|
|
|
mutex_unlock(&bflsc->device_mutex); |
|
|
|
|
return ret; |
|
|
|
|
} |
|
|
|
|
memcpy(data.midState, work->midstate, MIDSTATE_BYTES); |
|
|
|
|
memcpy(data.blockData, work->data + MERKLE_OFFSET, MERKLE_BYTES); |
|
|
|
|
try = 0; |
|
|
|
|
re_send: |
|
|
|
|
err = send_recv_ds(bflsc, dev, &stage, &sent, &amount, |
|
|
|
@ -1534,7 +1539,7 @@ re_send:
@@ -1534,7 +1539,7 @@ re_send:
|
|
|
|
|
case 1: |
|
|
|
|
if (!sent) { |
|
|
|
|
bflsc_applog(bflsc, dev, C_REQUESTQUEJOB, amount, err); |
|
|
|
|
return false; |
|
|
|
|
goto out; |
|
|
|
|
} else { |
|
|
|
|
// TODO: handle other errors ...
|
|
|
|
|
|
|
|
|
@ -1544,13 +1549,13 @@ re_send:
@@ -1544,13 +1549,13 @@ re_send:
|
|
|
|
|
goto re_send; |
|
|
|
|
|
|
|
|
|
bflsc_applog(bflsc, dev, C_REQUESTQUEJOBSTATUS, amount, err); |
|
|
|
|
return false; |
|
|
|
|
goto out; |
|
|
|
|
} |
|
|
|
|
break; |
|
|
|
|
case 2: |
|
|
|
|
if (!sent) { |
|
|
|
|
bflsc_applog(bflsc, dev, C_QUEJOB, amount, err); |
|
|
|
|
return false; |
|
|
|
|
goto out; |
|
|
|
|
} else { |
|
|
|
|
if (!isokerr(err, buf, amount)) { |
|
|
|
|
// TODO: check for QUEUE FULL and set work_queued to sc_info->que_size
|
|
|
|
@ -1563,7 +1568,7 @@ re_send:
@@ -1563,7 +1568,7 @@ re_send:
|
|
|
|
|
goto re_send; |
|
|
|
|
|
|
|
|
|
bflsc_applog(bflsc, dev, C_QUEJOBSTATUS, amount, err); |
|
|
|
|
return false; |
|
|
|
|
goto out; |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
break; |
|
|
|
@ -1574,13 +1579,13 @@ re_send:
@@ -1574,13 +1579,13 @@ re_send:
|
|
|
|
|
if (err < 0 || amount != BFLSC_QJOB_LEN) { |
|
|
|
|
mutex_unlock(&(bflsc->device_mutex)); |
|
|
|
|
bflsc_applog(bflsc, dev, C_REQUESTQUEJOB, amount, err); |
|
|
|
|
return false; |
|
|
|
|
goto out; |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
if (!getok(bflsc, C_REQUESTQUEJOBSTATUS, &err, &amount)) { |
|
|
|
|
mutex_unlock(&(bflsc->device_mutex)); |
|
|
|
|
bflsc_applog(bflsc, dev, C_REQUESTQUEJOBSTATUS, amount, err); |
|
|
|
|
return false; |
|
|
|
|
goto out; |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
len = sizeof(struct FullNonceRangeJob); |
|
|
|
@ -1589,7 +1594,7 @@ re_send:
@@ -1589,7 +1594,7 @@ re_send:
|
|
|
|
|
if (err < 0 || amount != len) { |
|
|
|
|
mutex_unlock(&(bflsc->device_mutex)); |
|
|
|
|
bflsc_applog(bflsc, dev, C_QUEJOB, amount, err); |
|
|
|
|
return false; |
|
|
|
|
goto out; |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
if (!getokerr(bflsc, C_QUEJOBSTATUS, &err, &amount, buf, sizeof(buf))) { |
|
|
|
@ -1603,7 +1608,7 @@ re_send:
@@ -1603,7 +1608,7 @@ re_send:
|
|
|
|
|
|
|
|
|
|
mutex_unlock(&(bflsc->device_mutex)); |
|
|
|
|
bflsc_applog(bflsc, dev, C_QUEJOBSTATUS, amount, err); |
|
|
|
|
return false; |
|
|
|
|
goto out; |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
mutex_unlock(&(bflsc->device_mutex)); |
|
|
|
@ -1614,14 +1619,16 @@ re_send:
@@ -1614,14 +1619,16 @@ re_send:
|
|
|
|
|
wr_unlock(&(sc_info->stat_lock)); |
|
|
|
|
|
|
|
|
|
work->subid = dev; |
|
|
|
|
|
|
|
|
|
return true; |
|
|
|
|
ret = true; |
|
|
|
|
out: |
|
|
|
|
if (unlikely(!ret)) |
|
|
|
|
work_completed(bflsc, work); |
|
|
|
|
return ret; |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
static bool bflsc_queue_full(struct cgpu_info *bflsc) |
|
|
|
|
{ |
|
|
|
|
struct bflsc_info *sc_info = (struct bflsc_info *)(bflsc->device_data); |
|
|
|
|
struct work *work = NULL; |
|
|
|
|
int i, dev, tried, que; |
|
|
|
|
bool ret = false; |
|
|
|
|
int tries = 0; |
|
|
|
@ -1673,19 +1680,12 @@ static bool bflsc_queue_full(struct cgpu_info *bflsc)
@@ -1673,19 +1680,12 @@ static bool bflsc_queue_full(struct cgpu_info *bflsc)
|
|
|
|
|
break; |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
if (!work) |
|
|
|
|
work = get_queued(bflsc); |
|
|
|
|
if (unlikely(!work)) |
|
|
|
|
break; |
|
|
|
|
if (bflsc_send_work(bflsc, dev, work, mandatory)) { |
|
|
|
|
work = NULL; |
|
|
|
|
if (bflsc_send_work(bflsc, dev, mandatory)) |
|
|
|
|
break; |
|
|
|
|
} else |
|
|
|
|
else |
|
|
|
|
tried = dev; |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
if (unlikely(work)) |
|
|
|
|
work_completed(bflsc, work); |
|
|
|
|
return ret; |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|