|
|
|
#ifndef __MINER_H__
|
|
|
|
#define __MINER_H__
|
|
|
|
|
|
|
|
#ifdef __cplusplus
|
|
|
|
extern "C" {
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#include "cpuminer-config.h"
|
|
|
|
|
|
|
|
#include <stdbool.h>
|
|
|
|
#include <inttypes.h>
|
|
|
|
#include <sys/time.h>
|
|
|
|
#include <pthread.h>
|
|
|
|
#include <jansson.h>
|
|
|
|
#include <curl/curl.h>
|
|
|
|
|
|
|
|
#ifdef WIN32
|
|
|
|
#define snprintf(...) _snprintf(__VA_ARGS__)
|
|
|
|
#define strdup(x) _strdup(x)
|
|
|
|
#define strncasecmp(x,y,z) _strnicmp(x,y,z)
|
|
|
|
#define strcasecmp(x,y) _stricmp(x,y)
|
|
|
|
typedef int ssize_t;
|
|
|
|
#undef HAVE_ALLOCA_H
|
|
|
|
#undef HAVE_SYSLOG_H
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#ifdef STDC_HEADERS
|
|
|
|
# include <stdlib.h>
|
|
|
|
# include <stddef.h>
|
|
|
|
#else
|
|
|
|
# ifdef HAVE_STDLIB_H
|
|
|
|
# include <stdlib.h>
|
|
|
|
# endif
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#ifdef HAVE_ALLOCA_H
|
|
|
|
# include <alloca.h>
|
|
|
|
#elif !defined alloca
|
|
|
|
# ifdef __GNUC__
|
|
|
|
# define alloca __builtin_alloca
|
|
|
|
# elif defined _AIX
|
|
|
|
# define alloca __alloca
|
|
|
|
# elif defined _MSC_VER
|
|
|
|
# include <malloc.h>
|
|
|
|
# define alloca _alloca
|
|
|
|
# elif !defined HAVE_ALLOCA
|
|
|
|
# ifdef __cplusplus
|
|
|
|
extern "C"
|
|
|
|
# endif
|
|
|
|
void *alloca (size_t);
|
|
|
|
# endif
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#ifdef HAVE_SYSLOG_H
|
|
|
|
#include <syslog.h>
|
|
|
|
#define LOG_BLUE 0x10 /* unique value */
|
|
|
|
#else
|
|
|
|
enum {
|
|
|
|
LOG_ERR,
|
|
|
|
LOG_WARNING,
|
|
|
|
LOG_NOTICE,
|
|
|
|
LOG_INFO,
|
|
|
|
LOG_DEBUG,
|
|
|
|
/* custom notices */
|
|
|
|
LOG_BLUE = 0x10,
|
|
|
|
};
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#undef unlikely
|
|
|
|
#undef likely
|
|
|
|
#if defined(__GNUC__) && (__GNUC__ > 2) && defined(__OPTIMIZE__)
|
|
|
|
#define unlikely(expr) (__builtin_expect(!!(expr), 0))
|
|
|
|
#define likely(expr) (__builtin_expect(!!(expr), 1))
|
|
|
|
#else
|
|
|
|
#define unlikely(expr) (expr)
|
|
|
|
#define likely(expr) (expr)
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#ifndef ARRAY_SIZE
|
|
|
|
#define ARRAY_SIZE(arr) (sizeof(arr) / sizeof((arr)[0]))
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#if ((__GNUC__ > 4) || (__GNUC__ == 4 && __GNUC_MINOR__ >= 3))
|
|
|
|
#define WANT_BUILTIN_BSWAP
|
|
|
|
#else
|
|
|
|
#define bswap_32(x) ((((x) << 24) & 0xff000000u) | (((x) << 8) & 0x00ff0000u) \
|
|
|
|
| (((x) >> 8) & 0x0000ff00u) | (((x) >> 24) & 0x000000ffu))
|
|
|
|
#endif
|
|
|
|
|
|
|
|
static inline uint32_t swab32(uint32_t v)
|
|
|
|
{
|
|
|
|
#ifdef WANT_BUILTIN_BSWAP
|
|
|
|
return __builtin_bswap32(v);
|
|
|
|
#else
|
|
|
|
return bswap_32(v);
|
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
|
|
|
#ifdef HAVE_SYS_ENDIAN_H
|
|
|
|
#include <sys/endian.h>
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#if !HAVE_DECL_BE32DEC
|
|
|
|
static inline uint32_t be32dec(const void *pp)
|
|
|
|
{
|
|
|
|
const uint8_t *p = (uint8_t const *)pp;
|
|
|
|
return ((uint32_t)(p[3]) + ((uint32_t)(p[2]) << 8) +
|
|
|
|
((uint32_t)(p[1]) << 16) + ((uint32_t)(p[0]) << 24));
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#if !HAVE_DECL_LE32DEC
|
|
|
|
static inline uint32_t le32dec(const void *pp)
|
|
|
|
{
|
|
|
|
const uint8_t *p = (uint8_t const *)pp;
|
|
|
|
return ((uint32_t)(p[0]) + ((uint32_t)(p[1]) << 8) +
|
|
|
|
((uint32_t)(p[2]) << 16) + ((uint32_t)(p[3]) << 24));
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#if !HAVE_DECL_BE32ENC
|
|
|
|
static inline void be32enc(void *pp, uint32_t x)
|
|
|
|
{
|
|
|
|
uint8_t *p = (uint8_t *)pp;
|
|
|
|
p[3] = x & 0xff;
|
|
|
|
p[2] = (x >> 8) & 0xff;
|
|
|
|
p[1] = (x >> 16) & 0xff;
|
|
|
|
p[0] = (x >> 24) & 0xff;
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#if !HAVE_DECL_LE32ENC
|
|
|
|
static inline void le32enc(void *pp, uint32_t x)
|
|
|
|
{
|
|
|
|
uint8_t *p = (uint8_t *)pp;
|
|
|
|
p[0] = x & 0xff;
|
|
|
|
p[1] = (x >> 8) & 0xff;
|
|
|
|
p[2] = (x >> 16) & 0xff;
|
|
|
|
p[3] = (x >> 24) & 0xff;
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#if !HAVE_DECL_BE16DEC
|
|
|
|
static inline uint16_t be16dec(const void *pp)
|
|
|
|
{
|
|
|
|
const uint8_t *p = (uint8_t const *)pp;
|
|
|
|
return ((uint16_t)(p[1]) + ((uint16_t)(p[0]) << 8));
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#if !HAVE_DECL_BE16ENC
|
|
|
|
static inline void be16enc(void *pp, uint16_t x)
|
|
|
|
{
|
|
|
|
uint8_t *p = (uint8_t *)pp;
|
|
|
|
p[1] = x & 0xff;
|
|
|
|
p[0] = (x >> 8) & 0xff;
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#if !HAVE_DECL_LE16DEC
|
|
|
|
static inline uint16_t le16dec(const void *pp)
|
|
|
|
{
|
|
|
|
const uint8_t *p = (uint8_t const *)pp;
|
|
|
|
return ((uint16_t)(p[0]) + ((uint16_t)(p[1]) << 8));
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#if !HAVE_DECL_LE16ENC
|
|
|
|
static inline void le16enc(void *pp, uint16_t x)
|
|
|
|
{
|
|
|
|
uint8_t *p = (uint8_t *)pp;
|
|
|
|
p[0] = x & 0xff;
|
|
|
|
p[1] = (x >> 8) & 0xff;
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#if JANSSON_MAJOR_VERSION >= 2
|
|
|
|
#define JSON_LOADS(str, err_ptr) json_loads((str), 0, (err_ptr))
|
|
|
|
#else
|
|
|
|
#define JSON_LOADS(str, err_ptr) json_loads((str), (err_ptr))
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#define USER_AGENT PACKAGE_NAME "/" PACKAGE_VERSION
|
|
|
|
|
|
|
|
void sha256_init(uint32_t *state);
|
|
|
|
void sha256_transform(uint32_t *state, const uint32_t *block, int swap);
|
|
|
|
void sha256d(unsigned char *hash, const unsigned char *data, int len);
|
|
|
|
|
|
|
|
#if defined(__ARM_NEON__) || defined(__i386__) || defined(__x86_64__)
|
|
|
|
#define HAVE_SHA256_4WAY 0
|
|
|
|
int sha256_use_4way();
|
|
|
|
void sha256_init_4way(uint32_t *state);
|
|
|
|
void sha256_transform_4way(uint32_t *state, const uint32_t *block, int swap);
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#if defined(__x86_64__) && defined(USE_AVX2)
|
|
|
|
#define HAVE_SHA256_8WAY 0
|
|
|
|
int sha256_use_8way();
|
|
|
|
void sha256_init_8way(uint32_t *state);
|
|
|
|
void sha256_transform_8way(uint32_t *state, const uint32_t *block, int swap);
|
|
|
|
#endif
|
|
|
|
|
|
|
|
extern int scanhash_sha256d(int thr_id, uint32_t *pdata,
|
|
|
|
const uint32_t *ptarget, uint32_t max_nonce, unsigned long *hashes_done);
|
|
|
|
|
|
|
|
extern unsigned char *scrypt_buffer_alloc();
|
|
|
|
|
|
|
|
extern int scanhash_scrypt(int thr_id, uint32_t *pdata,
|
|
|
|
unsigned char *scratchbuf, const uint32_t *ptarget,
|
|
|
|
uint32_t max_nonce, unsigned long *hashes_done);
|
|
|
|
|
|
|
|
extern int scanhash_heavy(int thr_id, uint32_t *pdata,
|
|
|
|
const uint32_t *ptarget, uint32_t max_nonce,
|
|
|
|
unsigned long *hashes_done, uint32_t maxvote, int blocklen);
|
|
|
|
|
|
|
|
extern int scanhash_fugue256(int thr_id, uint32_t *pdata,
|
|
|
|
const uint32_t *ptarget, uint32_t max_nonce,
|
|
|
|
unsigned long *hashes_done);
|
|
|
|
|
|
|
|
extern int scanhash_groestlcoin(int thr_id, uint32_t *pdata,
|
|
|
|
const uint32_t *ptarget, uint32_t max_nonce,
|
|
|
|
unsigned long *hashes_done);
|
|
|
|
|
|
|
|
extern int scanhash_myriad(int thr_id, uint32_t *pdata,
|
|
|
|
const uint32_t *ptarget, uint32_t max_nonce,
|
|
|
|
unsigned long *hashes_done);
|
|
|
|
|
|
|
|
extern int scanhash_jackpot(int thr_id, uint32_t *pdata,
|
|
|
|
const uint32_t *ptarget, uint32_t max_nonce,
|
|
|
|
unsigned long *hashes_done);
|
|
|
|
|
|
|
|
extern int scanhash_quark(int thr_id, uint32_t *pdata,
|
|
|
|
const uint32_t *ptarget, uint32_t max_nonce,
|
|
|
|
unsigned long *hashes_done);
|
|
|
|
|
|
|
|
extern int scanhash_anime(int thr_id, uint32_t *pdata,
|
|
|
|
const uint32_t *ptarget, uint32_t max_nonce,
|
|
|
|
unsigned long *hashes_done);
|
|
|
|
|
|
|
|
extern int scanhash_fresh(int thr_id, uint32_t *pdata,
|
|
|
|
const uint32_t *ptarget, uint32_t max_nonce,
|
|
|
|
unsigned long *hashes_done);
|
|
|
|
|
Release v1.4 with blake (NEOS)
Blake256: squashed commit...
Squashed commit of the following:
commit c370208bc92ef16557f66e5391faf2b1ad47726f
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Wed Sep 3 13:53:01 2014 +0200
hashlog: prepare store of scanned range
commit e2cf49a5e956f03deafd266d1a0dd087a2041c99
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Wed Sep 3 12:54:13 2014 +0200
stratum: store server time offset in context
commit 1a4391d7ff21397a128abf031f92733a8ac47437
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 12:40:52 2014 +0200
hashlog: prevent double computing on jobs already done
commit 049e57730116685755bd3ff214f0793cce7c773b
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Wed Sep 3 09:49:14 2014 +0200
tmp blake log
commit 43d3e93e1a97e569ead2437f759c6b8423d30c0a
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Wed Sep 3 09:29:51 2014 +0200
blake: set a max throughput
commit 7e595a36ea69027c8a28023399540a761e7686c3
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 21:13:37 2014 +0200
blake: cleanup, remove d_hash buf, not in a chain
host: only bencode if gpu hash was found
commit de80c7e9d1448f15541d08c5dbbf372d5bfeba48
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 12:40:44 2014 +0200
blake: remove unused parameter and fix index in d_hash
that reduce the speed to 92MH/s but the next commit
give us 30 more
so, todo: merge the whole checkhash proc in gpu_hash
and remove this d_hash buffer...
commit 2d42ae6de586a6ae8cbfd01806a273fd5cc4b262
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 05:09:31 2014 +0200
stratum: handle a small cache of submitted jobs
Prevent to send duplicated shares on some pools like hashharder..
This cache keeps submitted job/nounces of the last 15 minutes
so, remove exit on repeated duplicate shares,
the submitted cache now handles this problem.
Signed-off-by: Tanguy Pruvot <tanguy.pruvot@gmail.com>
commit 1b8c3c12fa5bb83afbb02f9d5f60586939f36d86
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 03:38:57 2014 +0200
debug: a new boolean to log or not json rpc data
commit 1f99aae0ff621f4f85f119d811a3f1a8d2204f60
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 18:49:23 2014 +0200
exit on repeated duplicate shares (to enhance)
create a new function proper_exit() to do common stuff on exit...
commit 530732458add6c4c3836606d028930f3581c0a5f
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 12:22:51 2014 +0200
blake: use a constant for threads, reduce mallocated d_hash size
and clean a bit more...
commit 0aeac878ef60840f3123354037cd56a89d2e94e6
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 06:12:55 2014 +0200
blake: tune up and cleanup, ~100 MH/s on a normal 750Ti
tested on linux and windows (x86 binary)...
but there is a high number of duplicated shares... weird
commit 4a52d0553b0076b984be480725fa67689c544647
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 10:22:32 2014 +0200
debug: show json methods, hide hash/target if ok
commit 1fb9becc1f2b6a15d8ccea4d8314df9ddf0af4ed
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 08:44:19 2014 +0200
cpu-miner: sort algos by name, show reject reason
commit bfe96c49b0bf321ed0776cb1cf31c4fe8a0a8b8d
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Aug 25 11:21:06 2014 +0200
release 1.4, update README...
commit c17d11e37758c37762a7664a731fda6e9a5454b1
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Sun Aug 31 08:57:48 2014 +0200
add "blake" 256, 14 rounds (for NEOS blake, not BlakeCoin)
also remove "missing" file, its old and not compatible with ubuntu 14.04
to test on windows
blake: clean and optimize
Release v1.4 with blake (NEOS)
10 years ago
|
|
|
extern int scanhash_blake32(int thr_id, uint32_t *pdata,
|
|
|
|
const uint32_t *ptarget, uint32_t max_nonce,
|
|
|
|
unsigned long *hashes_done);
|
|
|
|
|
|
|
|
extern int scanhash_nist5(int thr_id, uint32_t *pdata,
|
|
|
|
const uint32_t *ptarget, uint32_t max_nonce,
|
|
|
|
unsigned long *hashes_done);
|
|
|
|
|
|
|
|
extern int scanhash_whc(int thr_id, uint32_t *pdata,
|
|
|
|
const uint32_t *ptarget, uint32_t max_nonce,
|
|
|
|
unsigned long *hashes_done);
|
|
|
|
|
|
|
|
extern int scanhash_x11(int thr_id, uint32_t *pdata,
|
|
|
|
const uint32_t *ptarget, uint32_t max_nonce,
|
|
|
|
unsigned long *hashes_done);
|
|
|
|
|
|
|
|
extern int scanhash_x13(int thr_id, uint32_t *pdata,
|
|
|
|
const uint32_t *ptarget, uint32_t max_nonce,
|
|
|
|
unsigned long *hashes_done);
|
|
|
|
|
|
|
|
extern int scanhash_x14(int thr_id, uint32_t *pdata,
|
|
|
|
const uint32_t *ptarget, uint32_t max_nonce,
|
|
|
|
unsigned long *hashes_done);
|
|
|
|
|
|
|
|
extern int scanhash_x15(int thr_id, uint32_t *pdata,
|
|
|
|
const uint32_t *ptarget, uint32_t max_nonce,
|
|
|
|
unsigned long *hashes_done);
|
|
|
|
|
|
|
|
extern int scanhash_x17(int thr_id, uint32_t *pdata,
|
|
|
|
const uint32_t *ptarget, uint32_t max_nonce,
|
|
|
|
unsigned long *hashes_done);
|
|
|
|
|
|
|
|
struct thr_info {
|
|
|
|
int id;
|
|
|
|
pthread_t pth;
|
|
|
|
struct thread_q *q;
|
|
|
|
};
|
|
|
|
|
|
|
|
struct work_restart {
|
|
|
|
volatile unsigned long restart;
|
|
|
|
char padding[128 - sizeof(unsigned long)];
|
|
|
|
};
|
|
|
|
|
|
|
|
extern bool opt_debug;
|
Release v1.4 with blake (NEOS)
Blake256: squashed commit...
Squashed commit of the following:
commit c370208bc92ef16557f66e5391faf2b1ad47726f
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Wed Sep 3 13:53:01 2014 +0200
hashlog: prepare store of scanned range
commit e2cf49a5e956f03deafd266d1a0dd087a2041c99
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Wed Sep 3 12:54:13 2014 +0200
stratum: store server time offset in context
commit 1a4391d7ff21397a128abf031f92733a8ac47437
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 12:40:52 2014 +0200
hashlog: prevent double computing on jobs already done
commit 049e57730116685755bd3ff214f0793cce7c773b
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Wed Sep 3 09:49:14 2014 +0200
tmp blake log
commit 43d3e93e1a97e569ead2437f759c6b8423d30c0a
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Wed Sep 3 09:29:51 2014 +0200
blake: set a max throughput
commit 7e595a36ea69027c8a28023399540a761e7686c3
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 21:13:37 2014 +0200
blake: cleanup, remove d_hash buf, not in a chain
host: only bencode if gpu hash was found
commit de80c7e9d1448f15541d08c5dbbf372d5bfeba48
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 12:40:44 2014 +0200
blake: remove unused parameter and fix index in d_hash
that reduce the speed to 92MH/s but the next commit
give us 30 more
so, todo: merge the whole checkhash proc in gpu_hash
and remove this d_hash buffer...
commit 2d42ae6de586a6ae8cbfd01806a273fd5cc4b262
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 05:09:31 2014 +0200
stratum: handle a small cache of submitted jobs
Prevent to send duplicated shares on some pools like hashharder..
This cache keeps submitted job/nounces of the last 15 minutes
so, remove exit on repeated duplicate shares,
the submitted cache now handles this problem.
Signed-off-by: Tanguy Pruvot <tanguy.pruvot@gmail.com>
commit 1b8c3c12fa5bb83afbb02f9d5f60586939f36d86
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 03:38:57 2014 +0200
debug: a new boolean to log or not json rpc data
commit 1f99aae0ff621f4f85f119d811a3f1a8d2204f60
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 18:49:23 2014 +0200
exit on repeated duplicate shares (to enhance)
create a new function proper_exit() to do common stuff on exit...
commit 530732458add6c4c3836606d028930f3581c0a5f
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 12:22:51 2014 +0200
blake: use a constant for threads, reduce mallocated d_hash size
and clean a bit more...
commit 0aeac878ef60840f3123354037cd56a89d2e94e6
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 06:12:55 2014 +0200
blake: tune up and cleanup, ~100 MH/s on a normal 750Ti
tested on linux and windows (x86 binary)...
but there is a high number of duplicated shares... weird
commit 4a52d0553b0076b984be480725fa67689c544647
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 10:22:32 2014 +0200
debug: show json methods, hide hash/target if ok
commit 1fb9becc1f2b6a15d8ccea4d8314df9ddf0af4ed
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 08:44:19 2014 +0200
cpu-miner: sort algos by name, show reject reason
commit bfe96c49b0bf321ed0776cb1cf31c4fe8a0a8b8d
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Aug 25 11:21:06 2014 +0200
release 1.4, update README...
commit c17d11e37758c37762a7664a731fda6e9a5454b1
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Sun Aug 31 08:57:48 2014 +0200
add "blake" 256, 14 rounds (for NEOS blake, not BlakeCoin)
also remove "missing" file, its old and not compatible with ubuntu 14.04
to test on windows
blake: clean and optimize
Release v1.4 with blake (NEOS)
10 years ago
|
|
|
extern bool opt_debug_rpc;
|
|
|
|
extern bool opt_quiet;
|
|
|
|
extern bool opt_protocol;
|
|
|
|
extern int opt_timeout;
|
|
|
|
extern bool want_longpoll;
|
|
|
|
extern bool have_longpoll;
|
|
|
|
extern bool want_stratum;
|
|
|
|
extern bool have_stratum;
|
|
|
|
extern char *opt_cert;
|
|
|
|
extern char *opt_proxy;
|
|
|
|
extern long opt_proxy_type;
|
|
|
|
extern bool use_syslog;
|
|
|
|
extern bool use_colors;
|
|
|
|
extern pthread_mutex_t applog_lock;
|
|
|
|
extern struct thr_info *thr_info;
|
|
|
|
extern int longpoll_thr_id;
|
|
|
|
extern int stratum_thr_id;
|
|
|
|
extern struct work_restart *work_restart;
|
|
|
|
extern bool opt_trust_pool;
|
|
|
|
extern uint16_t opt_vote;
|
|
|
|
|
|
|
|
#define CL_N "\x1B[0m"
|
|
|
|
#define CL_RED "\x1B[31m"
|
|
|
|
#define CL_GRN "\x1B[32m"
|
|
|
|
#define CL_YLW "\x1B[33m"
|
|
|
|
#define CL_BLU "\x1B[34m"
|
|
|
|
#define CL_MAG "\x1B[35m"
|
|
|
|
#define CL_CYN "\x1B[36m"
|
|
|
|
|
|
|
|
#define CL_BLK "\x1B[22;30m" /* black */
|
|
|
|
#define CL_RD2 "\x1B[22;31m" /* red */
|
|
|
|
#define CL_GR2 "\x1B[22;32m" /* green */
|
Release v1.4 with blake (NEOS)
Blake256: squashed commit...
Squashed commit of the following:
commit c370208bc92ef16557f66e5391faf2b1ad47726f
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Wed Sep 3 13:53:01 2014 +0200
hashlog: prepare store of scanned range
commit e2cf49a5e956f03deafd266d1a0dd087a2041c99
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Wed Sep 3 12:54:13 2014 +0200
stratum: store server time offset in context
commit 1a4391d7ff21397a128abf031f92733a8ac47437
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 12:40:52 2014 +0200
hashlog: prevent double computing on jobs already done
commit 049e57730116685755bd3ff214f0793cce7c773b
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Wed Sep 3 09:49:14 2014 +0200
tmp blake log
commit 43d3e93e1a97e569ead2437f759c6b8423d30c0a
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Wed Sep 3 09:29:51 2014 +0200
blake: set a max throughput
commit 7e595a36ea69027c8a28023399540a761e7686c3
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 21:13:37 2014 +0200
blake: cleanup, remove d_hash buf, not in a chain
host: only bencode if gpu hash was found
commit de80c7e9d1448f15541d08c5dbbf372d5bfeba48
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 12:40:44 2014 +0200
blake: remove unused parameter and fix index in d_hash
that reduce the speed to 92MH/s but the next commit
give us 30 more
so, todo: merge the whole checkhash proc in gpu_hash
and remove this d_hash buffer...
commit 2d42ae6de586a6ae8cbfd01806a273fd5cc4b262
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 05:09:31 2014 +0200
stratum: handle a small cache of submitted jobs
Prevent to send duplicated shares on some pools like hashharder..
This cache keeps submitted job/nounces of the last 15 minutes
so, remove exit on repeated duplicate shares,
the submitted cache now handles this problem.
Signed-off-by: Tanguy Pruvot <tanguy.pruvot@gmail.com>
commit 1b8c3c12fa5bb83afbb02f9d5f60586939f36d86
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 03:38:57 2014 +0200
debug: a new boolean to log or not json rpc data
commit 1f99aae0ff621f4f85f119d811a3f1a8d2204f60
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 18:49:23 2014 +0200
exit on repeated duplicate shares (to enhance)
create a new function proper_exit() to do common stuff on exit...
commit 530732458add6c4c3836606d028930f3581c0a5f
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 12:22:51 2014 +0200
blake: use a constant for threads, reduce mallocated d_hash size
and clean a bit more...
commit 0aeac878ef60840f3123354037cd56a89d2e94e6
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 06:12:55 2014 +0200
blake: tune up and cleanup, ~100 MH/s on a normal 750Ti
tested on linux and windows (x86 binary)...
but there is a high number of duplicated shares... weird
commit 4a52d0553b0076b984be480725fa67689c544647
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 10:22:32 2014 +0200
debug: show json methods, hide hash/target if ok
commit 1fb9becc1f2b6a15d8ccea4d8314df9ddf0af4ed
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 08:44:19 2014 +0200
cpu-miner: sort algos by name, show reject reason
commit bfe96c49b0bf321ed0776cb1cf31c4fe8a0a8b8d
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Aug 25 11:21:06 2014 +0200
release 1.4, update README...
commit c17d11e37758c37762a7664a731fda6e9a5454b1
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Sun Aug 31 08:57:48 2014 +0200
add "blake" 256, 14 rounds (for NEOS blake, not BlakeCoin)
also remove "missing" file, its old and not compatible with ubuntu 14.04
to test on windows
blake: clean and optimize
Release v1.4 with blake (NEOS)
10 years ago
|
|
|
#define CL_YL2 "\x1B[22;33m" /* dark yellow */
|
|
|
|
#define CL_BL2 "\x1B[22;34m" /* blue */
|
|
|
|
#define CL_MA2 "\x1B[22;35m" /* magenta */
|
|
|
|
#define CL_CY2 "\x1B[22;36m" /* cyan */
|
|
|
|
#define CL_SIL "\x1B[22;37m" /* gray */
|
|
|
|
|
|
|
|
#define CL_GRY "\x1B[01;30m" /* dark gray */
|
|
|
|
#define CL_LRD "\x1B[01;31m" /* light red */
|
|
|
|
#define CL_LGR "\x1B[01;32m" /* light green */
|
Release v1.4 with blake (NEOS)
Blake256: squashed commit...
Squashed commit of the following:
commit c370208bc92ef16557f66e5391faf2b1ad47726f
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Wed Sep 3 13:53:01 2014 +0200
hashlog: prepare store of scanned range
commit e2cf49a5e956f03deafd266d1a0dd087a2041c99
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Wed Sep 3 12:54:13 2014 +0200
stratum: store server time offset in context
commit 1a4391d7ff21397a128abf031f92733a8ac47437
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 12:40:52 2014 +0200
hashlog: prevent double computing on jobs already done
commit 049e57730116685755bd3ff214f0793cce7c773b
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Wed Sep 3 09:49:14 2014 +0200
tmp blake log
commit 43d3e93e1a97e569ead2437f759c6b8423d30c0a
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Wed Sep 3 09:29:51 2014 +0200
blake: set a max throughput
commit 7e595a36ea69027c8a28023399540a761e7686c3
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 21:13:37 2014 +0200
blake: cleanup, remove d_hash buf, not in a chain
host: only bencode if gpu hash was found
commit de80c7e9d1448f15541d08c5dbbf372d5bfeba48
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 12:40:44 2014 +0200
blake: remove unused parameter and fix index in d_hash
that reduce the speed to 92MH/s but the next commit
give us 30 more
so, todo: merge the whole checkhash proc in gpu_hash
and remove this d_hash buffer...
commit 2d42ae6de586a6ae8cbfd01806a273fd5cc4b262
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 05:09:31 2014 +0200
stratum: handle a small cache of submitted jobs
Prevent to send duplicated shares on some pools like hashharder..
This cache keeps submitted job/nounces of the last 15 minutes
so, remove exit on repeated duplicate shares,
the submitted cache now handles this problem.
Signed-off-by: Tanguy Pruvot <tanguy.pruvot@gmail.com>
commit 1b8c3c12fa5bb83afbb02f9d5f60586939f36d86
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 03:38:57 2014 +0200
debug: a new boolean to log or not json rpc data
commit 1f99aae0ff621f4f85f119d811a3f1a8d2204f60
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 18:49:23 2014 +0200
exit on repeated duplicate shares (to enhance)
create a new function proper_exit() to do common stuff on exit...
commit 530732458add6c4c3836606d028930f3581c0a5f
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 12:22:51 2014 +0200
blake: use a constant for threads, reduce mallocated d_hash size
and clean a bit more...
commit 0aeac878ef60840f3123354037cd56a89d2e94e6
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 06:12:55 2014 +0200
blake: tune up and cleanup, ~100 MH/s on a normal 750Ti
tested on linux and windows (x86 binary)...
but there is a high number of duplicated shares... weird
commit 4a52d0553b0076b984be480725fa67689c544647
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 10:22:32 2014 +0200
debug: show json methods, hide hash/target if ok
commit 1fb9becc1f2b6a15d8ccea4d8314df9ddf0af4ed
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 08:44:19 2014 +0200
cpu-miner: sort algos by name, show reject reason
commit bfe96c49b0bf321ed0776cb1cf31c4fe8a0a8b8d
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Aug 25 11:21:06 2014 +0200
release 1.4, update README...
commit c17d11e37758c37762a7664a731fda6e9a5454b1
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Sun Aug 31 08:57:48 2014 +0200
add "blake" 256, 14 rounds (for NEOS blake, not BlakeCoin)
also remove "missing" file, its old and not compatible with ubuntu 14.04
to test on windows
blake: clean and optimize
Release v1.4 with blake (NEOS)
10 years ago
|
|
|
#define CL_LYL "\x1B[01;33m" /* tooltips */
|
|
|
|
#define CL_LBL "\x1B[01;34m" /* light blue */
|
|
|
|
#define CL_LMA "\x1B[01;35m" /* light magenta */
|
|
|
|
#define CL_LCY "\x1B[01;36m" /* light cyan */
|
|
|
|
|
|
|
|
#define CL_WHT "\x1B[01;37m" /* white */
|
|
|
|
|
|
|
|
extern void applog(int prio, const char *fmt, ...);
|
|
|
|
extern json_t *json_rpc_call(CURL *curl, const char *url, const char *userpass,
|
|
|
|
const char *rpc_req, bool, bool, int *);
|
|
|
|
extern char *bin2hex(const unsigned char *p, size_t len);
|
|
|
|
extern bool hex2bin(unsigned char *p, const char *hexstr, size_t len);
|
|
|
|
extern int timeval_subtract(struct timeval *result, struct timeval *x,
|
|
|
|
struct timeval *y);
|
|
|
|
extern bool fulltest(const uint32_t *hash, const uint32_t *target);
|
|
|
|
extern void diff_to_target(uint32_t *target, double diff);
|
|
|
|
|
|
|
|
struct stratum_job {
|
|
|
|
char *job_id;
|
|
|
|
unsigned char prevhash[32];
|
|
|
|
size_t coinbase_size;
|
|
|
|
unsigned char *coinbase;
|
|
|
|
unsigned char *xnonce2;
|
|
|
|
int merkle_count;
|
|
|
|
unsigned char **merkle;
|
|
|
|
unsigned char version[4];
|
|
|
|
unsigned char nbits[4];
|
|
|
|
unsigned char ntime[4];
|
|
|
|
bool clean;
|
|
|
|
unsigned char nreward[2];
|
|
|
|
double diff;
|
|
|
|
};
|
|
|
|
|
|
|
|
struct stratum_ctx {
|
|
|
|
char *url;
|
|
|
|
|
|
|
|
CURL *curl;
|
|
|
|
char *curl_url;
|
|
|
|
char curl_err_str[CURL_ERROR_SIZE];
|
|
|
|
curl_socket_t sock;
|
|
|
|
size_t sockbuf_size;
|
|
|
|
char *sockbuf;
|
|
|
|
pthread_mutex_t sock_lock;
|
|
|
|
|
|
|
|
double next_diff;
|
|
|
|
|
|
|
|
char *session_id;
|
|
|
|
size_t xnonce1_size;
|
|
|
|
unsigned char *xnonce1;
|
|
|
|
size_t xnonce2_size;
|
|
|
|
struct stratum_job job;
|
|
|
|
pthread_mutex_t work_lock;
|
Release v1.4 with blake (NEOS)
Blake256: squashed commit...
Squashed commit of the following:
commit c370208bc92ef16557f66e5391faf2b1ad47726f
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Wed Sep 3 13:53:01 2014 +0200
hashlog: prepare store of scanned range
commit e2cf49a5e956f03deafd266d1a0dd087a2041c99
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Wed Sep 3 12:54:13 2014 +0200
stratum: store server time offset in context
commit 1a4391d7ff21397a128abf031f92733a8ac47437
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 12:40:52 2014 +0200
hashlog: prevent double computing on jobs already done
commit 049e57730116685755bd3ff214f0793cce7c773b
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Wed Sep 3 09:49:14 2014 +0200
tmp blake log
commit 43d3e93e1a97e569ead2437f759c6b8423d30c0a
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Wed Sep 3 09:29:51 2014 +0200
blake: set a max throughput
commit 7e595a36ea69027c8a28023399540a761e7686c3
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 21:13:37 2014 +0200
blake: cleanup, remove d_hash buf, not in a chain
host: only bencode if gpu hash was found
commit de80c7e9d1448f15541d08c5dbbf372d5bfeba48
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 12:40:44 2014 +0200
blake: remove unused parameter and fix index in d_hash
that reduce the speed to 92MH/s but the next commit
give us 30 more
so, todo: merge the whole checkhash proc in gpu_hash
and remove this d_hash buffer...
commit 2d42ae6de586a6ae8cbfd01806a273fd5cc4b262
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 05:09:31 2014 +0200
stratum: handle a small cache of submitted jobs
Prevent to send duplicated shares on some pools like hashharder..
This cache keeps submitted job/nounces of the last 15 minutes
so, remove exit on repeated duplicate shares,
the submitted cache now handles this problem.
Signed-off-by: Tanguy Pruvot <tanguy.pruvot@gmail.com>
commit 1b8c3c12fa5bb83afbb02f9d5f60586939f36d86
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 03:38:57 2014 +0200
debug: a new boolean to log or not json rpc data
commit 1f99aae0ff621f4f85f119d811a3f1a8d2204f60
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 18:49:23 2014 +0200
exit on repeated duplicate shares (to enhance)
create a new function proper_exit() to do common stuff on exit...
commit 530732458add6c4c3836606d028930f3581c0a5f
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 12:22:51 2014 +0200
blake: use a constant for threads, reduce mallocated d_hash size
and clean a bit more...
commit 0aeac878ef60840f3123354037cd56a89d2e94e6
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 06:12:55 2014 +0200
blake: tune up and cleanup, ~100 MH/s on a normal 750Ti
tested on linux and windows (x86 binary)...
but there is a high number of duplicated shares... weird
commit 4a52d0553b0076b984be480725fa67689c544647
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 10:22:32 2014 +0200
debug: show json methods, hide hash/target if ok
commit 1fb9becc1f2b6a15d8ccea4d8314df9ddf0af4ed
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 08:44:19 2014 +0200
cpu-miner: sort algos by name, show reject reason
commit bfe96c49b0bf321ed0776cb1cf31c4fe8a0a8b8d
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Aug 25 11:21:06 2014 +0200
release 1.4, update README...
commit c17d11e37758c37762a7664a731fda6e9a5454b1
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Sun Aug 31 08:57:48 2014 +0200
add "blake" 256, 14 rounds (for NEOS blake, not BlakeCoin)
also remove "missing" file, its old and not compatible with ubuntu 14.04
to test on windows
blake: clean and optimize
Release v1.4 with blake (NEOS)
10 years ago
|
|
|
|
|
|
|
int srvtime_diff;
|
|
|
|
int bloc_height;
|
|
|
|
};
|
|
|
|
|
|
|
|
bool stratum_socket_full(struct stratum_ctx *sctx, int timeout);
|
|
|
|
bool stratum_send_line(struct stratum_ctx *sctx, char *s);
|
|
|
|
char *stratum_recv_line(struct stratum_ctx *sctx);
|
|
|
|
bool stratum_connect(struct stratum_ctx *sctx, const char *url);
|
|
|
|
void stratum_disconnect(struct stratum_ctx *sctx);
|
|
|
|
bool stratum_subscribe(struct stratum_ctx *sctx);
|
|
|
|
bool stratum_authorize(struct stratum_ctx *sctx, const char *user, const char *pass);
|
|
|
|
bool stratum_handle_method(struct stratum_ctx *sctx, const char *s);
|
|
|
|
|
Release v1.4 with blake (NEOS)
Blake256: squashed commit...
Squashed commit of the following:
commit c370208bc92ef16557f66e5391faf2b1ad47726f
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Wed Sep 3 13:53:01 2014 +0200
hashlog: prepare store of scanned range
commit e2cf49a5e956f03deafd266d1a0dd087a2041c99
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Wed Sep 3 12:54:13 2014 +0200
stratum: store server time offset in context
commit 1a4391d7ff21397a128abf031f92733a8ac47437
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 12:40:52 2014 +0200
hashlog: prevent double computing on jobs already done
commit 049e57730116685755bd3ff214f0793cce7c773b
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Wed Sep 3 09:49:14 2014 +0200
tmp blake log
commit 43d3e93e1a97e569ead2437f759c6b8423d30c0a
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Wed Sep 3 09:29:51 2014 +0200
blake: set a max throughput
commit 7e595a36ea69027c8a28023399540a761e7686c3
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 21:13:37 2014 +0200
blake: cleanup, remove d_hash buf, not in a chain
host: only bencode if gpu hash was found
commit de80c7e9d1448f15541d08c5dbbf372d5bfeba48
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 12:40:44 2014 +0200
blake: remove unused parameter and fix index in d_hash
that reduce the speed to 92MH/s but the next commit
give us 30 more
so, todo: merge the whole checkhash proc in gpu_hash
and remove this d_hash buffer...
commit 2d42ae6de586a6ae8cbfd01806a273fd5cc4b262
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 05:09:31 2014 +0200
stratum: handle a small cache of submitted jobs
Prevent to send duplicated shares on some pools like hashharder..
This cache keeps submitted job/nounces of the last 15 minutes
so, remove exit on repeated duplicate shares,
the submitted cache now handles this problem.
Signed-off-by: Tanguy Pruvot <tanguy.pruvot@gmail.com>
commit 1b8c3c12fa5bb83afbb02f9d5f60586939f36d86
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 03:38:57 2014 +0200
debug: a new boolean to log or not json rpc data
commit 1f99aae0ff621f4f85f119d811a3f1a8d2204f60
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 18:49:23 2014 +0200
exit on repeated duplicate shares (to enhance)
create a new function proper_exit() to do common stuff on exit...
commit 530732458add6c4c3836606d028930f3581c0a5f
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 12:22:51 2014 +0200
blake: use a constant for threads, reduce mallocated d_hash size
and clean a bit more...
commit 0aeac878ef60840f3123354037cd56a89d2e94e6
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 06:12:55 2014 +0200
blake: tune up and cleanup, ~100 MH/s on a normal 750Ti
tested on linux and windows (x86 binary)...
but there is a high number of duplicated shares... weird
commit 4a52d0553b0076b984be480725fa67689c544647
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 10:22:32 2014 +0200
debug: show json methods, hide hash/target if ok
commit 1fb9becc1f2b6a15d8ccea4d8314df9ddf0af4ed
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 08:44:19 2014 +0200
cpu-miner: sort algos by name, show reject reason
commit bfe96c49b0bf321ed0776cb1cf31c4fe8a0a8b8d
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Aug 25 11:21:06 2014 +0200
release 1.4, update README...
commit c17d11e37758c37762a7664a731fda6e9a5454b1
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Sun Aug 31 08:57:48 2014 +0200
add "blake" 256, 14 rounds (for NEOS blake, not BlakeCoin)
also remove "missing" file, its old and not compatible with ubuntu 14.04
to test on windows
blake: clean and optimize
Release v1.4 with blake (NEOS)
10 years ago
|
|
|
void hashlog_remember_submit(char* jobid, uint32_t nounce);
|
|
|
|
void hashlog_remember_scan_range(char* jobid, uint32_t scanned_from, uint32_t scanned_to);
|
|
|
|
uint32_t hashlog_already_submittted(char* jobid, uint32_t nounce);
|
|
|
|
uint32_t hashlog_get_last_sent(char* jobid);
|
|
|
|
uint64_t hashlog_get_scan_range(char* jobid);
|
|
|
|
void hashlog_purge_old(void);
|
|
|
|
void hashlog_purge_job(char* jobid);
|
|
|
|
void hashlog_purge_all(void);
|
|
|
|
void hashlog_dump_job(char* jobid);
|
|
|
|
|
|
|
|
struct thread_q;
|
|
|
|
|
|
|
|
extern struct thread_q *tq_new(void);
|
|
|
|
extern void tq_free(struct thread_q *tq);
|
|
|
|
extern bool tq_push(struct thread_q *tq, void *data);
|
|
|
|
extern void *tq_pop(struct thread_q *tq, const struct timespec *abstime);
|
|
|
|
extern void tq_freeze(struct thread_q *tq);
|
|
|
|
extern void tq_thaw(struct thread_q *tq);
|
|
|
|
|
Release v1.4 with blake (NEOS)
Blake256: squashed commit...
Squashed commit of the following:
commit c370208bc92ef16557f66e5391faf2b1ad47726f
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Wed Sep 3 13:53:01 2014 +0200
hashlog: prepare store of scanned range
commit e2cf49a5e956f03deafd266d1a0dd087a2041c99
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Wed Sep 3 12:54:13 2014 +0200
stratum: store server time offset in context
commit 1a4391d7ff21397a128abf031f92733a8ac47437
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 12:40:52 2014 +0200
hashlog: prevent double computing on jobs already done
commit 049e57730116685755bd3ff214f0793cce7c773b
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Wed Sep 3 09:49:14 2014 +0200
tmp blake log
commit 43d3e93e1a97e569ead2437f759c6b8423d30c0a
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Wed Sep 3 09:29:51 2014 +0200
blake: set a max throughput
commit 7e595a36ea69027c8a28023399540a761e7686c3
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 21:13:37 2014 +0200
blake: cleanup, remove d_hash buf, not in a chain
host: only bencode if gpu hash was found
commit de80c7e9d1448f15541d08c5dbbf372d5bfeba48
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 12:40:44 2014 +0200
blake: remove unused parameter and fix index in d_hash
that reduce the speed to 92MH/s but the next commit
give us 30 more
so, todo: merge the whole checkhash proc in gpu_hash
and remove this d_hash buffer...
commit 2d42ae6de586a6ae8cbfd01806a273fd5cc4b262
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 05:09:31 2014 +0200
stratum: handle a small cache of submitted jobs
Prevent to send duplicated shares on some pools like hashharder..
This cache keeps submitted job/nounces of the last 15 minutes
so, remove exit on repeated duplicate shares,
the submitted cache now handles this problem.
Signed-off-by: Tanguy Pruvot <tanguy.pruvot@gmail.com>
commit 1b8c3c12fa5bb83afbb02f9d5f60586939f36d86
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 03:38:57 2014 +0200
debug: a new boolean to log or not json rpc data
commit 1f99aae0ff621f4f85f119d811a3f1a8d2204f60
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 18:49:23 2014 +0200
exit on repeated duplicate shares (to enhance)
create a new function proper_exit() to do common stuff on exit...
commit 530732458add6c4c3836606d028930f3581c0a5f
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 12:22:51 2014 +0200
blake: use a constant for threads, reduce mallocated d_hash size
and clean a bit more...
commit 0aeac878ef60840f3123354037cd56a89d2e94e6
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 06:12:55 2014 +0200
blake: tune up and cleanup, ~100 MH/s on a normal 750Ti
tested on linux and windows (x86 binary)...
but there is a high number of duplicated shares... weird
commit 4a52d0553b0076b984be480725fa67689c544647
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 10:22:32 2014 +0200
debug: show json methods, hide hash/target if ok
commit 1fb9becc1f2b6a15d8ccea4d8314df9ddf0af4ed
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 08:44:19 2014 +0200
cpu-miner: sort algos by name, show reject reason
commit bfe96c49b0bf321ed0776cb1cf31c4fe8a0a8b8d
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Aug 25 11:21:06 2014 +0200
release 1.4, update README...
commit c17d11e37758c37762a7664a731fda6e9a5454b1
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Sun Aug 31 08:57:48 2014 +0200
add "blake" 256, 14 rounds (for NEOS blake, not BlakeCoin)
also remove "missing" file, its old and not compatible with ubuntu 14.04
to test on windows
blake: clean and optimize
Release v1.4 with blake (NEOS)
10 years ago
|
|
|
void proper_exit(int reason);
|
|
|
|
|
|
|
|
size_t time2str(char* buf, time_t timer);
|
|
|
|
char* atime2str(time_t timer);
|
|
|
|
|
|
|
|
void applog_hash(unsigned char *hash);
|
|
|
|
|
|
|
|
void print_hash_tests(void);
|
Release v1.4 with blake (NEOS)
Blake256: squashed commit...
Squashed commit of the following:
commit c370208bc92ef16557f66e5391faf2b1ad47726f
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Wed Sep 3 13:53:01 2014 +0200
hashlog: prepare store of scanned range
commit e2cf49a5e956f03deafd266d1a0dd087a2041c99
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Wed Sep 3 12:54:13 2014 +0200
stratum: store server time offset in context
commit 1a4391d7ff21397a128abf031f92733a8ac47437
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 12:40:52 2014 +0200
hashlog: prevent double computing on jobs already done
commit 049e57730116685755bd3ff214f0793cce7c773b
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Wed Sep 3 09:49:14 2014 +0200
tmp blake log
commit 43d3e93e1a97e569ead2437f759c6b8423d30c0a
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Wed Sep 3 09:29:51 2014 +0200
blake: set a max throughput
commit 7e595a36ea69027c8a28023399540a761e7686c3
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 21:13:37 2014 +0200
blake: cleanup, remove d_hash buf, not in a chain
host: only bencode if gpu hash was found
commit de80c7e9d1448f15541d08c5dbbf372d5bfeba48
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 12:40:44 2014 +0200
blake: remove unused parameter and fix index in d_hash
that reduce the speed to 92MH/s but the next commit
give us 30 more
so, todo: merge the whole checkhash proc in gpu_hash
and remove this d_hash buffer...
commit 2d42ae6de586a6ae8cbfd01806a273fd5cc4b262
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 05:09:31 2014 +0200
stratum: handle a small cache of submitted jobs
Prevent to send duplicated shares on some pools like hashharder..
This cache keeps submitted job/nounces of the last 15 minutes
so, remove exit on repeated duplicate shares,
the submitted cache now handles this problem.
Signed-off-by: Tanguy Pruvot <tanguy.pruvot@gmail.com>
commit 1b8c3c12fa5bb83afbb02f9d5f60586939f36d86
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 03:38:57 2014 +0200
debug: a new boolean to log or not json rpc data
commit 1f99aae0ff621f4f85f119d811a3f1a8d2204f60
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 18:49:23 2014 +0200
exit on repeated duplicate shares (to enhance)
create a new function proper_exit() to do common stuff on exit...
commit 530732458add6c4c3836606d028930f3581c0a5f
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 12:22:51 2014 +0200
blake: use a constant for threads, reduce mallocated d_hash size
and clean a bit more...
commit 0aeac878ef60840f3123354037cd56a89d2e94e6
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 06:12:55 2014 +0200
blake: tune up and cleanup, ~100 MH/s on a normal 750Ti
tested on linux and windows (x86 binary)...
but there is a high number of duplicated shares... weird
commit 4a52d0553b0076b984be480725fa67689c544647
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 10:22:32 2014 +0200
debug: show json methods, hide hash/target if ok
commit 1fb9becc1f2b6a15d8ccea4d8314df9ddf0af4ed
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 08:44:19 2014 +0200
cpu-miner: sort algos by name, show reject reason
commit bfe96c49b0bf321ed0776cb1cf31c4fe8a0a8b8d
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Aug 25 11:21:06 2014 +0200
release 1.4, update README...
commit c17d11e37758c37762a7664a731fda6e9a5454b1
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Sun Aug 31 08:57:48 2014 +0200
add "blake" 256, 14 rounds (for NEOS blake, not BlakeCoin)
also remove "missing" file, its old and not compatible with ubuntu 14.04
to test on windows
blake: clean and optimize
Release v1.4 with blake (NEOS)
10 years ago
|
|
|
void animehash(void *state, const void *input);
|
|
|
|
void blake32hash(void *output, const void *input);
|
|
|
|
void fresh_hash(void *state, const void *input);
|
|
|
|
void fugue256_hash(unsigned char* output, const unsigned char* input, int len);
|
|
|
|
void heavycoin_hash(unsigned char* output, const unsigned char* input, int len);
|
Release v1.4 with blake (NEOS)
Blake256: squashed commit...
Squashed commit of the following:
commit c370208bc92ef16557f66e5391faf2b1ad47726f
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Wed Sep 3 13:53:01 2014 +0200
hashlog: prepare store of scanned range
commit e2cf49a5e956f03deafd266d1a0dd087a2041c99
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Wed Sep 3 12:54:13 2014 +0200
stratum: store server time offset in context
commit 1a4391d7ff21397a128abf031f92733a8ac47437
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 12:40:52 2014 +0200
hashlog: prevent double computing on jobs already done
commit 049e57730116685755bd3ff214f0793cce7c773b
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Wed Sep 3 09:49:14 2014 +0200
tmp blake log
commit 43d3e93e1a97e569ead2437f759c6b8423d30c0a
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Wed Sep 3 09:29:51 2014 +0200
blake: set a max throughput
commit 7e595a36ea69027c8a28023399540a761e7686c3
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 21:13:37 2014 +0200
blake: cleanup, remove d_hash buf, not in a chain
host: only bencode if gpu hash was found
commit de80c7e9d1448f15541d08c5dbbf372d5bfeba48
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 12:40:44 2014 +0200
blake: remove unused parameter and fix index in d_hash
that reduce the speed to 92MH/s but the next commit
give us 30 more
so, todo: merge the whole checkhash proc in gpu_hash
and remove this d_hash buffer...
commit 2d42ae6de586a6ae8cbfd01806a273fd5cc4b262
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 05:09:31 2014 +0200
stratum: handle a small cache of submitted jobs
Prevent to send duplicated shares on some pools like hashharder..
This cache keeps submitted job/nounces of the last 15 minutes
so, remove exit on repeated duplicate shares,
the submitted cache now handles this problem.
Signed-off-by: Tanguy Pruvot <tanguy.pruvot@gmail.com>
commit 1b8c3c12fa5bb83afbb02f9d5f60586939f36d86
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 03:38:57 2014 +0200
debug: a new boolean to log or not json rpc data
commit 1f99aae0ff621f4f85f119d811a3f1a8d2204f60
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 18:49:23 2014 +0200
exit on repeated duplicate shares (to enhance)
create a new function proper_exit() to do common stuff on exit...
commit 530732458add6c4c3836606d028930f3581c0a5f
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 12:22:51 2014 +0200
blake: use a constant for threads, reduce mallocated d_hash size
and clean a bit more...
commit 0aeac878ef60840f3123354037cd56a89d2e94e6
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 06:12:55 2014 +0200
blake: tune up and cleanup, ~100 MH/s on a normal 750Ti
tested on linux and windows (x86 binary)...
but there is a high number of duplicated shares... weird
commit 4a52d0553b0076b984be480725fa67689c544647
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 10:22:32 2014 +0200
debug: show json methods, hide hash/target if ok
commit 1fb9becc1f2b6a15d8ccea4d8314df9ddf0af4ed
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 08:44:19 2014 +0200
cpu-miner: sort algos by name, show reject reason
commit bfe96c49b0bf321ed0776cb1cf31c4fe8a0a8b8d
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Aug 25 11:21:06 2014 +0200
release 1.4, update README...
commit c17d11e37758c37762a7664a731fda6e9a5454b1
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Sun Aug 31 08:57:48 2014 +0200
add "blake" 256, 14 rounds (for NEOS blake, not BlakeCoin)
also remove "missing" file, its old and not compatible with ubuntu 14.04
to test on windows
blake: clean and optimize
Release v1.4 with blake (NEOS)
10 years ago
|
|
|
unsigned int jackpothash(void *state, const void *input);
|
|
|
|
void groestlhash(void *state, const void *input);
|
|
|
|
void myriadhash(void *state, const void *input);
|
|
|
|
void nist5hash(void *state, const void *input);
|
|
|
|
void quarkhash(void *state, const void *input);
|
|
|
|
void wcoinhash(void *state, const void *input);
|
|
|
|
void x11hash(void *output, const void *input);
|
|
|
|
void x13hash(void *output, const void *input);
|
|
|
|
void x14hash(void *output, const void *input);
|
|
|
|
void x15hash(void *output, const void *input);
|
|
|
|
void x17hash(void *output, const void *input);
|
|
|
|
|
|
|
|
#ifdef __cplusplus
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#endif /* __MINER_H__ */
|