|
|
|
|
extern "C"
|
|
|
|
|
{
|
|
|
|
|
#include "sph/sph_blake.h"
|
|
|
|
|
#include "sph/sph_bmw.h"
|
|
|
|
|
#include "sph/sph_groestl.h"
|
|
|
|
|
#include "sph/sph_skein.h"
|
|
|
|
|
#include "sph/sph_jh.h"
|
|
|
|
|
#include "sph/sph_keccak.h"
|
|
|
|
|
#include "miner.h"
|
|
|
|
|
#include "cuda_helper.h"
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
// aus cpu-miner.c
|
|
|
|
|
extern int device_map[8];
|
|
|
|
|
|
|
|
|
|
// Speicher f<EFBFBD>r Input/Output der verketteten Hashfunktionen
|
|
|
|
|
static uint32_t *d_hash[8];
|
|
|
|
|
|
|
|
|
|
// Speicher zur Generierung der Noncevektoren f<EFBFBD>r die bedingten Hashes
|
|
|
|
|
static uint32_t *d_animeNonces[8];
|
|
|
|
|
static uint32_t *d_branch1Nonces[8];
|
|
|
|
|
static uint32_t *d_branch2Nonces[8];
|
|
|
|
|
static uint32_t *d_branch3Nonces[8];
|
|
|
|
|
|
|
|
|
|
extern void quark_blake512_cpu_init(int thr_id, int threads);
|
|
|
|
|
extern void quark_blake512_cpu_hash_64(int thr_id, int threads, uint32_t startNounce, uint32_t *d_nonceVector, uint32_t *d_hash, int order);
|
|
|
|
|
|
|
|
|
|
extern void quark_bmw512_cpu_init(int thr_id, int threads);
|
|
|
|
|
extern void quark_bmw512_cpu_setBlock_80(void *pdata);
|
|
|
|
|
extern void quark_bmw512_cpu_hash_80(int thr_id, int threads, uint32_t startNounce, uint32_t *d_outputHash, int order);
|
|
|
|
|
extern void quark_bmw512_cpu_hash_64(int thr_id, int threads, uint32_t startNounce, uint32_t *d_nonceVector, uint32_t *d_outputHash, int order);
|
|
|
|
|
|
|
|
|
|
extern void quark_groestl512_cpu_init(int thr_id, int threads);
|
|
|
|
|
extern void quark_groestl512_cpu_hash_64(int thr_id, int threads, uint32_t startNounce, uint32_t *d_nonceVector, uint32_t *d_hash, int order);
|
|
|
|
|
extern void quark_doublegroestl512_cpu_hash_64(int thr_id, int threads, uint32_t startNounce, uint32_t *d_nonceVector, uint32_t *d_hash, int order);
|
|
|
|
|
|
|
|
|
|
extern void quark_skein512_cpu_init(int thr_id, int threads);
|
|
|
|
|
extern void quark_skein512_cpu_hash_64(int thr_id, int threads, uint32_t startNounce, uint32_t *d_nonceVector, uint32_t *d_hash, int order);
|
|
|
|
|
|
|
|
|
|
extern void quark_keccak512_cpu_init(int thr_id, int threads);
|
|
|
|
|
extern void quark_keccak512_cpu_hash_64(int thr_id, int threads, uint32_t startNounce, uint32_t *d_nonceVector, uint32_t *d_hash, int order);
|
|
|
|
|
|
|
|
|
|
extern void quark_jh512_cpu_init(int thr_id, int threads);
|
|
|
|
|
extern void quark_jh512_cpu_hash_64(int thr_id, int threads, uint32_t startNounce, uint32_t *d_nonceVector, uint32_t *d_hash, int order);
|
|
|
|
|
|
|
|
|
|
extern void cuda_check_cpu_init(int thr_id, int threads);
|
|
|
|
|
extern void cuda_check_cpu_setTarget(const void *ptarget);
|
|
|
|
|
extern uint32_t cuda_check_cpu_hash_64(int thr_id, int threads, uint32_t startNounce, uint32_t *d_nonceVector, uint32_t *d_inputHash, int order);
|
|
|
|
|
|
|
|
|
|
extern void quark_compactTest_cpu_init(int thr_id, int threads);
|
|
|
|
|
extern void quark_compactTest_cpu_hash_64(int thr_id, int threads, uint32_t startNounce, uint32_t *inpHashes, uint32_t *d_validNonceTable,
|
|
|
|
|
uint32_t *d_nonces1, size_t *nrm1,
|
|
|
|
|
uint32_t *d_nonces2, size_t *nrm2,
|
|
|
|
|
int order);
|
|
|
|
|
extern void quark_compactTest_single_false_cpu_hash_64(int thr_id, int threads, uint32_t startNounce, uint32_t *inpHashes, uint32_t *d_validNonceTable,
|
|
|
|
|
uint32_t *d_nonces1, size_t *nrm1,
|
|
|
|
|
int order);
|
|
|
|
|
|
|
|
|
|
// Original Quarkhash Funktion aus einem miner Quelltext
|
Release v1.4 with blake (NEOS)
Blake256: squashed commit...
Squashed commit of the following:
commit c370208bc92ef16557f66e5391faf2b1ad47726f
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Wed Sep 3 13:53:01 2014 +0200
hashlog: prepare store of scanned range
commit e2cf49a5e956f03deafd266d1a0dd087a2041c99
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Wed Sep 3 12:54:13 2014 +0200
stratum: store server time offset in context
commit 1a4391d7ff21397a128abf031f92733a8ac47437
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 12:40:52 2014 +0200
hashlog: prevent double computing on jobs already done
commit 049e57730116685755bd3ff214f0793cce7c773b
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Wed Sep 3 09:49:14 2014 +0200
tmp blake log
commit 43d3e93e1a97e569ead2437f759c6b8423d30c0a
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Wed Sep 3 09:29:51 2014 +0200
blake: set a max throughput
commit 7e595a36ea69027c8a28023399540a761e7686c3
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 21:13:37 2014 +0200
blake: cleanup, remove d_hash buf, not in a chain
host: only bencode if gpu hash was found
commit de80c7e9d1448f15541d08c5dbbf372d5bfeba48
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 12:40:44 2014 +0200
blake: remove unused parameter and fix index in d_hash
that reduce the speed to 92MH/s but the next commit
give us 30 more
so, todo: merge the whole checkhash proc in gpu_hash
and remove this d_hash buffer...
commit 2d42ae6de586a6ae8cbfd01806a273fd5cc4b262
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 05:09:31 2014 +0200
stratum: handle a small cache of submitted jobs
Prevent to send duplicated shares on some pools like hashharder..
This cache keeps submitted job/nounces of the last 15 minutes
so, remove exit on repeated duplicate shares,
the submitted cache now handles this problem.
Signed-off-by: Tanguy Pruvot <tanguy.pruvot@gmail.com>
commit 1b8c3c12fa5bb83afbb02f9d5f60586939f36d86
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Tue Sep 2 03:38:57 2014 +0200
debug: a new boolean to log or not json rpc data
commit 1f99aae0ff621f4f85f119d811a3f1a8d2204f60
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 18:49:23 2014 +0200
exit on repeated duplicate shares (to enhance)
create a new function proper_exit() to do common stuff on exit...
commit 530732458add6c4c3836606d028930f3581c0a5f
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 12:22:51 2014 +0200
blake: use a constant for threads, reduce mallocated d_hash size
and clean a bit more...
commit 0aeac878ef60840f3123354037cd56a89d2e94e6
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 06:12:55 2014 +0200
blake: tune up and cleanup, ~100 MH/s on a normal 750Ti
tested on linux and windows (x86 binary)...
but there is a high number of duplicated shares... weird
commit 4a52d0553b0076b984be480725fa67689c544647
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 10:22:32 2014 +0200
debug: show json methods, hide hash/target if ok
commit 1fb9becc1f2b6a15d8ccea4d8314df9ddf0af4ed
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Sep 1 08:44:19 2014 +0200
cpu-miner: sort algos by name, show reject reason
commit bfe96c49b0bf321ed0776cb1cf31c4fe8a0a8b8d
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Mon Aug 25 11:21:06 2014 +0200
release 1.4, update README...
commit c17d11e37758c37762a7664a731fda6e9a5454b1
Author: Tanguy Pruvot <tanguy.pruvot@gmail.com>
Date: Sun Aug 31 08:57:48 2014 +0200
add "blake" 256, 14 rounds (for NEOS blake, not BlakeCoin)
also remove "missing" file, its old and not compatible with ubuntu 14.04
to test on windows
blake: clean and optimize
Release v1.4 with blake (NEOS)
10 years ago
|
|
|
|
extern "C" void animehash(void *state, const void *input)
|
|
|
|
|
{
|
|
|
|
|
sph_blake512_context ctx_blake;
|
|
|
|
|
sph_bmw512_context ctx_bmw;
|
|
|
|
|
sph_groestl512_context ctx_groestl;
|
|
|
|
|
sph_jh512_context ctx_jh;
|
|
|
|
|
sph_keccak512_context ctx_keccak;
|
|
|
|
|
sph_skein512_context ctx_skein;
|
|
|
|
|
|
|
|
|
|
unsigned char hash[64];
|
|
|
|
|
|
|
|
|
|
sph_bmw512_init(&ctx_bmw);
|
|
|
|
|
// ZBMW;
|
|
|
|
|
sph_bmw512 (&ctx_bmw, (const void*) input, 80);
|
|
|
|
|
sph_bmw512_close(&ctx_bmw, (void*) hash);
|
|
|
|
|
|
|
|
|
|
sph_blake512_init(&ctx_blake);
|
|
|
|
|
// ZBLAKE;
|
|
|
|
|
sph_blake512 (&ctx_blake, hash, 64);
|
|
|
|
|
sph_blake512_close(&ctx_blake, (void*) hash);
|
|
|
|
|
|
|
|
|
|
if (hash[0] & 0x8)
|
|
|
|
|
{
|
|
|
|
|
sph_groestl512_init(&ctx_groestl);
|
|
|
|
|
// ZGROESTL;
|
|
|
|
|
sph_groestl512 (&ctx_groestl, (const void*) hash, 64);
|
|
|
|
|
sph_groestl512_close(&ctx_groestl, (void*) hash);
|
|
|
|
|
}
|
|
|
|
|
else
|
|
|
|
|
{
|
|
|
|
|
sph_skein512_init(&ctx_skein);
|
|
|
|
|
// ZSKEIN;
|
|
|
|
|
sph_skein512 (&ctx_skein, (const void*) hash, 64);
|
|
|
|
|
sph_skein512_close(&ctx_skein, (void*) hash);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
sph_groestl512_init(&ctx_groestl);
|
|
|
|
|
// ZGROESTL;
|
|
|
|
|
sph_groestl512 (&ctx_groestl, (const void*) hash, 64);
|
|
|
|
|
sph_groestl512_close(&ctx_groestl, (void*) hash);
|
|
|
|
|
|
|
|
|
|
sph_jh512_init(&ctx_jh);
|
|
|
|
|
// ZJH;
|
|
|
|
|
sph_jh512 (&ctx_jh, (const void*) hash, 64);
|
|
|
|
|
sph_jh512_close(&ctx_jh, (void*) hash);
|
|
|
|
|
|
|
|
|
|
if (hash[0] & 0x8)
|
|
|
|
|
{
|
|
|
|
|
sph_blake512_init(&ctx_blake);
|
|
|
|
|
// ZBLAKE;
|
|
|
|
|
sph_blake512 (&ctx_blake, (const void*) hash, 64);
|
|
|
|
|
sph_blake512_close(&ctx_blake, (void*) hash);
|
|
|
|
|
}
|
|
|
|
|
else
|
|
|
|
|
{
|
|
|
|
|
sph_bmw512_init(&ctx_bmw);
|
|
|
|
|
// ZBMW;
|
|
|
|
|
sph_bmw512 (&ctx_bmw, (const void*) hash, 64);
|
|
|
|
|
sph_bmw512_close(&ctx_bmw, (void*) hash);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
sph_keccak512_init(&ctx_keccak);
|
|
|
|
|
// ZKECCAK;
|
|
|
|
|
sph_keccak512 (&ctx_keccak, (const void*) hash, 64);
|
|
|
|
|
sph_keccak512_close(&ctx_keccak, (void*) hash);
|
|
|
|
|
|
|
|
|
|
sph_skein512_init(&ctx_skein);
|
|
|
|
|
// SKEIN;
|
|
|
|
|
sph_skein512 (&ctx_skein, (const void*) hash, 64);
|
|
|
|
|
sph_skein512_close(&ctx_skein, (void*) hash);
|
|
|
|
|
|
|
|
|
|
if (hash[0] & 0x8)
|
|
|
|
|
{
|
|
|
|
|
sph_keccak512_init(&ctx_keccak);
|
|
|
|
|
// ZKECCAK;
|
|
|
|
|
sph_keccak512 (&ctx_keccak, (const void*) hash, 64);
|
|
|
|
|
sph_keccak512_close(&ctx_keccak, (void*) hash);
|
|
|
|
|
}
|
|
|
|
|
else
|
|
|
|
|
{
|
|
|
|
|
sph_jh512_init(&ctx_jh);
|
|
|
|
|
// ZJH;
|
|
|
|
|
sph_jh512 (&ctx_jh, (const void*) hash, 64);
|
|
|
|
|
sph_jh512_close(&ctx_jh, (void*) hash);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
memcpy(state, hash, 32);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
struct HashPredicate
|
|
|
|
|
{
|
|
|
|
|
HashPredicate(uint32_t *hashes, uint32_t startNonce) :
|
|
|
|
|
m_hashes(hashes),
|
|
|
|
|
m_startNonce(startNonce)
|
|
|
|
|
{ }
|
|
|
|
|
|
|
|
|
|
__device__
|
|
|
|
|
bool operator()(const uint32_t x)
|
|
|
|
|
{
|
|
|
|
|
uint32_t *hash = &m_hashes[(x - m_startNonce)*16];
|
|
|
|
|
return hash[0] & 0x8;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
uint32_t *m_hashes;
|
|
|
|
|
uint32_t m_startNonce;
|
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
extern bool opt_benchmark;
|
|
|
|
|
|
|
|
|
|
extern "C" int scanhash_anime(int thr_id, uint32_t *pdata,
|
|
|
|
|
const uint32_t *ptarget, uint32_t max_nonce,
|
|
|
|
|
unsigned long *hashes_done)
|
|
|
|
|
{
|
|
|
|
|
const uint32_t first_nonce = pdata[19];
|
|
|
|
|
|
|
|
|
|
if (opt_benchmark)
|
|
|
|
|
((uint32_t*)ptarget)[7] = 0x00000f;
|
|
|
|
|
|
|
|
|
|
const uint32_t Htarg = ptarget[7];
|
|
|
|
|
|
|
|
|
|
const int throughput = 256*2048; // 100;
|
|
|
|
|
|
|
|
|
|
static bool init[8] = {0,0,0,0,0,0,0,0};
|
|
|
|
|
if (!init[thr_id])
|
|
|
|
|
{
|
|
|
|
|
cudaSetDevice(device_map[thr_id]);
|
|
|
|
|
|
|
|
|
|
// Konstanten kopieren, Speicher belegen
|
|
|
|
|
cudaMalloc(&d_hash[thr_id], 16 * sizeof(uint32_t) * throughput);
|
|
|
|
|
|
|
|
|
|
quark_blake512_cpu_init(thr_id, throughput);
|
|
|
|
|
quark_groestl512_cpu_init(thr_id, throughput);
|
|
|
|
|
quark_skein512_cpu_init(thr_id, throughput);
|
|
|
|
|
quark_bmw512_cpu_init(thr_id, throughput);
|
|
|
|
|
quark_keccak512_cpu_init(thr_id, throughput);
|
|
|
|
|
quark_jh512_cpu_init(thr_id, throughput);
|
|
|
|
|
cuda_check_cpu_init(thr_id, throughput);
|
|
|
|
|
quark_compactTest_cpu_init(thr_id, throughput);
|
|
|
|
|
|
|
|
|
|
cudaMalloc(&d_animeNonces[thr_id], sizeof(uint32_t)*throughput);
|
|
|
|
|
cudaMalloc(&d_branch1Nonces[thr_id], sizeof(uint32_t)*throughput);
|
|
|
|
|
cudaMalloc(&d_branch2Nonces[thr_id], sizeof(uint32_t)*throughput);
|
|
|
|
|
cudaMalloc(&d_branch3Nonces[thr_id], sizeof(uint32_t)*throughput);
|
|
|
|
|
|
|
|
|
|
init[thr_id] = true;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
uint32_t endiandata[20];
|
|
|
|
|
for (int k=0; k < 20; k++)
|
|
|
|
|
be32enc(&endiandata[k], ((uint32_t*)pdata)[k]);
|
|
|
|
|
|
|
|
|
|
quark_bmw512_cpu_setBlock_80((void*)endiandata);
|
|
|
|
|
cuda_check_cpu_setTarget(ptarget);
|
|
|
|
|
|
|
|
|
|
do {
|
|
|
|
|
int order = 0;
|
|
|
|
|
size_t nrm1=0, nrm2=0, nrm3=0;
|
|
|
|
|
|
|
|
|
|
// erstes BMW512 Hash mit CUDA
|
|
|
|
|
quark_bmw512_cpu_hash_80(thr_id, throughput, pdata[19], d_hash[thr_id], order++);
|
|
|
|
|
|
|
|
|
|
// das ist der unbedingte Branch f<EFBFBD>r Blake512
|
|
|
|
|
quark_blake512_cpu_hash_64(thr_id, throughput, pdata[19], NULL, d_hash[thr_id], order++);
|
|
|
|
|
|
|
|
|
|
quark_compactTest_single_false_cpu_hash_64(thr_id, throughput, pdata[19], d_hash[thr_id], NULL,
|
|
|
|
|
d_branch3Nonces[thr_id], &nrm3,
|
|
|
|
|
order++);
|
|
|
|
|
|
|
|
|
|
// nur den Skein Branch weiterverfolgen
|
|
|
|
|
quark_skein512_cpu_hash_64(thr_id, nrm3, pdata[19], d_branch3Nonces[thr_id], d_hash[thr_id], order++);
|
|
|
|
|
|
|
|
|
|
// das ist der unbedingte Branch f<EFBFBD>r Groestl512
|
|
|
|
|
quark_groestl512_cpu_hash_64(thr_id, nrm3, pdata[19], d_branch3Nonces[thr_id], d_hash[thr_id], order++);
|
|
|
|
|
|
|
|
|
|
// das ist der unbedingte Branch f<EFBFBD>r JH512
|
|
|
|
|
quark_jh512_cpu_hash_64(thr_id, nrm3, pdata[19], d_branch3Nonces[thr_id], d_hash[thr_id], order++);
|
|
|
|
|
|
|
|
|
|
// quarkNonces in branch1 und branch2 aufsplitten gem<EFBFBD>ss if (hash[0] & 0x8)
|
|
|
|
|
quark_compactTest_cpu_hash_64(thr_id, nrm3, pdata[19], d_hash[thr_id], d_branch3Nonces[thr_id],
|
|
|
|
|
d_branch1Nonces[thr_id], &nrm1,
|
|
|
|
|
d_branch2Nonces[thr_id], &nrm2,
|
|
|
|
|
order++);
|
|
|
|
|
|
|
|
|
|
// das ist der bedingte Branch f<EFBFBD>r Blake512
|
|
|
|
|
quark_blake512_cpu_hash_64(thr_id, nrm1, pdata[19], d_branch1Nonces[thr_id], d_hash[thr_id], order++);
|
|
|
|
|
|
|
|
|
|
// das ist der bedingte Branch f<EFBFBD>r Bmw512
|
|
|
|
|
quark_bmw512_cpu_hash_64(thr_id, nrm2, pdata[19], d_branch2Nonces[thr_id], d_hash[thr_id], order++);
|
|
|
|
|
|
|
|
|
|
// das ist der unbedingte Branch f<EFBFBD>r Keccak512
|
|
|
|
|
quark_keccak512_cpu_hash_64(thr_id, nrm3, pdata[19], d_branch3Nonces[thr_id], d_hash[thr_id], order++);
|
|
|
|
|
|
|
|
|
|
// das ist der unbedingte Branch f<EFBFBD>r Skein512
|
|
|
|
|
quark_skein512_cpu_hash_64(thr_id, nrm3, pdata[19], d_branch3Nonces[thr_id], d_hash[thr_id], order++);
|
|
|
|
|
|
|
|
|
|
// quarkNonces in branch1 und branch2 aufsplitten gem<EFBFBD>ss if (hash[0] & 0x8)
|
|
|
|
|
quark_compactTest_cpu_hash_64(thr_id, nrm3, pdata[19], d_hash[thr_id], d_branch3Nonces[thr_id],
|
|
|
|
|
d_branch1Nonces[thr_id], &nrm1,
|
|
|
|
|
d_branch2Nonces[thr_id], &nrm2,
|
|
|
|
|
order++);
|
|
|
|
|
|
|
|
|
|
// das ist der bedingte Branch f<EFBFBD>r Keccak512
|
|
|
|
|
quark_keccak512_cpu_hash_64(thr_id, nrm1, pdata[19], d_branch1Nonces[thr_id], d_hash[thr_id], order++);
|
|
|
|
|
|
|
|
|
|
// das ist der bedingte Branch f<EFBFBD>r JH512
|
|
|
|
|
quark_jh512_cpu_hash_64(thr_id, nrm2, pdata[19], d_branch2Nonces[thr_id], d_hash[thr_id], order++);
|
|
|
|
|
|
|
|
|
|
// Scan nach Gewinner Hashes auf der GPU
|
|
|
|
|
uint32_t foundNonce = cuda_check_cpu_hash_64(thr_id, nrm3, pdata[19], d_branch3Nonces[thr_id], d_hash[thr_id], order++);
|
|
|
|
|
if (foundNonce != 0xffffffff)
|
|
|
|
|
{
|
|
|
|
|
uint32_t vhash64[8];
|
|
|
|
|
be32enc(&endiandata[19], foundNonce);
|
|
|
|
|
animehash(vhash64, endiandata);
|
|
|
|
|
|
|
|
|
|
if ((vhash64[7]<=Htarg) && fulltest(vhash64, ptarget)) {
|
|
|
|
|
|
|
|
|
|
pdata[19] = foundNonce;
|
|
|
|
|
*hashes_done = (foundNonce - first_nonce + 1)/2;
|
|
|
|
|
return 1;
|
|
|
|
|
} else {
|
|
|
|
|
applog(LOG_INFO, "GPU #%d: result for nonce $%08X does not validate on CPU!", thr_id, foundNonce);
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
pdata[19] += throughput;
|
|
|
|
|
|
|
|
|
|
} while (pdata[19] < max_nonce && !work_restart[thr_id].restart);
|
|
|
|
|
|
|
|
|
|
*hashes_done = (pdata[19] - first_nonce + 1)/2;
|
|
|
|
|
return 0;
|
|
|
|
|
}
|