package main import ( "fmt" "log" "net" "strconv" "strings" "sync" "time" "github.com/btcsuite/btcd/wire" ) const ( // NOUNCE is used to check if we connect to ourselves // as we don't listen we can use a fixed value nounce = 0x0539a019ca550825 minPort = 0 maxPort = 65535 crawlDelay = 22 // seconds between start crawlwer ticks auditDelay = 22 // minutes between audit channel ticks dnsDelay = 57 // seconds between updates to active dns record list maxFails = 58 // max number of connect fails before we delete a node. Just over 24 hours(checked every 33 minutes) maxTo = 250 // max seconds (4min 10 sec) for all comms to node to complete before we timeout ) const ( dnsInvalid = iota // dnsV4Std // ip v4 using network standard port dnsV4Non // ip v4 using network non standard port dnsV6Std // ipv6 using network standard port dnsV6Non // ipv6 using network non standard port maxDNSTypes // used in main to allocate slice ) const ( // node status statusRG = iota // reported good status. A remote node has reported this ip but we have not connected statusCG // confirmed good. We have connected to the node and received addresses statusWG // was good. node was confirmed good but now having problems statusNG // no good. Will be removed from theList after 24 hours to redure bouncing ip addresses maxStatusTypes // used in main to allocate slice ) type dnsseeder struct { id wire.BitcoinNet // Magic number - Unique ID for this network. Sent in header of all messages theList map[string]*node // the list of current nodes mtx sync.RWMutex // protect thelist dnsHost string // dns host we will serve results for this domain name string // Short name for the network desc string // Long description for the network initialIP string // Initial ip address to connect to and ask for addresses if we have no seeders seeders []string // slice of seeders to pull ip addresses when starting this seeder maxStart []uint32 // max number of goroutines to start each run for each status type delay []int64 // number of seconds to wait before we connect to a known client for each status counts NodeCounts // structure to hold stats for this seeder pver uint32 // minimum block height for the seeder ttl uint32 // DNS TTL to use for this seeder maxSize int // max number of clients before we start restricting new entries port uint16 // default network port this seeder uses } type result struct { nas []*wire.NetAddress // slice of node addresses returned from a node msg *crawlError // error string or nil if no problems node string // theList key to the node that was crawled version int32 // remote node protocol version services wire.ServiceFlag // remote client supported services lastBlock int32 // last block seen by the node strVersion string // remote client user agent } // initCrawlers needs to be run before the startCrawlers so it can get // a list of current ip addresses from the other seeders and therefore // start the crawl process func (s *dnsseeder) initSeeder() { // range over existing seeders for the network and get starting ip addresses from them for _, aseeder := range s.seeders { c := 0 if aseeder == "" { continue } newRRs, err := net.LookupHost(aseeder) if err != nil { log.Printf("%s: unable to do initial lookup to seeder %s %v\n", s.name, aseeder, err) continue } for _, ip := range newRRs { if newIP := net.ParseIP(ip); newIP != nil { // 1 at the end is the services flag if x := s.addNa(wire.NewNetAddressIPPort(newIP, s.port, 1)); x == true { c++ } } } if config.verbose { log.Printf("%s: completed import of %v addresses from %s\n", s.name, c, aseeder) } } // load ip addresses into system and start crawling from them if len(s.theList) == 0 && s.initialIP != "" { for _, initialIP := range strings.Split(s.initialIP, ",") { if newIP := net.ParseIP(initialIP); newIP != nil { // 1 at the end is the services flag if x := s.addNa(wire.NewNetAddressIPPort(newIP, s.port, 1)); x == true { log.Printf("%s: crawling with initial IP %s \n", s.name, s.initialIP) } } } } if len(s.theList) == 0 { log.Printf("%s: Error: No ip addresses from seeders so I have nothing to crawl.\n", s.name) for _, v := range s.seeders { log.Printf("%s: Seeder: %s\n", s.name, v) } log.Printf("%s: Initial IP: %s\n", s.name, s.initialIP) } } // runSeeder runs a seeder in an endless goroutine func (s *dnsseeder) runSeeder(done <-chan struct{}, wg *sync.WaitGroup) { defer wg.Done() // receive the results from the crawl goroutines resultsChan := make(chan *result) // load data from other seeders so we can start crawling nodes s.initSeeder() // start initial scan now so we don't have to wait for the timers to fire s.startCrawlers(resultsChan) // create timing channels for regular tasks auditChan := time.NewTicker(time.Minute * auditDelay).C crawlChan := time.NewTicker(time.Second * crawlDelay).C dnsChan := time.NewTicker(time.Second * dnsDelay).C dowhile := true for dowhile == true { select { case r := <-resultsChan: // process a results structure from a crawl s.processResult(r) case <-dnsChan: // update the system with the latest selection of dns records s.loadDNS() case <-auditChan: // keep theList clean and tidy s.auditNodes() case <-crawlChan: // start a scan to crawl nodes s.startCrawlers(resultsChan) case <-done: // done channel closed so exit the select and shutdown the seeder dowhile = false } } fmt.Printf("shutting down seeder: %s\n", s.name) // end the goroutine & defer will call wg.Done() } // startCrawlers is called on a time basis to start maxcrawlers new // goroutines if there are spare goroutine slots available func (s *dnsseeder) startCrawlers(resultsChan chan *result) { s.mtx.RLock() defer s.mtx.RUnlock() tcount := uint32(len(s.theList)) if tcount == 0 { if config.debug { log.Printf("%s - debug - startCrawlers fail: no node available\n", s.name) } return } started := make([]uint32, maxStatusTypes) totals := make([]uint32, maxStatusTypes) // range on a map will not return items in the same order each time // so this is a random'ish selection for _, nd := range s.theList { totals[nd.status]++ if nd.crawlActive == true { continue } // do we already have enough started at this status if started[nd.status] >= s.maxStart[nd.status] { continue } // don't crawl a node to quickly if (time.Now().Unix() - s.delay[nd.status]) <= nd.lastTry.Unix() { continue } // all looks good so start a go routine to crawl the remote node nd.crawlActive = true nd.crawlStart = time.Now() go crawlNode(resultsChan, s, nd) started[nd.status]++ } // update the global stats in another goroutine to free the main goroutine // for other work go updateNodeCounts(s, tcount, started, totals) // returns and read lock released } // processResult will add new nodes to the list and update the status of the crawled node func (s *dnsseeder) processResult(r *result) { var nd *node s.mtx.Lock() defer s.mtx.Unlock() if _, ok := s.theList[r.node]; ok { nd = s.theList[r.node] } else { log.Printf("%s: warning - ignoring results from unknown node: %s\n", s.name, r.node) return } // now nd has been set to a valid pointer we can use it in a defer defer crawlEnd(nd) // msg is a crawlerror or nil if r.msg != nil { // update the fact that we have not connected to this node nd.lastTry = time.Now() nd.connectFails++ nd.statusStr = r.msg.Error() // update the status of this failed node switch nd.status { case statusRG: // if we are full then any RG failures will skip directly to NG if len(s.theList) > s.maxSize { nd.status = statusNG // not able to connect to this node so ignore } else { if nd.rating += 25; nd.rating > 30 { nd.status = statusWG } } case statusCG: if nd.rating += 25; nd.rating >= 50 { nd.status = statusWG } case statusWG: if nd.rating += 15; nd.rating >= 100 { nd.status = statusNG // not able to connect to this node so ignore } } // no more to do so return which will shutdown the goroutine & call // the deffered cleanup if config.verbose { log.Printf("%s: failed crawl node: %s s:r:f: %v:%v:%v %s\n", s.name, net.JoinHostPort(nd.na.IP.String(), strconv.Itoa(int(nd.na.Port))), nd.status, nd.rating, nd.connectFails, nd.statusStr) } return } // succesful connection and addresses received so mark status nd.status = statusCG cs := nd.lastConnect nd.rating = 0 nd.connectFails = 0 nd.lastConnect = time.Now() nd.lastTry = nd.lastConnect nd.statusStr = "ok: received remote address list" // update the node from the results nd.version = r.version nd.services = r.services nd.lastBlock = r.lastBlock nd.strVersion = r.strVersion added := 0 // if we are full then skip adding more possible clients if len(s.theList) < s.maxSize { // do not accept more than one third of maxSize addresses from one node oneThird := int(float64(s.maxSize / 3)) // loop through all the received network addresses and add to thelist if not present for _, na := range r.nas { // a new network address so add to the system if x := s.addNa(na); x == true { if added++; added > oneThird { break } } } } if config.verbose { log.Printf("%s: crawl done: node: %s s:r:f: %v:%v:%v addr: %v:%v CrawlTime: %s Last connect: %v ago\n", s.name, net.JoinHostPort(nd.na.IP.String(), strconv.Itoa(int(nd.na.Port))), nd.status, nd.rating, nd.connectFails, len(r.nas), added, time.Since(nd.crawlStart).String(), time.Since(cs).String()) } } // crawlEnd is run as a defer to make sure node status is correctly updated func crawlEnd(nd *node) { nd.crawlActive = false } // addNa validates and adds a network address to theList func (s *dnsseeder) addNa(nNa *wire.NetAddress) bool { if len(s.theList) > s.maxSize { return false } // generate the key and add to theList k := net.JoinHostPort(nNa.IP.String(), strconv.Itoa(int(nNa.Port))) if _, dup := s.theList[k]; dup == true { return false } if nNa.Port <= minPort || nNa.Port >= maxPort { return false } // if the reported timestamp suggests the netaddress has not been seen in the last 24 hours // then ignore this netaddress if (time.Now().Add(-(time.Hour * 24))).After(nNa.Timestamp) { return false } nt := node{ na: nNa, lastConnect: time.Now(), version: 0, status: statusRG, dnsType: dnsV4Std, } // select the dns type based on the remote address type and port if x := nt.na.IP.To4(); x == nil { // not ipv4 if nNa.Port != s.port { nt.dnsType = dnsV6Non // produce the nonstdIP nt.nonstdIP = getNonStdIP(nt.na.IP, nt.na.Port) } else { nt.dnsType = dnsV6Std } } else { // ipv4 if nNa.Port != s.port { nt.dnsType = dnsV4Non // force ipv4 address into a 4 byte buffer nt.na.IP = nt.na.IP.To4() // produce the nonstdIP nt.nonstdIP = getNonStdIP(nt.na.IP, nt.na.Port) } } // add the new node details to theList s.theList[k] = &nt return true } // getNonStdIP is given an IP address and a port and returns a fake IP address // that is encoded with the original IP and port number. Remote clients can match // the two and work out the real IP and port from the two IP addresses. func getNonStdIP(rip net.IP, port uint16) net.IP { b := []byte{0x0, 0x0, 0x0, 0x0} crcAddr := crc16(rip.To4()) b[0] = byte(crcAddr >> 8) b[1] = byte((crcAddr & 0xff)) b[2] = byte(port >> 8) b[3] = byte(port & 0xff) encip := net.IPv4(b[0], b[1], b[2], b[3]) if config.debug { log.Printf("debug - encode nonstd - realip: %s port: %v encip: %s crc: %x\n", rip.String(), port, encip.String(), crcAddr) } return encip } // crc16 produces a crc16 from a byte slice func crc16(bs []byte) uint16 { var x, crc uint16 crc = 0xffff for _, v := range bs { x = crc>>8 ^ uint16(v) x ^= x >> 4 crc = (crc << 8) ^ (x << 12) ^ (x << 5) ^ x } return crc } func (s *dnsseeder) auditNodes() { c := 0 // set this early so for this audit run all NG clients will be purged // and space will be made for new, possible CG clients iAmFull := len(s.theList) > s.maxSize // cgGoal is 75% of the max statusCG clients we can crawl with the current network delay & maxStart settings. // This allows us to cycle statusCG users to keep the list fresh cgGoal := int(float64(float64(s.delay[statusCG]/crawlDelay)*float64(s.maxStart[statusCG])) * 0.75) cgCount := 0 log.Printf("%s: Audit start. statusCG Goal: %v System Uptime: %s\n", s.name, cgGoal, time.Since(config.uptime).String()) s.mtx.Lock() defer s.mtx.Unlock() for k, nd := range s.theList { if nd.crawlActive == true { if time.Now().Unix()-nd.crawlStart.Unix() >= 300 { log.Printf("warning - long running crawl > 5 minutes ====\n- %s status:rating:fails %v:%v:%v crawl start: %s last status: %s\n====\n", k, nd.status, nd.rating, nd.connectFails, nd.crawlStart.String(), nd.statusStr) } } // Audit task is to remove node that we have not been able to connect to if nd.status == statusNG && nd.connectFails > maxFails { if config.verbose { log.Printf("%s: purging node %s after %v failed connections\n", s.name, k, nd.connectFails) } c++ // remove the map entry and mark the old node as // nil so garbage collector will remove it s.theList[k] = nil delete(s.theList, k) } // If seeder is full then remove old NG clients and fill up with possible new CG clients if nd.status == statusNG && iAmFull { if config.verbose { log.Printf("%s: seeder full purging node %s\n", s.name, k) } c++ // remove the map entry and mark the old node as // nil so garbage collector will remove it s.theList[k] = nil delete(s.theList, k) } // check if we need to purge statusCG to freshen the list if nd.status == statusCG { if cgCount++; cgCount > cgGoal { // we have enough statusCG clients so purge remaining to cycle through the list if config.verbose { log.Printf("%s: seeder cycle statusCG - purging node %s\n", s.name, k) } c++ // remove the map entry and mark the old node as // nil so garbage collector will remove it s.theList[k] = nil delete(s.theList, k) } } } if config.verbose { log.Printf("%s: Audit complete. %v nodes purged\n", s.name, c) } } // teatload loads the dns records with time based test data func (s *dnsseeder) loadDNS() { updateDNS(s) } // getSeederByName returns a pointer to the seeder based on its name or nil if not found func getSeederByName(name string) *dnsseeder { for _, s := range config.seeders { if s.name == name { return s } } return nil } // isDuplicateSeeder returns true if the seeder details already exist in the application func isDuplicateSeeder(s *dnsseeder) (bool, error) { // check for duplicate seeders with the same details for _, v := range config.seeders { if v.id == s.id { return true, fmt.Errorf("Duplicate Magic id. Already loaded for %s so can not be used for %s", v.id, v.name, s.name) } if v.dnsHost == s.dnsHost { return true, fmt.Errorf("Duplicate DNS names. Already loaded %s for %s so can not be used for %s", v.dnsHost, v.name, s.name) } } return false, nil } /* */