mirror of
https://github.com/r4sas/Niflheim-api
synced 2025-01-12 07:58:00 +00:00
some minor fixes
This commit is contained in:
parent
f7c20b67f8
commit
5c3a33e0b9
40
vserv.py
40
vserv.py
@ -39,44 +39,44 @@ def isdatabase(db_path):
|
||||
if not os.path.exists(db_path):
|
||||
os.makedirs(db_path)
|
||||
try:
|
||||
conn = sqlite3.connect(db_path + 'yggindex.db')
|
||||
c = conn.cursor()
|
||||
dbconn = sqlite3.connect(db_path + 'yggindex.db')
|
||||
c = dbconn.cursor()
|
||||
c.execute('''create table yggindex(ipv6 varchar(45) UNIQUE, coords varchar(50),\
|
||||
dt datetime default (strftime('%s','now')))''')
|
||||
c.execute('''create table timeseries(max varchar(45),\
|
||||
dt datetime default (strftime('%s','now')))''')
|
||||
c.execute('''create table contrib(ipv6 varchar(45) UNIQUE,\
|
||||
ut unixtime default (strftime('%s','now')))''')
|
||||
conn.commit()
|
||||
dbconn.commit()
|
||||
except Error as e:
|
||||
print(e)
|
||||
finally:
|
||||
conn.close()
|
||||
dbconn.close()
|
||||
else:
|
||||
print("found database will not create a new one")
|
||||
|
||||
|
||||
def insert_new_entry(db_path, ipv6, coords):
|
||||
try:
|
||||
conn = sqlite3.connect(db_path + "yggindex.db")
|
||||
c.execute('''INSERT OR REPLACE INTO yggindex(ipv6, coords) VALUES(?, ?)''',\
|
||||
dbconn = sqlite3.connect(db_path + "yggindex.db")
|
||||
dbconn.execute('''INSERT OR REPLACE INTO yggindex(ipv6, coords) VALUES(?, ?)''',\
|
||||
(ipv6, coords))
|
||||
conn.commit()
|
||||
conn.close()
|
||||
dbconn.commit()
|
||||
dbconn.close()
|
||||
except Error as e:
|
||||
print e
|
||||
|
||||
def contrib_entry(db_path, ipv6):
|
||||
try:
|
||||
conn = sqlite3.connect(db_path + "yggindex.db")
|
||||
c.execute('''INSERT OR REPLACE INTO contrib(ipv6) VALUES(''' + "'"\
|
||||
dbconn = sqlite3.connect(db_path + "yggindex.db")
|
||||
dbconn.execute('''INSERT OR REPLACE INTO contrib(ipv6) VALUES(''' + "'"\
|
||||
+ ipv6 + "'" + ''')''')
|
||||
conn.commit()
|
||||
conn.close()
|
||||
dbconn.commit()
|
||||
dbconn.close()
|
||||
except Error as e:
|
||||
print e
|
||||
|
||||
def error_check_insert_into_db(dht, switchpeers, addr):
|
||||
def error_check_insert_into_db(dht, switchpeers, ipv6):
|
||||
try:
|
||||
if dht.get("status") == "success":
|
||||
for x, y in dht["response"]["dht"].iteritems():
|
||||
@ -88,17 +88,17 @@ def error_check_insert_into_db(dht, switchpeers, addr):
|
||||
if valid_ipv6_check(x[1]["ip"]) and check_coords(x[1]["coords"]):
|
||||
insert_new_entry(DB_PATH, x[1]["ip"], x[1]["coords"])
|
||||
|
||||
contrib_entry(addr)
|
||||
contrib_entry(DB_PATH, ipv6)
|
||||
except:
|
||||
print"error in json file, aborting"
|
||||
|
||||
|
||||
def proccess_incoming_data(datty, addr):
|
||||
print str(time.time()) + " " + addr
|
||||
def proccess_incoming_data(datty, ipv6):
|
||||
print str(time.time()) + " " + str(ipv6)
|
||||
try:
|
||||
ddata = datty.recv(1024 * 20)
|
||||
data = json.loads(ddata.decode())
|
||||
error_check_insert_into_db(data["dhtpack"], data["switchpack"])
|
||||
ddata = datty.recv(1024 * 20)
|
||||
data = json.loads(ddata.decode())
|
||||
error_check_insert_into_db(data["dhtpack"], data["switchpack"], ipv6)
|
||||
except:
|
||||
print "ignoring, data was not json"
|
||||
|
||||
@ -112,6 +112,6 @@ conn.listen(30)
|
||||
while True:
|
||||
try:
|
||||
dataraw, addr = conn.accept()
|
||||
thread.start_new_thread(proccess_incoming_data, (dataraw, addr))
|
||||
thread.start_new_thread(proccess_incoming_data, (dataraw, addr[0]))
|
||||
except Exception:
|
||||
print "bloop"
|
||||
|
Loading…
Reference in New Issue
Block a user