Fixed Double Paste
This commit is contained in:
parent
a98fbf6057
commit
4324925782
63
lib/ircbot.py
Normal file
63
lib/ircbot.py
Normal file
@ -0,0 +1,63 @@
|
||||
import base64
|
||||
import gc
|
||||
import json
|
||||
import os
|
||||
import random
|
||||
import sys
|
||||
import time
|
||||
import signal
|
||||
import traceback
|
||||
import urlparse
|
||||
import twisted
|
||||
import settings
|
||||
#if settings.BOT_ENABLED:
|
||||
from twisted.words.protocols import irc
|
||||
class IRCClient(irc.IRCClient):
|
||||
nickname = settings.BOT_NICK
|
||||
channel = settings.BOT_CHANNEL
|
||||
|
||||
def lineReceived(self, line):
|
||||
log.debug(line)
|
||||
irc.IRCClient.lineReceived(self, line)
|
||||
|
||||
def signedOn(self):
|
||||
self.in_channel = False
|
||||
irc.IRCClient.signedOn(self)
|
||||
self.factory.resetDelay()
|
||||
self.join(self.channel)
|
||||
|
||||
@defer.inlineCallbacks
|
||||
def new_share(share):
|
||||
if not self.in_channel:
|
||||
return
|
||||
if share.pow_hash <= share.header['bits'].target and abs(share.timestamp - time.time()) < 10*60:
|
||||
yield deferral.sleep(random.expovariate(1/60))
|
||||
message = '\x02%s BLOCK FOUND by %s! %s%064x' % (net.NAME.upper(), bitcoin_data.script2_to_address(share.new_script, net.PARENT), net.PARENT.BLOCK_EXPLORER_URL_PREFIX, share.header_hash)
|
||||
if all('%x' % (share.header_hash,) not in old_message for old_message in self.recent_messages):
|
||||
self.say(self.channel, message)
|
||||
self._remember_message(message)
|
||||
self.watch_id = node.tracker.verified.added.watch(new_share)
|
||||
self.recent_messages = []
|
||||
|
||||
def joined(self, channel):
|
||||
self.in_channel = True
|
||||
|
||||
def left(self, channel):
|
||||
self.in_channel = False
|
||||
|
||||
def _remember_message(self, message):
|
||||
self.recent_messages.append(message)
|
||||
while len(self.recent_messages) > 100:
|
||||
self.recent_messages.pop(0)
|
||||
|
||||
def privmsg(self, user, channel, message):
|
||||
if channel == self.channel:
|
||||
self._remember_message(message)
|
||||
|
||||
def connectionLost(self, reason):
|
||||
node.tracker.verified.added.unwatch(self.watch_id)
|
||||
print 'IRC connection lost:', reason.getErrorMessage()
|
||||
|
||||
class IRCClientFactory(protocol.ReconnectingClientFactory):
|
||||
protocol = IRCClient
|
||||
reactor.connectTCP(settings.BOT_NETWORK,settings.BOT_PORT, IRCClientFactory())
|
||||
@ -4,128 +4,6 @@ from datetime import datetime
|
||||
import Queue
|
||||
import signal
|
||||
import Cache
|
||||
import notify_email
|
||||
from sets import Set
|
||||
|
||||
import lib.settings as settings
|
||||
|
||||
import lib.logger
|
||||
log = lib.logger.get_logger('DBInterface')
|
||||
|
||||
class DBInterface():
|
||||
def __init__(self):
|
||||
self.dbi = self.connectDB()
|
||||
|
||||
def init_main(self):
|
||||
self.dbi.check_tables()
|
||||
|
||||
self.q = Queue.Queue()
|
||||
self.queueclock = None
|
||||
|
||||
self.cache = Cache.Cache()
|
||||
|
||||
self.email = notify_email.NOTIFY_EMAIL()
|
||||
|
||||
self.nextStatsUpdate = 0
|
||||
|
||||
self.scheduleImport()
|
||||
|
||||
self.next_force_import_time = time.time() + settings.DB_LOADER_FORCE_TIME
|
||||
self.next_force_notify_time = time.time() + settings.NOTIFY_DEADMINER_INTERVAL
|
||||
|
||||
|
||||
signal.signal(signal.SIGINT, self.signal_handler)
|
||||
|
||||
def signal_handler(self, signal, frame):
|
||||
log.warning("SIGINT Detected, shutting down")
|
||||
self.do_import(self.dbi, True)
|
||||
reactor.stop()
|
||||
|
||||
def set_bitcoinrpc(self, bitcoinrpc):
|
||||
self.bitcoinrpc = bitcoinrpc
|
||||
|
||||
def connectDB(self):
|
||||
if settings.DATABASE_DRIVER == "sqlite":
|
||||
log.debug('DB_Sqlite INIT')
|
||||
import DB_Sqlite
|
||||
return DB_Sqlite.DB_Sqlite()
|
||||
elif settings.DATABASE_DRIVER == "mysql":
|
||||
if settings.VARIABLE_DIFF:
|
||||
log.debug("DB_Mysql_Vardiff INIT")
|
||||
import DB_Mysql_Vardiff
|
||||
return DB_Mysql_Vardiff.DB_Mysql_Vardiff()
|
||||
else:
|
||||
log.debug('DB_Mysql INIT')
|
||||
import DB_Mysql
|
||||
return DB_Mysql.DB_Mysql()
|
||||
elif settings.DATABASE_DRIVER == "postgresql":
|
||||
log.debug('DB_Postgresql INIT')
|
||||
import DB_Postgresql
|
||||
return DB_Postgresql.DB_Postgresql()
|
||||
elif settings.DATABASE_DRIVER == "none":
|
||||
log.debug('DB_None INIT')
|
||||
import DB_None
|
||||
return DB_None.DB_None()
|
||||
else:
|
||||
log.error('Invalid DATABASE_DRIVER -- using NONE')
|
||||
log.debug('DB_None INIT')
|
||||
import DB_None
|
||||
return DB_None.DB_None()
|
||||
|
||||
def scheduleImport(self):
|
||||
# This schedule's the Import
|
||||
if settings.DATABASE_DRIVER == "sqlite":
|
||||
use_thread = False
|
||||
else:
|
||||
use_thread = True
|
||||
|
||||
if use_thread:
|
||||
self.queueclock = reactor.callLater(settings.DB_LOADER_CHECKTIME , self.run_import_thread)
|
||||
else:
|
||||
self.queueclock = reactor.callLater(settings.DB_LOADER_CHECKTIME , self.run_import)
|
||||
|
||||
def run_import_thread(self):
|
||||
log.debug("run_import_thread current size: %d", self.q.qsize())
|
||||
|
||||
if self.q.qsize() >= settings.DB_LOADER_REC_MIN or time.time() >= self.next_force_import_time: # Don't incur thread overhead if we're not going to run
|
||||
reactor.callInThread(self.import_thread)
|
||||
|
||||
self.scheduleImport()
|
||||
|
||||
def run_import(self):
|
||||
log.debug("DBInterface.run_import called")
|
||||
|
||||
self.do_import(self.dbi, False)
|
||||
|
||||
self.scheduleImport()
|
||||
|
||||
def import_thread(self):
|
||||
# Here we are in the thread.
|
||||
dbi = self.connectDB()
|
||||
self.do_import(dbi, False)
|
||||
|
||||
dbi.close()
|
||||
|
||||
def _update_pool_info(self, data):
|
||||
self.dbi.update_pool_info({ 'blocks' : data['blocks'], 'balance' : data['balance'],
|
||||
'connections' : data['connections'], 'difficulty' : data['difficulty'] })
|
||||
|
||||
def do_import(self, dbi, force):
|
||||
log.debug("DBInterface.do_import called. force: %s, queue size: %s", 'yes' if force == True else 'no', self.q.qsize())
|
||||
|
||||
# Flush the whole queue on force
|
||||
forcesize = 0
|
||||
if force == True:
|
||||
forcesize = self.q.qsize()
|
||||
|
||||
# Only run if we have data
|
||||
while self.q.empty() == False and (force == True or self.q.qsize() >= settings.DB_LOADER_REC_MIN or time.time() >= self.next_force_import_time or forcesize > 0):
|
||||
from twisted.internet import reactor, defer
|
||||
import time
|
||||
from datetime import datetime
|
||||
import Queue
|
||||
import signal
|
||||
import Cache
|
||||
from sets import Set
|
||||
import notify_email
|
||||
import lib.settings as settings
|
||||
@ -341,225 +219,3 @@ class DBInterface():
|
||||
def clear_worker_diff(self):
|
||||
return self.dbi.clear_worker_diff()
|
||||
|
||||
from twisted.internet import reactor, defer
|
||||
import time
|
||||
from datetime import datetime
|
||||
import Queue
|
||||
import signal
|
||||
import Cache
|
||||
from sets import Set
|
||||
import notify_email
|
||||
import lib.settings as settings
|
||||
|
||||
import lib.logger
|
||||
log = lib.logger.get_logger('DBInterface')
|
||||
|
||||
class DBInterface():
|
||||
def __init__(self):
|
||||
self.dbi = self.connectDB()
|
||||
|
||||
def init_main(self):
|
||||
self.dbi.check_tables()
|
||||
self.q = Queue.Queue()
|
||||
self.queueclock = None
|
||||
|
||||
self.cache = Cache.Cache()
|
||||
self.email = notify_email.NOTIFY_EMAIL()
|
||||
self.nextStatsUpdate = 0
|
||||
|
||||
self.scheduleImport()
|
||||
|
||||
self.next_force_import_time = time.time() + settings.DB_LOADER_FORCE_TIME
|
||||
self.next_force_notify_time = time.time() + settings.NOTIFY_DEADMINER_INTERVAL
|
||||
signal.signal(signal.SIGINT, self.signal_handler)
|
||||
|
||||
def signal_handler(self, signal, frame):
|
||||
log.warning("SIGINT Detected, shutting down")
|
||||
self.do_import(self.dbi, True)
|
||||
reactor.stop()
|
||||
|
||||
def set_bitcoinrpc(self, bitcoinrpc):
|
||||
self.bitcoinrpc = bitcoinrpc
|
||||
|
||||
def connectDB(self):
|
||||
if settings.DATABASE_DRIVER == "sqlite":
|
||||
log.debug('DB_Sqlite INIT')
|
||||
import DB_Sqlite
|
||||
return DB_Sqlite.DB_Sqlite()
|
||||
elif settings.DATABASE_DRIVER == "mysql":
|
||||
if settings.VARIABLE_DIFF:
|
||||
log.debug("DB_Mysql_Vardiff INIT")
|
||||
import DB_Mysql_Vardiff
|
||||
return DB_Mysql_Vardiff.DB_Mysql_Vardiff()
|
||||
else:
|
||||
log.debug('DB_Mysql INIT')
|
||||
import DB_Mysql
|
||||
return DB_Mysql.DB_Mysql()
|
||||
elif settings.DATABASE_DRIVER == "postgresql":
|
||||
log.debug('DB_Postgresql INIT')
|
||||
import DB_Postgresql
|
||||
return DB_Postgresql.DB_Postgresql()
|
||||
elif settings.DATABASE_DRIVER == "none":
|
||||
log.debug('DB_None INIT')
|
||||
import DB_None
|
||||
return DB_None.DB_None()
|
||||
else:
|
||||
log.error('Invalid DATABASE_DRIVER -- using NONE')
|
||||
log.debug('DB_None INIT')
|
||||
import DB_None
|
||||
return DB_None.DB_None()
|
||||
|
||||
def scheduleImport(self):
|
||||
# This schedule's the Import
|
||||
if settings.DATABASE_DRIVER == "sqlite":
|
||||
use_thread = False
|
||||
else:
|
||||
use_thread = True
|
||||
|
||||
if use_thread:
|
||||
self.queueclock = reactor.callLater(settings.DB_LOADER_CHECKTIME , self.run_import_thread)
|
||||
else:
|
||||
self.queueclock = reactor.callLater(settings.DB_LOADER_CHECKTIME , self.run_import)
|
||||
|
||||
def run_import_thread(self):
|
||||
log.debug("run_import_thread current size: %d", self.q.qsize())
|
||||
|
||||
if self.q.qsize() >= settings.DB_LOADER_REC_MIN or time.time() >= self.next_force_import_time: # Don't incur thread overhead if we're not going to run
|
||||
reactor.callInThread(self.import_thread)
|
||||
|
||||
self.scheduleImport()
|
||||
|
||||
def run_import(self):
|
||||
log.debug("DBInterface.run_import called")
|
||||
|
||||
self.do_import(self.dbi, False)
|
||||
|
||||
self.scheduleImport()
|
||||
|
||||
def import_thread(self):
|
||||
# Here we are in the thread.
|
||||
dbi = self.connectDB()
|
||||
self.do_import(dbi, False)
|
||||
|
||||
dbi.close()
|
||||
|
||||
def _update_pool_info(self, data):
|
||||
self.dbi.update_pool_info({ 'blocks' : data['blocks'], 'balance' : data['balance'],
|
||||
'connections' : data['connections'], 'difficulty' : data['difficulty'] })
|
||||
|
||||
def do_import(self, dbi, force):
|
||||
log.debug("DBInterface.do_import called. force: %s, queue size: %s", 'yes' if force == True else 'no', self.q.qsize())
|
||||
|
||||
# Flush the whole queue on force
|
||||
forcesize = 0
|
||||
if force == True:
|
||||
forcesize = self.q.qsize()
|
||||
|
||||
# Only run if we have data
|
||||
while self.q.empty() == False and (force == True or self.q.qsize() >= settings.DB_LOADER_REC_MIN or time.time() >= self.next_force_import_time or forcesize > 0):
|
||||
self.next_force_import_time = time.time() + settings.DB_LOADER_FORCE_TIME
|
||||
|
||||
force = False
|
||||
# Put together the data we want to import
|
||||
sqldata = []
|
||||
datacnt = 0
|
||||
|
||||
while self.q.empty() == False and datacnt < settings.DB_LOADER_REC_MAX:
|
||||
datacnt += 1
|
||||
data = self.q.get()
|
||||
sqldata.append(data)
|
||||
self.q.task_done()
|
||||
|
||||
forcesize -= datacnt
|
||||
|
||||
# try to do the import, if we fail, log the error and put the data back in the queue
|
||||
try:
|
||||
log.info("Inserting %s Share Records", datacnt)
|
||||
dbi.import_shares(sqldata)
|
||||
except Exception as e:
|
||||
log.error("Insert Share Records Failed: %s", e.args[0])
|
||||
for k, v in enumerate(sqldata):
|
||||
self.q.put(v)
|
||||
break # Allows us to sleep a little
|
||||
|
||||
def queue_share(self, data):
|
||||
self.q.put(data)
|
||||
|
||||
def found_block(self, data):
|
||||
try:
|
||||
log.info("Updating Found Block Share Record")
|
||||
self.do_import(self.dbi, True) # We can't Update if the record is not there.
|
||||
self.dbi.found_block(data)
|
||||
except Exception as e:
|
||||
log.error("Update Found Block Share Record Failed: %s", e.args[0])
|
||||
|
||||
def check_password(self, username, password):
|
||||
if username == "":
|
||||
log.info("Rejected worker for blank username")
|
||||
return False
|
||||
allowed_chars = Set('0123456789abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ_-.')
|
||||
if Set(username).issubset(allowed_chars) != True:
|
||||
log.info("Username contains bad arguments")
|
||||
return False
|
||||
if username.count('.') > 1:
|
||||
log.info("Username contains multiple . ")
|
||||
return False
|
||||
|
||||
# Force username and password to be strings
|
||||
username = str(username)
|
||||
password = str(password)
|
||||
if not settings.USERS_CHECK_PASSWORD and self.user_exists(username):
|
||||
return True
|
||||
elif self.cache.get(username) == password:
|
||||
return True
|
||||
elif self.dbi.check_password(username, password):
|
||||
self.cache.set(username, password)
|
||||
return True
|
||||
elif settings.USERS_AUTOADD == True:
|
||||
if self.dbi.get_uid(username) != False:
|
||||
uid = self.dbi.get_uid(username)
|
||||
self.dbi.insert_worker(uid, username, password)
|
||||
self.cache.set(username, password)
|
||||
return True
|
||||
|
||||
log.info("Authentication for %s failed" % username)
|
||||
return False
|
||||
|
||||
def list_users(self):
|
||||
return self.dbi.list_users()
|
||||
|
||||
def get_user(self, id):
|
||||
return self.dbi.get_user(id)
|
||||
|
||||
def user_exists(self, username):
|
||||
if self.cache.get(username) is not None:
|
||||
return True
|
||||
user = self.dbi.get_user(username)
|
||||
return user is not None
|
||||
|
||||
def insert_user(self, username, password):
|
||||
return self.dbi.insert_user(username, password)
|
||||
|
||||
def delete_user(self, username):
|
||||
self.mc.delete(username)
|
||||
self.usercache = {}
|
||||
return self.dbi.delete_user(username)
|
||||
|
||||
def update_user(self, username, password):
|
||||
self.mc.delete(username)
|
||||
self.mc.set(username, password)
|
||||
return self.dbi.update_user(username, password)
|
||||
|
||||
def update_worker_diff(self, username, diff):
|
||||
return self.dbi.update_worker_diff(username, diff)
|
||||
|
||||
def get_pool_stats(self):
|
||||
return self.dbi.get_pool_stats()
|
||||
|
||||
def get_workers_stats(self):
|
||||
return self.dbi.get_workers_stats()
|
||||
|
||||
def clear_worker_diff(self):
|
||||
return self.dbi.clear_worker_diff()
|
||||
|
||||
|
||||
|
||||
Loading…
Reference in New Issue
Block a user