X-Git-Url: https://plomlompom.com/repos/?p=plomlombot-irc.git;a=blobdiff_plain;f=plomlombot.py;h=7f412ea63d5505e11f68b90893c515cbdbe129e4;hp=36228f31b2ccafe949d381958169077a9e6bf77a;hb=e7fab400e3f2bfa5ea5320a7ce8eebf8a01afcdd;hpb=9a7e2920a88ea32412427c1395b47cdf6d2775cb diff --git a/plomlombot.py b/plomlombot.py old mode 100644 new mode 100755 index 36228f3..7f412ea --- a/plomlombot.py +++ b/plomlombot.py @@ -1,33 +1,50 @@ +#!/usr/bin/python3 + +import argparse import socket import datetime import select import time import re -import urllib.request -import html +import requests +import bs4 +import random +import hashlib +import os +import plomsearch + +# Defaults, may be overwritten by command line arguments. +SERVER = "irc.freenode.net" +PORT = 6667 +TIMEOUT = 240 +USERNAME = "plomlombot" +NICKNAME = USERNAME +TWTFILE = "" +DBDIR = os.path.expanduser("~/plomlombot_db") + + +class ExceptionForRestart(Exception): + pass -servernet = "irc.freenode.net" -port = 6667 -servername = "" -timeout = 240 -username = "plomlombot" -nickname = username -channel = "#zrolaps-test" class IO: - def __init__(self, server, port): + def __init__(self, server, port, timeout): + self.timeout = timeout self.socket = socket.socket() self.socket.connect((server, port)) self.socket.setblocking(0) self.line_buffer = [] self.rune_buffer = "" self.last_pong = time.time() + self.servername = self.recv_line(send_ping=False).split(" ")[0][1:] - def _pingtest(self): - if self.last_pong + timeout < time.time(): - raise RuntimeError("server not answering") - self.send_line("PING " + nickname + " " + servername) + def _pingtest(self, send_ping=True): + if self.last_pong + self.timeout < time.time(): + print("SERVER NOT ANSWERING") + raise ExceptionForRestart + if send_ping: + self.send_line("PING " + self.servername) def send_line(self, msg): msg = msg.replace("\r", " ") @@ -35,69 +52,334 @@ class IO: if len(msg.encode("utf-8")) > 510: print("NOT SENT LINE TO SERVER (too long): " + msg) print("LINE TO SERVER: " - + str(datetime.datetime.now()) + ": " + msg) + + str(datetime.datetime.now()) + ": " + msg) msg = msg + "\r\n" msg_len = len(msg) total_sent_len = 0 while total_sent_len < msg_len: sent_len = self.socket.send(bytes(msg[total_sent_len:], "UTF-8")) if sent_len == 0: - raise RuntimeError("socket connection broken") + print("SOCKET CONNECTION BROKEN") + raise ExceptionForRestart total_sent_len += sent_len - def recv_line_wrapped(self): + def _recv_line_wrapped(self, send_ping=True): if len(self.line_buffer) > 0: return self.line_buffer.pop(0) while True: - ready = select.select([self.socket], [], [], int(timeout / 2)) + ready = select.select([self.socket], [], [], int(self.timeout / 2)) if not ready[0]: - self._pingtest() + self._pingtest(send_ping) return None self.last_pong = time.time() - received_runes = self.socket.recv(1024).decode("UTF-8") + received_bytes = self.socket.recv(1024) + try: + received_runes = received_bytes.decode("UTF-8") + except UnicodeDecodeError: + received_runes = received_bytes.decode("latin1") if len(received_runes) == 0: - raise RuntimeError("socket connection broken") - self.rune_buffer += received_runes + print("SOCKET CONNECTION BROKEN") + raise ExceptionForRestart + self.rune_buffer += received_runes lines_split = str.split(self.rune_buffer, "\r\n") self.line_buffer += lines_split[:-1] self.rune_buffer = lines_split[-1] if len(self.line_buffer) > 0: return self.line_buffer.pop(0) - def recv_line(self): - line = self.recv_line_wrapped() + def recv_line(self, send_ping=True): + line = self._recv_line_wrapped(send_ping) if line: print("LINE FROM SERVER " + str(datetime.datetime.now()) + ": " + - line) + line) return line -def url_check(msg): - matches = re.findall("(https?://[^\s]+)", msg) - for i in range(len(matches)): - url = matches[i] - webpage = urllib.request.urlopen(url, timeout=15) - content_type = webpage.info().get_content_type() - charset = webpage.info().get_content_charset() - if not charset or not content_type in ('text/html', 'text/xml', - 'application/xhtml+xml'): - continue - content = webpage.read().decode(charset) - title = str(content).split('')[1].split('')[0] - title = html.unescape(title) - io.send_line("PRIVMSG " + target + " :page title for url: " + title) - -io = IO(servernet, port) -io.send_line("NICK " + nickname) -io.send_line("USER " + username + " 0 * : ") -io.send_line("JOIN " + channel) -servername = io.recv_line().split(" ")[0][1:] -while 1: - line = io.recv_line() - if not line: - continue - tokens = line.split(" ") - if len(tokens) > 1: - if tokens[1] == "PRIVMSG": + +def handle_command(command, argument, notice, target, session): + hash_string = hashlib.md5(target.encode("utf-8")).hexdigest() + quotesfile_name = session.dbdir + "/quotes_" + hash_string + + def addquote(): + if not os.access(quotesfile_name, os.F_OK): + quotesfile = open(quotesfile_name, "w") + quotesfile.write("QUOTES FOR " + target + ":\n") + quotesfile.close() + quotesfile = open(quotesfile_name, "a") + quotesfile.write(argument + "\n") + quotesfile.close() + quotesfile = open(quotesfile_name, "r") + lines = quotesfile.readlines() + quotesfile.close() + notice("ADDED QUOTE #" + str(len(lines) - 1)) + + def quote(): + + def help(): + notice("SYNTAX: !quote [int] OR !quote search QUERY") + notice("QUERY may be a boolean grouping of quoted or unquoted " + + "search terms, examples:") + notice("!quote search foo") + notice("!quote search foo AND (bar OR NOT baz)") + notice("!quote search \"foo\\\"bar\" AND ('NOT\"' AND \"'foo'\"" + + " OR 'bar\\'baz')") + + if "" == argument: + tokens = [] + else: + tokens = argument.split(" ") + if (len(tokens) > 1 and tokens[0] != "search") or \ + (len(tokens) == 1 and + (tokens[0] == "search" or not tokens[0].isdigit())): + help() + return + if not os.access(quotesfile_name, os.F_OK): + notice("NO QUOTES AVAILABLE") + return + quotesfile = open(quotesfile_name, "r") + lines = quotesfile.readlines() + quotesfile.close() + lines = lines[1:] + if len(tokens) == 1: + i = int(tokens[0]) + if i == 0 or i > len(lines): + notice("THERE'S NO QUOTE OF THAT INDEX") + return + i = i - 1 + elif len(tokens) > 1: + query = str.join(" ", tokens[1:]) + try: + results = plomsearch.search(query, lines) + except plomsearch.LogicParserError as err: + notice("FAILED QUERY PARSING: " + str(err)) + return + if len(results) == 0: + notice("NO QUOTES MATCHING QUERY") + else: + for result in results: + notice("QUOTE #" + str(result[0] + 1) + " : " + result[1]) + return + else: + i = random.randrange(len(lines)) + notice("QUOTE #" + str(i + 1) + ": " + lines[i]) + + def markov(): + from random import choice, shuffle + select_length = 2 + selections = [] + + def markov(snippet): + usable_selections = [] + for i in range(select_length, 0, -1): + for selection in selections: + add = True + for j in range(i): + j += 1 + if snippet[-j] != selection[-(j+1)]: + add = False + break + if add: + usable_selections += [selection] + if [] != usable_selections: + break + if [] == usable_selections: + usable_selections = selections + selection = choice(usable_selections) + return selection[select_length] + + hash_string = hashlib.md5(target.encode("utf-8")).hexdigest() + markovfeed_name = session.dbdir + "/markovfeed_" + hash_string + if not os.access(markovfeed_name, os.F_OK): + notice("NOT ENOUGH TEXT TO MARKOV.") + return + + # Lowercase incoming lines, ensure they end in a sentence end mark. + file = open(markovfeed_name, "r") + lines = file.readlines() + file.close() + tokens = [] + sentence_end_markers = ".!?)(" + for line in lines: + line = line.lower().replace("\n", "") + if line[-1] not in sentence_end_markers: + line += "." + tokens += line.split() + if len(tokens) <= select_length: + notice("NOT ENOUGH TEXT TO MARKOV.") + return + + # Replace URLs with escape string for now, so that the Markov selector + # won't see them as different strings. Stash replaced URLs in urls. + urls = [] + url_escape = "\nURL" + url_starts = ["http://", "https://", "") + 1 + except ValueError: + pass + urls += [tokens[i][:length]] + tokens[i] = url_escape + tokens[i][length:] + break + + # For each snippet of select_length, use markov() to find continuation + # token from selections. Replace present users' names with malkovich. + # Start snippets with the beginning of a sentence, if possible. + for i in range(len(tokens) - select_length): + token_list = [] + for j in range(select_length + 1): + token_list += [tokens[i + j]] + selections += [token_list] + snippet = [] + for i in range(select_length): + snippet += [""] + shuffle(selections) + for i in range(len(selections)): + if selections[i][0][-1] in sentence_end_markers: + for i in range(select_length): + snippet[i] = selections[i][i + 1] + break + msg = "" + malkovich = "malkovich" + while 1: + new_end = markov(snippet) + for name in session.users_in_chan: + if new_end[:len(name)] == name.lower(): + new_end = malkovich + new_end[len(name):] + break + if len(msg) + len(new_end) > 200: + break + msg += new_end + " " + for i in range(select_length - 1): + snippet[i] = snippet[i + 1] + snippet[select_length - 1] = new_end + + # Replace occurences of url escape string with random choice from urls. + while True: + index = msg.find(url_escape) + if index < 0: + break + msg = msg.replace(url_escape, choice(urls), 1) + + # More meaningful ways to randomly end sentences. + notice(msg + malkovich + ".") + + def twt(): + def try_open(mode): + try: + twtfile = open(session.twtfile, mode) + except (PermissionError, FileNotFoundError) as err: + notice("CAN'T ACCESS OR CREATE TWT FILE: " + str(err)) + return None + return twtfile + + from datetime import datetime + if not os.access(session.twtfile, os.F_OK): + twtfile = try_open("w") + if None == twtfile: + return + twtfile.close() + twtfile = try_open("a") + if None == twtfile: + return + twtfile.write(datetime.utcnow().isoformat() + "\t" + argument + "\n") + twtfile.close() + notice("WROTE TWT.") + + if "addquote" == command: + addquote() + elif "quote" == command: + quote() + elif "markov" == command: + markov() + elif "twt" == command: + twt() + + +def handle_url(url, notice, show_url=False): + + def mobile_twitter_hack(url): + re1 = 'https?://(mobile.twitter.com/)[^/]+(/status/)' + re2 = 'https?://mobile.twitter.com/([^/]+)/status/([^\?/]+)' + m = re.search(re1, url) + if m and m.group(1) == 'mobile.twitter.com/' \ + and m.group(2) == '/status/': + m = re.search(re2, url) + url = 'https://twitter.com/' + m.group(1) + '/status/' + m.group(2) + handle_url(url, notice, True) + return True + + try: + r = requests.get(url, timeout=15) + except (requests.exceptions.TooManyRedirects, + requests.exceptions.ConnectionError, + requests.exceptions.InvalidURL, + UnicodeError, + requests.exceptions.InvalidSchema) as error: + notice("TROUBLE FOLLOWING URL: " + str(error)) + return + if mobile_twitter_hack(url): + return + title = bs4.BeautifulSoup(r.text, "html5lib").title + if title and title.string: + prefix = "PAGE TITLE: " + if show_url: + prefix = "PAGE TITLE FOR <" + url + ">: " + notice(prefix + title.string.strip()) + else: + notice("PAGE HAS NO TITLE TAG") + + +class Session: + + def __init__(self, io, username, nickname, channel, twtfile, dbdir): + self.io = io + self.nickname = nickname + self.channel = channel + self.users_in_chan = [] + self.twtfile = twtfile + self.dbdir = dbdir + self.io.send_line("NICK " + self.nickname) + self.io.send_line("USER " + username + " 0 * : ") + self.io.send_line("JOIN " + self.channel) + hash_string = hashlib.md5(self.channel.encode("utf-8")).hexdigest() + self.logdir = self.dbdir + "/irclogs_" + hash_string + "/" + if not os.path.exists(self.logdir): + os.makedirs(self.logdir) + + def loop(self): + + def log(line): + now = datetime.datetime.utcnow() + logfile = open(self.logdir + now.strftime("%Y-%m-%d") + ".txt", "a") + form = "%Y-%m-%d %H:%M:%S UTC\t" + logfile.write(now.strftime(form) + " " + line + "\n") + logfile.close() + + def handle_privmsg(tokens): + + def handle_input(msg, target): + + def notice(msg): + self.io.send_line("NOTICE " + target + " :" + msg) + + matches = re.findall("(https?://[^\s>]+)", msg) + for i in range(len(matches)): + handle_url(matches[i], notice) + if "!" == msg[0]: + tokens = msg[1:].split() + argument = str.join(" ", tokens[1:]) + handle_command(tokens[0], argument, notice, target, self) + return + hash_string = hashlib.md5(target.encode("utf-8")).hexdigest() + markovfeed_name = self.dbdir + "/markovfeed_" + hash_string + file = open(markovfeed_name, "a") + file.write(msg + "\n") + file.close() + sender = "" for rune in tokens[0]: if rune == "!": @@ -111,9 +393,87 @@ while 1: if rune != ":": receiver += rune target = sender - if receiver != nickname: + if receiver != self.nickname: target = receiver msg = str.join(" ", tokens[3:])[1:] - url_check(msg) - if tokens[0] == "PING": - io.send_line("PONG " + tokens[1]) + if target == self.channel: + log("<" + sender + "> " + msg) + handle_input(msg, target) + + def name_from_join_or_part(tokens): + token = tokens[0][1:] + index_cut = token.find("@") + index_ex = token.find("!") + if index_ex > 0 and index_ex < index_cut: + index_cut = index_ex + return token[:index_cut] + + while True: + line = self.io.recv_line() + if not line: + continue + tokens = line.split(" ") + if len(tokens) > 1: + if tokens[0] == "PING": + self.io.send_line("PONG " + tokens[1]) + elif tokens[1] == "PRIVMSG": + handle_privmsg(tokens) + elif tokens[1] == "353": + names = tokens[5:] + names[0] = names[0][1:] + for i in range(len(names)): + names[i] = names[i].replace("@", "").replace("+", "") + self.users_in_chan += names + log(line) + elif tokens[1] == "JOIN": + name = name_from_join_or_part(tokens) + if name != self.nickname: + self.users_in_chan += [name] + log(line) + elif tokens[1] == "PART": + name = name_from_join_or_part(tokens) + del(self.users_in_chan[self.users_in_chan.index(name)]) + log(line) + else: + log(line) + + +def parse_command_line_arguments(): + parser = argparse.ArgumentParser() + parser.add_argument("-s, --server", action="store", dest="server", + default=SERVER, + help="server or server net to connect to (default: " + + SERVER + ")") + parser.add_argument("-p, --port", action="store", dest="port", type=int, + default=PORT, help="port to connect to (default : " + + str(PORT) + ")") + parser.add_argument("-w, --wait", action="store", dest="timeout", + type=int, default=TIMEOUT, + help="timeout in seconds after which to attempt " + + "reconnect (default: " + str(TIMEOUT) + ")") + parser.add_argument("-u, --username", action="store", dest="username", + default=USERNAME, help="username to use (default: " + + USERNAME + ")") + parser.add_argument("-n, --nickname", action="store", dest="nickname", + default=NICKNAME, help="nickname to use (default: " + + NICKNAME + ")") + parser.add_argument("-t, --twtxtfile", action="store", dest="twtfile", + default=TWTFILE, help="twtxt file to use (default: " + + TWTFILE + ")") + parser.add_argument("-d, --dbdir", action="store", dest="dbdir", + default=DBDIR, help="directory to store DB files in") + parser.add_argument("CHANNEL", action="store", help="channel to join") + opts, unknown = parser.parse_known_args() + return opts + + +opts = parse_command_line_arguments() +while True: + try: + io = IO(opts.server, opts.port, opts.timeout) + session = Session(io, opts.username, opts.nickname, opts.CHANNEL, + opts.twtfile, opts.dbdir) + session.loop() + except ExceptionForRestart: + io.socket.close() + continue