X-Git-Url: https://plomlompom.com/repos/%7Broute%7D?a=blobdiff_plain;f=plomlombot.py;h=c32b6dd3c0c42b79f86bb00c96d6f8b7d9017fd2;hb=7bbd08d498b289f2f3cad8b85c8582aa6ceffc04;hp=81be31fdb5f438349469e7d63afac9b4b20b1a7e;hpb=3b3ac1e18357e60525a6be9b5665c504e0ec4e6c;p=plomlombot-irc.git diff --git a/plomlombot.py b/plomlombot.py index 81be31f..c32b6dd 100755 --- a/plomlombot.py +++ b/plomlombot.py @@ -6,9 +6,12 @@ import datetime import select import time import re -import urllib.request -import http.client -import html +import requests +import bs4 +import random +import hashlib +import os +import plomsearch # Defaults, may be overwritten by command line arguments. SERVER = "irc.freenode.net" @@ -67,7 +70,11 @@ class IO: self._pingtest(send_ping) return None self.last_pong = time.time() - received_runes = self.socket.recv(1024).decode("UTF-8") + received_bytes = self.socket.recv(1024) + try: + received_runes = received_bytes.decode("UTF-8") + except UnicodeDecodeError: + received_runes = received_bytes.decode("latin1") if len(received_runes) == 0: print("SOCKET CONNECTION BROKEN") raise ExceptionForRestart @@ -99,36 +106,109 @@ def lineparser_loop(io, nickname): def act_on_privmsg(tokens): + def notice(msg): + io.send_line("NOTICE " + target + " :" + msg) + def url_check(msg): - def notice(msg): - io.send_line("NOTICE " + target + " :" + msg) + def handle_url(url, show_url=False): + + def mobile_twitter_hack(url): + re1 = 'https?://(mobile.twitter.com/)[^/]+(/status/)' + re2 = 'https?://mobile.twitter.com/([^/]+)/status/' \ + + '([^\?/]+)' + m = re.search(re1, url) + if m and m.group(1) == 'mobile.twitter.com/' \ + and m.group(2) == '/status/': + m = re.search(re2, url) + url = 'https://twitter.com/' + m.group(1) + '/status/' \ + + m.group(2) + handle_url(url, True) + return True - matches = re.findall("(https?://[^\s>]+)", msg) - for i in range(len(matches)): - url = matches[i] - request = urllib.request.Request(url, headers={ - "User-Agent": "plomlombot" - }) try: - webpage = urllib.request.urlopen(request, timeout=15) - except (urllib.error.HTTPError, urllib.error.URLError, - UnicodeError, http.client.BadStatusLine) as error: + r = requests.get(url, timeout=15) + except (requests.exceptions.TooManyRedirects, + requests.exceptions.ConnectionError, + requests.exceptions.InvalidURL, + requests.exceptions.InvalidSchema) as error: notice("TROUBLE FOLLOWING URL: " + str(error)) - continue - charset = webpage.info().get_content_charset() - if not charset: - charset = "utf-8" - content_type = webpage.info().get_content_type() - if content_type not in ('text/html', 'text/xml', - 'application/xhtml+xml'): - notice("TROUBLE READING PAGE TITLE: bad content type " - + content_type) - continue - content = webpage.read().decode(charset) - title = str(content).split('