- webpage = urllib.request.urlopen(url, timeout=15)
- except (urllib.error.HTTPError, urllib.error.URLError,
- UnicodeError) as error:
- print("TROUBLE FOLLOWING URL: " + str(error))
- continue
- charset = webpage.info().get_content_charset()
- if not charset:
- charset = "utf-8"
- content_type = webpage.info().get_content_type()
- if content_type not in ('text/html', 'text/xml',
- 'application/xhtml+xml'):
- print("TROUBLE INTERPRETING URL: bad content type "
- + content_type)
- continue
- content = webpage.read().decode(charset)
- title = str(content).split('<title>')[1].split('</title>')[0]
- title = html.unescape(title)
- io.send_line("PRIVMSG " + target + " :page title for url: "
- + title)
+ r = requests.get(url, timeout=15)
+ except (requests.exceptions.TooManyRedirects,
+ requests.exceptions.ConnectionError,
+ requests.exceptions.InvalidURL,
+ requests.exceptions.InvalidSchema) as error:
+ notice("TROUBLE FOLLOWING URL: " + str(error))
+ return
+ if mobile_twitter_hack(url):
+ return
+ title = bs4.BeautifulSoup(r.text).title
+ if title:
+ prefix = "PAGE TITLE: "
+ if show_url:
+ prefix = "PAGE TITLE FOR <" + url + ">: "
+ notice(prefix + title.string.strip())
+ else:
+ notice("PAGE HAS NO TITLE TAG")
+
+ matches = re.findall("(https?://[^\s>]+)", msg)
+ for i in range(len(matches)):
+ handle_url(matches[i])
+
+ def command_check(msg):
+ if msg[0] != "!":
+ return
+ tokens = msg[1:].split()
+ hash_string = hashlib.md5(target.encode("utf-8")).hexdigest()
+ quotesfile_name = "quotes_" + hash_string
+ if tokens[0] == "addquote":
+ if not os.access(quotesfile_name, os.F_OK):
+ quotesfile = open(quotesfile_name, "w")
+ quotesfile.write("QUOTES FOR " + target + ":\n")
+ quotesfile.close()
+ quotesfile = open(quotesfile_name, "a")
+ quotesfile.write(str.join(" ", tokens[1:]) + "\n")
+ quotesfile.close()
+ quotesfile = open(quotesfile_name, "r")
+ lines = quotesfile.readlines()
+ quotesfile.close()
+ notice("ADDED QUOTE #" + str(len(lines) - 1))
+ elif tokens[0] == "quote":
+ if (len(tokens) > 2 and tokens[1] != "search") or \
+ (len(tokens) == 2 and
+ (tokens[1] == "search" or not tokens[1].isdigit())):
+ notice("SYNTAX: !quote [int] OR !quote search QUERY")
+ notice("QUERY may be a boolean grouping of quoted or "\
+ + "unquoted search terms, examples:")
+ notice("!quote search foo")
+ notice("!quote search foo AND (bar OR NOT baz)")
+ notice("!quote search \"foo\\\"bar\" AND "\
+ + "('NOT\"' AND \"'foo'\" OR 'bar\\'baz')")
+ return
+ if not os.access(quotesfile_name, os.F_OK):
+ notice("NO QUOTES AVAILABLE")
+ return
+ quotesfile = open(quotesfile_name, "r")
+ lines = quotesfile.readlines()
+ quotesfile.close()
+ lines = lines[1:]
+ if len(tokens) == 2:
+ i = int(tokens[1])
+ if i == 0 or i > len(lines):
+ notice("THERE'S NO QUOTE OF THAT INDEX")
+ return
+ i = i - 1
+ elif len(tokens) > 2:
+ query = str.join(" ", tokens[2:])
+ try:
+ results = plomsearch.search(query, lines)
+ except plomsearch.LogicParserError as err:
+ notice("FAILED QUERY PARSING: " + str(err))
+ return
+ if len(results) == 0:
+ notice("NO QUOTES MATCHING QUERY")
+ else:
+ for result in results:
+ notice("QUOTE #" + str(result[0] + 1) + " : "
+ + result[1])
+ return
+ else:
+ i = random.randrange(len(lines))
+ notice("QUOTE #" + str(i + 1) + ": " + lines[i])