X-Git-Url: https://git.sommitrealweird.co.uk/rss2maildir.git/blobdiff_plain/9a52177bb38437b5b498704b81643461c0ce08c5..0c879280385e1ebc874403357f8f6ce4664e0b4e:/rss2maildir.py diff --git a/rss2maildir.py b/rss2maildir.py index 32a39f6..c9a2cee 100755 --- a/rss2maildir.py +++ b/rss2maildir.py @@ -252,25 +252,42 @@ class HTML2Text(HTMLParser): data = data + "\n".join(textwrap.wrap(self.currentparagraph, 70)) return data +def open_url(method, url): + redirectcount = 0 + while redirectcount < 3: + (type, rest) = urllib.splittype(url) + (host, path) = urllib.splithost(rest) + (host, port) = urllib.splitport(host) + if port == None: + port = 80 + try: + conn = httplib.HTTPConnection("%s:%s" %(host, port)) + conn.request(method, path) + response = conn.getresponse() + if response.status in [301, 302, 303, 307]: + headers = response.getheaders() + for header in headers: + if header[0] == "location": + url = header[1] + elif response.status == 200: + return response + except: + pass + redirectcount = redirectcount + 1 + return None + def parse_and_deliver(maildir, url, statedir): feedhandle = None headers = None # first check if we know about this feed already feeddb = dbm.open(os.path.join(statedir, "feeds"), "c") - # we need all the parts of the url - (type, rest) = urllib.splittype(url) - (host, path) = urllib.splithost(rest) - (host, port) = urllib.splitport(host) - if port == None: - port = 80 if feeddb.has_key(url): data = feeddb[url] data = cgi.parse_qs(data) - # now do a head on the feed to see if it's been updated - conn = httplib.HTTPConnection("%s:%s" %(host, port)) - conn.request("HEAD", path) - response = conn.getresponse() - headers = response.getheaders() + response = open_url("HEAD", url) + headers = None + if response: + headers = response.getheaders() ischanged = False try: for header in headers: @@ -289,24 +306,23 @@ def parse_and_deliver(maildir, url, statedir): except: ischanged = True if ischanged: - conn = httplib.HTTPConnection("%s:%s" %(host, port)) - conn.request("GET", path) - response = conn.getresponse() - headers = response.getheaders() - feedhandle = response + response = open_url("GET", url) + if response != None: + headers = response.getheaders() + feedhandle = response + else: + sys.stderr.write("Failed to fetch feed: %s\n" %(url)) + return else: return # don't need to do anything, nothings changed. else: - conn = httplib.HTTPConnection("%s:%s" %(host, port)) - conn.request("GET", path) - response = None - try: - response = conn.getresponse() - except: - print "Failed to fetch feed: %s" %(url) + response = open_url("GET", url) + if response != None: + headers = response.getheaders() + feedhandle = response + else: + sys.stderr.write("Failed to fetch feed: %s\n" %(url)) return - headers = response.getheaders() - feedhandle = response fp = feedparser.parse(feedhandle) db = dbm.open(os.path.join(statedir, "seen"), "c") @@ -323,6 +339,15 @@ def parse_and_deliver(maildir, url, statedir): prevmessageid = None + # check if there's a guid too - if that exists and we match the md5, + # return + if item.has_key("guid"): + if db.has_key(url + "|" + item["guid"]): + data = db[url + "|" + item["guid"]] + data = cgi.parse_qs(data) + if data["contentmd5"][0] == md5sum: + continue + if db.has_key(url + "|" + item["link"]): data = db[url + "|" + item["link"]] data = cgi.parse_qs(data) @@ -367,6 +392,9 @@ def parse_and_deliver(maildir, url, statedir): textparser = HTML2Text() textparser.feed(content.encode("utf-8")) textcontent = textparser.gettext() + textcontent = "%s\n\nItem URL: %s" %( \ + textcontent, \ + item["link"] ) textpart = MIMEText(textcontent.encode("utf-8"), "plain", "utf-8") msg.attach(textpart) msg.attach(htmlpart) @@ -393,12 +421,31 @@ def parse_and_deliver(maildir, url, statedir): # now add to the database about the item if prevmessageid: messageid = prevmessageid + " " + messageid - data = urllib.urlencode(( - ("message-id", messageid), \ - ("created", createddate), \ - ("contentmd5", md5sum) \ - )) - db[url + "|" + item["link"]] = data + if item.has_key("guid") and item["guid"] != item["link"]: + data = urllib.urlencode(( \ + ("message-id", messageid), \ + ("created", createddate), \ + ("contentmd5", md5sum) \ + )) + db[url + "|" + item["guid"]] = data + try: + data = db[url + "|" + item["link"]] + data = cgi.parse_qs(data) + newdata = urllib.urlencode(( \ + ("message-id", messageid), \ + ("created", data["created"][0]), \ + ("contentmd5", data["contentmd5"][0]) \ + )) + db[url + "|" + item["link"]] = newdata + except: + db[url + "|" + item["link"]] = data + else: + data = urllib.urlencode(( \ + ("message-id", messageid), \ + ("created", createddate), \ + ("contentmd5", md5sum) \ + )) + db[url + "|" + item["link"]] = data if headers: data = []