scan-feeds: Import feeds into the same file.
This allows me to accumulate feeds over time in case someone's blog is down at any given moment.
This commit is contained in:
parent
11dec64ca0
commit
ab6966baf1
|
@ -9,7 +9,11 @@ import dateutil.tz as tz
|
||||||
with open('bloggers.yml') as f:
|
with open('bloggers.yml') as f:
|
||||||
users = yaml.safe_load(f.read())
|
users = yaml.safe_load(f.read())
|
||||||
|
|
||||||
log = {}
|
try:
|
||||||
|
with open('out/report.yml') as f:
|
||||||
|
log = yaml.safe_load(f.read())
|
||||||
|
except IOError:
|
||||||
|
log = {}
|
||||||
|
|
||||||
START = datetime.datetime(2009, 12, 21, 6)
|
START = datetime.datetime(2009, 12, 21, 6)
|
||||||
|
|
||||||
|
@ -37,19 +41,17 @@ def parse_feeds(weeks, uri):
|
||||||
|
|
||||||
while len(weeks) <= wn:
|
while len(weeks) <= wn:
|
||||||
weeks.append([])
|
weeks.append([])
|
||||||
weeks[wn].append(dict(
|
|
||||||
date=date,
|
post = dict(date=date,
|
||||||
title=post.title,
|
title=post.title,
|
||||||
url=get_link(post)))
|
url=get_link(post))
|
||||||
|
if post['url'] not in [p['url'] for p in weeks[wn]]:
|
||||||
|
weeks[wn].append(post)
|
||||||
|
|
||||||
for (username, u) in users.items():
|
for (username, u) in users.items():
|
||||||
weeks = []
|
weeks = log.setdefault(username, [])
|
||||||
print "[%s]" % (username)
|
|
||||||
for l in u['links']:
|
for l in u['links']:
|
||||||
parse_feeds(weeks, l[2])
|
parse_feeds(weeks, l[2])
|
||||||
log[username] = weeks
|
|
||||||
for (i, w) in enumerate(weeks):
|
|
||||||
print " [%d]: %s" % (i, w)
|
|
||||||
|
|
||||||
with open('out/report.yml', 'w') as f:
|
with open('out/report.yml', 'w') as f:
|
||||||
yaml.safe_dump(log, f)
|
yaml.safe_dump(log, f)
|
||||||
|
|
Loading…
Reference in New Issue