more scrapers
[disclosr.git] / documents / scrapers / d72744fb1e5d6e87af9a5ea16cc27fa5.py
blob:a/documents/scrapers/d72744fb1e5d6e87af9a5ea16cc27fa5.py -> blob:b/documents/scrapers/d72744fb1e5d6e87af9a5ea16cc27fa5.py
--- a/documents/scrapers/d72744fb1e5d6e87af9a5ea16cc27fa5.py
+++ b/documents/scrapers/d72744fb1e5d6e87af9a5ea16cc27fa5.py
@@ -17,7 +17,7 @@
         (url, mime_type, content) = scrape.fetchURL(scrape.docsdb,
              self.getURL(), "foidocuments", self.getAgencyID())
 
-        d = pq(content.read())
+        d = pq(content)
         d.make_links_absolute(base_url = self.getURL())
         for item in d('.item-list').items():
             title= item('h3').text()
@@ -30,7 +30,6 @@
             doc = foidocsdb.get(dochash)
             if doc is None:
                 print "saving " + dochash
-                edate = date.today().strftime("%Y-%m-%d")
                 doc = {'_id': dochash, 'agencyID': self.getAgencyID()
                 , 'url': self.getURL(), 'docID': dochash,
                 "links": links,