-import urllib
-import BeautifulSoup
+import urllib, re
+import BeautifulSoup, json
import lib, htcache
soup = BeautifulSoup.BeautifulSoup
-class imgstream(lib.imgstream):
- def __init__(self, url):
- self.bk = urllib.urlopen(url)
- self.ctype = self.bk.info()["Content-Type"]
- self.clen = int(self.bk.info()["Content-Length"])
-
- def fileno(self):
- return self.bk.fileno()
-
- def close(self):
- self.bk.close()
-
- def read(self, sz = None):
- if sz is None:
- return self.bk.read()
- else:
- return self.bk.read(sz)
-
class page(lib.page):
def __init__(self, chapter, stack, n, url):
self.stack = stack
return self.ciurl
def open(self):
- return imgstream(self.iurl())
+ return lib.stdimgstream(self.iurl())
def __str__(self):
return self.name
return el
class manga(lib.manga):
+ cure = re.compile(r"/c[\d.]+/$")
+
def __init__(self, lib, id, name, url):
self.lib = lib
self.id = id
except KeyError:
pass
url = n.a["href"].encode("us-ascii")
- if url[-7:] != "/1.html":
+ if url[-7:] == "/1.html":
+ url = url[:-6]
+ elif self.cure.search(url) is not None:
+ pass
+ else:
raise Exception("parse error: unexpected chapter URL for %r: %s" % (self, url))
- vol.ch.append(chapter(vol, vol.stack + [(vol, o)], chid, name, url[:-6]))
+ vol.ch.append(chapter(vol, vol.stack + [(vol, o)], chid, name, url))
cvol.append(vol)
self.cvol = cvol
return self.cvol
ls = self.alphapage(pno)
i = 0
+ def search(self, expr):
+ resp = urllib.urlopen(self.base + ("ajax/search.php?term=%s" % urllib.quote(expr)))
+ try:
+ rc = json.load(resp)
+ finally:
+ resp.close()
+ return [manga(self, id.encode("utf8"), name, self.base + ("manga/%s/" % id.encode("utf8"))) for num, name, id, genres, author in rc]
+
def byid(self, id):
url = self.base + ("manga/%s/" % id)
page = soup(htcache.fetch(url))