# coding=utf-8 from lxml import etree import sqlite3 import re import requests import random import time import logging user_agent = [ "Mozilla/5.0 (Macintosh; U; Intel Mac OS X 10_6_8; en-us) AppleWebKit/534.50 (KHTML, like Gecko) Version/5.1 " "Safari/534.50", "Mozilla/5.0 (Windows; U; Windows NT 6.1; en-us) AppleWebKit/534.50 (KHTML, like Gecko) Version/5.1 " "Safari/534.50", "Mozilla/5.0 (Windows NT 10.0; WOW64; rv:38.0) Gecko/20100101 Firefox/38.0", "Mozilla/5.0 (Windows NT 10.0; WOW64; Trident/7.0; .NET4.0C; .NET4.0E; .NET CLR 2.0.50727; .NET CLR " "3.0.30729; .NET CLR 3.5.30729; InfoPath.3; rv:11.0) like Gecko", "Mozilla/5.0 (compatible; MSIE 9.0; Windows NT 6.1; Trident/5.0)", "Mozilla/4.0 (compatible; MSIE 8.0; Windows NT 6.0; Trident/4.0)", "Mozilla/4.0 (compatible; MSIE 7.0; Windows NT 6.0)", "Mozilla/4.0 (compatible; MSIE 6.0; Windows NT 5.1)", "Mozilla/5.0 (Macintosh; Intel Mac OS X 10.6; rv:2.0.1) Gecko/20100101 Firefox/4.0.1", "Mozilla/5.0 (Windows NT 6.1; rv:2.0.1) Gecko/20100101 Firefox/4.0.1", "Opera/9.80 (Macintosh; Intel Mac OS X 10.6.8; U; en) Presto/2.8.131 Version/11.11", "Opera/9.80 (Windows NT 6.1; U; en) Presto/2.8.131 Version/11.11", "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_7_0) AppleWebKit/535.11 (KHTML, like Gecko) Chrome/17.0.963.56 " "Safari/535.11", "Mozilla/4.0 (compatible; MSIE 7.0; Windows NT 5.1; Maxthon 2.0)", "Mozilla/4.0 (compatible; MSIE 7.0; Windows NT 5.1; TencentTraveler 4.0)", "Mozilla/4.0 (compatible; MSIE 7.0; Windows NT 5.1)", "Mozilla/4.0 (compatible; MSIE 7.0; Windows NT 5.1; The World)", "Mozilla/4.0 (compatible; MSIE 7.0; Windows NT 5.1; Trident/4.0; SE 2.X MetaSr 1.0; SE 2.X MetaSr 1.0; .NET " "CLR 2.0.50727; SE 2.X MetaSr 1.0)", "Mozilla/4.0 (compatible; MSIE 7.0; Windows NT 5.1; 360SE)", "Mozilla/4.0 (compatible; MSIE 7.0; Windows NT 5.1; Avant Browser)", "Mozilla/4.0 (compatible; MSIE 7.0; Windows NT 5.1)", "Mozilla/5.0 (iPhone; U; CPU iPhone OS 4_3_3 like Mac OS X; en-us) AppleWebKit/533.17.9 (KHTML, like Gecko) " "Version/5.0.2 Mobile/8J2 Safari/6533.18.5", "Mozilla/5.0 (iPod; U; CPU iPhone OS 4_3_3 like Mac OS X; en-us) AppleWebKit/533.17.9 (KHTML, like Gecko) " "Version/5.0.2 Mobile/8J2 Safari/6533.18.5", "Mozilla/5.0 (iPad; U; CPU OS 4_3_3 like Mac OS X; en-us) AppleWebKit/533.17.9 (KHTML, like Gecko) " "Version/5.0.2 Mobile/8J2 Safari/6533.18.5", "Mozilla/5.0 (Linux; U; Android 2.3.7; en-us; Nexus One Build/FRF91) AppleWebKit/533.1 (KHTML, like Gecko) " "Version/4.0 Mobile Safari/533.1", "MQQBrowser/26 Mozilla/5.0 (Linux; U; Android 2.3.7; zh-cn; MB200 Build/GRJ22; CyanogenMod-7) " "AppleWebKit/533.1 (KHTML, like Gecko) Version/4.0 Mobile Safari/533.1", "Opera/9.80 (Android 2.3.4; Linux; Opera Mobi/build-1107180945; U; en-GB) Presto/2.8.149 Version/11.10", "Mozilla/5.0 (Linux; U; Android 3.0; en-us; Xoom Build/HRI39) AppleWebKit/534.13 (KHTML, like Gecko) " "Version/4.0 Safari/534.13", "Mozilla/5.0 (BlackBerry; U; BlackBerry 9800; en) AppleWebKit/534.1+ (KHTML, like Gecko) Version/6.0.0.337 " "Mobile Safari/534.1+", "Mozilla/5.0 (hp-tablet; Linux; hpwOS/3.0.0; U; en-US) AppleWebKit/534.6 (KHTML, like Gecko) " "wOSBrowser/233.70 Safari/534.6 TouchPad/1.0", "Mozilla/5.0 (SymbianOS/9.4; Series60/5.0 NokiaN97-1/20.0.019; Profile/MIDP-2.1 Configuration/CLDC-1.1) " "AppleWebKit/525 (KHTML, like Gecko) BrowserNG/7.1.18124", "Mozilla/5.0 (compatible; MSIE 9.0; Windows Phone OS 7.5; Trident/5.0; IEMobile/9.0; HTC; Titan)", "UCWEB7.0.2.37/28/999", "NOKIA5700/ UCWEB7.0.2.37/28/999", "Openwave/ UCWEB7.0.2.37/28/999", "Mozilla/4.0 (compatible; MSIE 6.0; ) Opera/UCWEB7.0.2.37/28/999", # iPhone 6: "Mozilla/6.0 (iPhone; CPU iPhone OS 8_0 like Mac OS X) AppleWebKit/536.26 (KHTML, like Gecko) Version/8.0 " "Mobile/10A5376e Safari/8536.25", ] def get_user_agent(): headers = {"User-Agent": random.choice(user_agent)} return headers class Scraper: database: str tag: str page: int def __init__(self): self.database = "book.db" self.tag = "" self.page = 0 self.pattern_number = re.compile(r"\d+\.?\d*") logging.basicConfig(filename="scraper.log", level=logging.ERROR) def get_current_progress(self) -> (): conn = sqlite3.connect(self.database) results = conn.execute("SELECT tag, page from progress where id = '0'") for row in results: return row[0], row[1] return "", 0 def save_current_progress(self, current_tag, current_page): conn = sqlite3.connect(self.database) conn.execute( "UPDATE progress set tag = '{}', page = {} where id = '0'".format( current_tag, current_page ) ) conn.commit() conn.close() def start_grab(self) -> bool: self.create_tables() scraper.grab_tag() current_tag, current_page = self.get_current_progress() tags = self.get_tag_list() for i in range(0, len(tags)): no = 0 if i == 0 and current_tag == tags[i]: no = current_page while self.grab_book_list(tags[i], no): no = no + 20 return True def create_tables(self): conn = sqlite3.connect(self.database) try: conn.execute("CREATE TABLE tags (tag TEXT PRIMARY KEY)") conn.commit() except sqlite3.Error as e: logging.error(str(e)) conn.rollback() try: conn.execute( "CREATE TABLE book (" "id TEXT PRIMARY KEY, title TEXT, author TEXT, " "publisher TEXT, original_title TEXT, " "translator TEXT, pub_year TEXT, pages INTEGER, " "price INTEGER, currency_unit TEXT, binding TEXT, " "isbn TEXT, author_intro TEXT, book_intro text, " "content TEXT, tags TEXT, picture BLOB)" ) conn.commit() except sqlite3.Error as e: logging.error(str(e)) conn.rollback() try: conn.execute( "CREATE TABLE progress (id TEXT PRIMARY KEY, tag TEXT, page integer )" ) conn.execute("INSERT INTO progress values('0', '', 0)") conn.commit() except sqlite3.Error as e: logging.error(str(e)) conn.rollback() def grab_tag(self): url = "https://book.douban.com/tag/?view=cloud" r = requests.get(url, headers=get_user_agent()) r.encoding = "utf-8" h: etree.ElementBase = etree.HTML(r.text) tags: [] = h.xpath( '/html/body/div[@id="wrapper"]/div[@id="content"]' '/div[@class="grid-16-8 clearfix"]/div[@class="article"]' '/div[@class=""]/div[@class="indent tag_cloud"]' "/table/tbody/tr/td/a/@href" ) conn = sqlite3.connect(self.database) c = conn.cursor() try: for tag in tags: t: str = tag.strip("/tag") c.execute("INSERT INTO tags VALUES ('{}')".format(t)) c.close() conn.commit() conn.close() except sqlite3.Error as e: logging.error(str(e)) conn.rollback() return False return True def grab_book_list(self, tag="小说", pageno=1) -> bool: logging.info("start to grab tag {} page {}...".format(tag, pageno)) self.save_current_progress(tag, pageno) url = "https://book.douban.com/tag/{}?start={}&type=T".format(tag, pageno) r = requests.get(url, headers=get_user_agent()) r.encoding = "utf-8" h: etree.Element = etree.HTML(r.text) li_list: [] = h.xpath( '/html/body/div[@id="wrapper"]/div[@id="content"]' '/div[@class="grid-16-8 clearfix"]' '/div[@class="article"]/div[@id="subject_list"]' '/ul/li/div[@class="info"]/h2/a/@href' ) next_page = h.xpath( '/html/body/div[@id="wrapper"]/div[@id="content"]' '/div[@class="grid-16-8 clearfix"]' '/div[@class="article"]/div[@id="subject_list"]' '/div[@class="paginator"]/span[@class="next"]/a[@href]' ) has_next = True if len(next_page) == 0: has_next = False if len(li_list) == 0: return False for li in li_list: li.strip("") book_id = li.strip("/").split("/")[-1] try: delay = float(random.randint(0, 200)) / 100.0 time.sleep(delay) self.crow_book_info(book_id) except BaseException as e: logging.error( logging.error("error when scrape {}, {}".format(book_id, str(e))) ) return has_next def get_tag_list(self) -> [str]: ret = [] conn = sqlite3.connect(self.database) results = conn.execute( "SELECT tags.tag from tags join progress where tags.tag >= progress.tag" ) for row in results: ret.append(row[0]) return ret def crow_book_info(self, book_id) -> bool: conn = sqlite3.connect(self.database) for _ in conn.execute("SELECT id from book where id = ('{}')".format(book_id)): return url = "https://book.douban.com/subject/{}/".format(book_id) r = requests.get(url, headers=get_user_agent()) r.encoding = "utf-8" h: etree.Element = etree.HTML(r.text) e_text = h.xpath('/html/body/div[@id="wrapper"]/h1/span/text()') if len(e_text) == 0: return False title = e_text[0] elements = h.xpath( '/html/body/div[@id="wrapper"]' '/div[@id="content"]/div[@class="grid-16-8 clearfix"]' '/div[@class="article"]' ) if len(elements) == 0: return False e_article = elements[0] book_intro = "" author_intro = "" content = "" tags = "" e_book_intro = e_article.xpath( 'div[@class="related_info"]' '/div[@class="indent"][@id="link-report"]/*' '/div[@class="intro"]/*/text()' ) for line in e_book_intro: line = line.strip() if line != "": book_intro = book_intro + line + "\n" e_author_intro = e_article.xpath( 'div[@class="related_info"]' '/div[@class="indent "]/*' '/div[@class="intro"]/*/text()' ) for line in e_author_intro: line = line.strip() if line != "": author_intro = author_intro + line + "\n" e_content = e_article.xpath( 'div[@class="related_info"]' '/div[@class="indent"][@id="dir_' + book_id + '_full"]/text()' ) for line in e_content: line = line.strip() if line != "": content = content + line + "\n" e_tags = e_article.xpath( 'div[@class="related_info"]/' 'div[@id="db-tags-section"]/' 'div[@class="indent"]/span/a/text()' ) for line in e_tags: line = line.strip() if line != "": tags = tags + line + "\n" e_subject = e_article.xpath( 'div[@class="indent"]' '/div[@class="subjectwrap clearfix"]' '/div[@class="subject clearfix"]' ) pic_href = e_subject[0].xpath('div[@id="mainpic"]/a/@href') picture = None if len(pic_href) > 0: res = requests.get(pic_href[0], headers=get_user_agent()) picture = res.content info_children = e_subject[0].xpath('div[@id="info"]/child::node()') e_array = [] e_dict = dict() for e in info_children: if isinstance(e, etree._ElementUnicodeResult): e_dict["text"] = e elif isinstance(e, etree._Element): if e.tag == "br": e_array.append(e_dict) e_dict = dict() else: e_dict[e.tag] = e book_info = dict() for d in e_array: label = "" span = d.get("span") a_label = span.xpath("span/text()") if len(a_label) > 0 and label == "": label = a_label[0].strip() a_label = span.xpath("text()") if len(a_label) > 0 and label == "": label = a_label[0].strip() label = label.strip(":") text = d.get("text").strip() e_a = d.get("a") text.strip() text.strip(":") if label == "作者" or label == "译者": a = span.xpath("a/text()") if text == "" and len(a) == 1: text = a[0].strip() if text == "" and e_a is not None: text_a = e_a.xpath("text()") if len(text_a) > 0: text = text_a[0].strip() text = re.sub(r"\s+", " ", text) if text != "": book_info[label] = text sql = ( "INSERT INTO book(" "id, title, author, " "publisher, original_title, translator, " "pub_year, pages, price, " "currency_unit, binding, isbn, " "author_intro, book_intro, content, " "tags, picture)" "VALUES(" "?, ?, ?, " "?, ?, ?, " "?, ?, ?, " "?, ?, ?, " "?, ?, ?, " "?, ?)" ) unit = None price = None pages = None conn = sqlite3.connect(self.database) try: s_price = book_info.get("定价") if s_price is None: # price cannot be NULL logging.error( "error when scrape book_id {}, cannot retrieve price...", book_id ) return None else: e = re.findall(self.pattern_number, s_price) if len(e) != 0: number = e[0] unit = s_price.replace(number, "").strip() price = int(float(number) * 100) s_pages = book_info.get("页数") if s_pages is not None: # pages can be NULL e = re.findall(self.pattern_number, s_pages) if len(e) != 0: pages = int(e[0]) conn.execute( sql, ( book_id, title, book_info.get("作者"), book_info.get("出版社"), book_info.get("原作名"), book_info.get("译者"), book_info.get("出版年"), pages, price, unit, book_info.get("装帧"), book_info.get("ISBN"), author_intro, book_intro, content, tags, picture, ), ) conn.commit() except sqlite3.Error as e: logging(str(e)) conn.rollback() except TypeError as e: logging.error("error when scrape {}, {}".format(book_id, str(e))) conn.rollback() return False conn.close() return True if __name__ == "__main__": scraper = Scraper() scraper.start_grab()