スクレイピングは茨の道/Scraping is a thorny road

6094c91088af5e8ea5094aa750d13570?s=47 cottondesu
December 17, 2018

スクレイピングは茨の道/Scraping is a thorny road

前回LT「Python + Selenium + Beautiful Soup でスクレイピング」の
1.おさらい
2.失敗から学ぶ
3.ソースの修正
4.結論

6094c91088af5e8ea5094aa750d13570?s=128

cottondesu

December 17, 2018
Tweet

Transcript

  1. 13.

    class main(): options = Options() # ChromeͷύεʢStableνϟωϧͰ--headless͕࢖͑ΔΑ͏ʹͳͬͨΒෆཁͳ͸ͣʣ options.binary_location = '/Applications/Google

    Chrome.app/Contents/MacOS/Google Chrome' # ϔουϨεϞʔυΛ༗ޮʹ͢Δʢ࣍ͷߦΛίϝϯτΞ΢τ͢Δͱը໘͕දࣔ͞ΕΔʣɻ options.add_argument('--headless') # ChromeͷWebDriverΦϒδΣΫτΛ࡞੒͢Δɻ driver = webdriver.Chrome( os.environ["CHROMEDRIVER"], chrome_options=options) # ϩάΠϯαΠτURL driver.get(os.environ["URL"]) # ϩάΠϯID driver.find_element_by_xpath( "//div[@class='sub-content']/dl/dd/div/input").send_keys( os.environ["LOGIN_USER_ID"]) # ϩάΠϯύεϫʔυ username = driver.find_element_by_xpath( "//div[@class='sub-content']/dl/dd[2]/div/input").send_keys( os.environ["LOGIN_USER_PASS"]) # ϩάΠϯϘλϯԡԼ driver.find_element_by_name("login").click() # ޱ࠲؅ཧʹભҠ driver.find_element_by_xpath(“//div[@id='link']/ul/li[10]").click() html = driver.page_source soup = BeautifulSoup(html, 'html.parser')
  2. 14.

    # อ༗ޱ਺ɺऔಘ୯Ձɺऔಘ୯ՁɺධՁଛӹͷऔಘ item_name = ['อ༗ޱ਺ɹɹ |', 'औಘ୯Ձɹɹ |', 'ج४Ձֹɹɹ |',

    'ධՁଛӹɹɹ |'] # ౤ࢿ৴ୗ໊ͷऔಘ investmentname = [] investments = soup.find_all('td', class_='mbody', colspan="3") for investment in investments: investmentname.append(investment.a.text) inves_num = 0 row_num = 0 various_values = soup.find_all('tr', bgcolor='#eaf4e8', align="right") for various_value in various_values: # ౤ࢿ৴ୗ໊ͷදࣔ print(investmentname[inves_num]) # อ༗ޱ਺ͷදࣔ print(item_name[row_num], various_value.td.text) for other in various_value.td.find_next_siblings("td"): row_num = row_num + 1 # อ༗ޱ਺Ҏ֎ͷදࣔ print(item_name[row_num], other.text) row_num = 0 inves_num = inves_num + 1 # શͯͷΫοΩʔΛ࡟আ driver.delete_all_cookies() driver.quit() if __name__ == "__main__": main()
  3. 15.
  4. 17.
  5. 20.

    class main(): try: options = Options() # ChromeͷύεʢStableνϟωϧͰ--headless͕࢖͑ΔΑ͏ʹͳͬͨΒෆཁͳ͸ͣʣ options.binary_location =

    '/Applications/Google Chrome.app/Contents/MacOS/Google Chrome' # ϔουϨεϞʔυΛ༗ޮʹ͢Δʢ࣍ͷߦΛίϝϯτΞ΢τ͢Δͱը໘͕දࣔ͞ΕΔʣɻ options.add_argument('--headless') # ChromeͷWebDriverΦϒδΣΫτΛ࡞੒͢Δɻ driver = webdriver.Chrome( os.environ["CHROMEDRIVER"], chrome_options=options) # ϩάΠϯαΠτURL driver.get(os.environ["URL"]) # ϩάΠϯID driver.find_element_by_xpath( "//div[@class='sub-content']/dl/dd/div/input").send_keys( os.environ["LOGIN_USER_ID"]) # ϩάΠϯύεϫʔυ username = driver.find_element_by_xpath( "//div[@class='sub-content']/dl/dd[2]/div/input").send_keys( os.environ["LOGIN_USER_PASS"]) # ϩάΠϯϘλϯԡԼ driver.find_element_by_name("login").click() # ޱ࠲؅ཧʹભҠ driver.find_element_by_xpath(“//div[@id='link']/ul/li[10]").click() html = driver.page_source soup = BeautifulSoup(html, 'html.parser') मਖ਼Օॴ͸੺࿮
  6. 21.

    # อ༗ޱ਺ɺऔಘ୯Ձɺऔಘ୯ՁɺධՁଛӹͷऔಘ item_name = ['อ༗ޱ਺ɹɹ |', 'औಘ୯Ձɹɹ |', 'ج४Ձֹɹɹ |',

    'ධՁଛӹɹɹ |'] # ౤ࢿ৴ୗ໊ͷऔಘ investmentname = [] investments = soup.find_all('td', class_='mbody', colspan="3") for investment in investments: investmentname.append(investment.a.text) inves_num = 0 row_num = 0 various_values = soup.find_all('tr', bgcolor='#eaf4e8', align="right") for various_value in various_values: # ౤ࢿ৴ୗ໊ͷදࣔ print(investmentname[inves_num]) # อ༗ޱ਺ͷදࣔ print(item_name[row_num], various_value.td.text) for other in various_value.td.find_next_siblings("td"): row_num = row_num + 1 # อ༗ޱ਺Ҏ֎ͷදࣔ print(item_name[row_num], other.text) row_num = 0 inves_num = inves_num + 1 except NoSuchElementException as e: print("seleniumͷૢ࡞தʹΤϥʔ͕ൃੜ͠·ͨ͠ɻ") traceback.print_exc() finally: # શͯͷΫοΩʔΛ࡟আ driver.delete_all_cookies() driver.quit() if __name__ == "__main__": main() मਖ਼Օॴ͸੺࿮
  7. 22.