如何解决python beautifulsoup4突然TypeError:'NoneType'对象不可调用但它不是None
我检查了很多问题/答案,但我可以找到我的解决方案。 这段代码以前很完美,但两天前,它坚持抛出如下错误:
Traceback (most recent call last):
File "D:\Python\experiments\20210213.py",line 85,in <module>
ws['B1'] = get_titles[loop_excel].contents[0].strip()
TypeError: 'NoneType' object is not callable
1,我用:
soup = BeautifulSoup(html_text,'html.parser')
2,然后获取标题:
get_titles = soup.find_all('li',class_='chapter')
3、循环插入标题到excel中:
wb = load_workbook('20210209.xlsx')
time.sleep(5)
ws = wb['Sheet1']
loop_excel = 0
for get_download in get_downloads_soup:
ws.insert_rows(1,amount=1)
ws['A1'] = datetime.now(pytz.timezone('Asia/Hong_Kong')).strftime("%Y-%m-%d %H:%M:%S")
ws['B1'] = get_titles_s[loop_excel].contents[0].strip()
loop_excel = loop_excel + 1
我尝试使用打印来识别 get_titles/contents 等的类型和长度,它有效,
<class 'bs4.element.ResultSet'> 66
所以,我真的不知道为什么会弹出这个错误?
整个事情看起来像:
options = Options()
options.binary_location = r'C:\Program Files\Mozilla Firefox\firefox.exe'
driver = webdriver.Firefox(executable_path=r'C:\Users\louis\geckodriver.exe')
time.sleep(14)
# continue
articles_total_number = 333
page_number_start = 1
loop_download_first_page = 0
downloaded_number = 0
urlwhole = (f'blahblah')
driver.get(urlwhole)
WebDriverWait(driver,100).until(EC.visibility_of_element_located((By.CLASS_NAME,'chapters')))
time.sleep(30)
# supplementary first run
for page_number in range(page_number_start,page_number_start + 1):
#preparation
time.sleep(5)
get_downloads_selenium =
driver.find_elements_by_css_selector('input.downloads') #200 resultset of selenium FirefoxWebElements
time.sleep(5)
html_text = driver.page_source
time.sleep(15)
soup = BeautifulSoup(html_text,'html.parser')
time.sleep(15)
get_titles = soup.find_all('li',class_='chapters')
time.sleep(15)
get_chapternumbers = soup.find_all('a',class_="srch-result") # 200 resultset of bs element
print(type(get_chapternumbers),len(get_chapternumbers))
get_downloads_soup = soup.find_all('input',class_='downloads')# 200 resultset of bs element
print(type(get_downloads_soup),len(get_downloads_soup))
time.sleep(25)
# record info by excel
wb = load_workbook('20210209.xlsx')
time.sleep(5)
ws = wb['Sheet1']
loop_excel = 0
for get_download in get_downloads_soup:
ws.insert_rows(1,amount=1)
ws['A1'] = datetime.now(pytz.timezone('Asia/Hong_Kong')).strftime("%Y-%m-%d %H:%M:%S")
ws['B1'] = get_titles[loop_excel].contents[0].strip()
ws['C1'] = get_chapternumbers[loop_excel].get_text()
ws['D1'] = page_number
loop_excel = loop_excel + 1
wb.save('20210209.xlsx')
wb.close()
print(f'page {page_number} excel recorded!')
我也对record-to-excel过程做了一些测试,我相信get_titles[]有问题,标题是:
<li class="chapters">Safety analysis and other purposes <a href="blahblah/00000651">View details</a></li>
因为 ws['A1'] 运行良好,而 ws['C1'] 也很好。所以请关注这条线。
get_titles[loop_excel].contents[0].strip()
解决方法
我终于确定了它:
<li class="chapters">
<i>In</i>
<i>vitro</i>
blahblah1
<i>in</i>
<i>vitro</i>
blahblah2
<a href="//0008373">View details</a>
</li>
这就是 get_titles[40] 返回的内容。这就是问题所在。 content[0] 是第一个标签:In,这是一个 Nonetype?为什么这是一个 Nonetype?
版权声明:本文内容由互联网用户自发贡献,该文观点与技术仅代表作者本人。本站仅提供信息存储空间服务,不拥有所有权,不承担相关法律责任。如发现本站有涉嫌侵权/违法违规的内容, 请发送邮件至 dio@foxmail.com 举报,一经查实,本站将立刻删除。