from selenium import webdriver
from selenium.webdriver.common.by import By
from selenium.webdriver.support.ui import WebDriverWait
from selenium.webdriver.support import expected_conditions as EC
from selenium.common.exceptions import TimeoutException
from bs4 import BeautifulSoup
import pandas as pd
import time
chrome_options = webdriver.ChromeOptions()
chrome_options.add_argument("--disable-extensions")
chrome_options.add_argument("--disable-gpu")
chrome_options.add_argument("--no-sandbox")
chrome_options.add_argument("--headless")
chrome_options.add_argument("--disable-dev-shm-usage")
driver = webdriver.Chrome(options=chrome_options)
driver.set_page_load_timeout(30)
driver.get("https://www.nomfoundation.org/nom-project/Ho-Xuan-Huong/Ho-Xuan-Huong-of-poems?uiLang=vn")
data = {"Poem_Title": [], "Line_Number": [], "Nom_Text": [], "Translation": []}
num_pages = 4
for page in range(num_pages):
WebDriverWait(driver, 10).until(EC.presence_of_element_located((By.CLASS_NAME, "pagination")))
soup = BeautifulSoup(driver.page_source, 'html.parser')
for poem_link in soup.select('table a'):
title = poem_link.text.split(" - ")[0].strip()
link = "https://www.nomfoundation.org" + poem_link['href']
try:
driver.get(link)
except TimeoutException:
print(f"Timeout khi truy cập: {link}")
continue
poem_soup = BeautifulSoup(driver.page_source, 'html.parser')
nom_text_element = poem_soup.select_one('td.hnTextBodyGray')
translation_text_element = poem_soup.select_one('td.alt2')
if nom_text_element and translation_text_element:
nom_lines = nom_text_element.decode_contents().replace('
', '
').split('
')
nom_lines = [BeautifulSoup(line, 'html.parser').get_text(strip=True) for line in nom_lines if line.strip()]
translation_lines = translation_text_element.decode_contents().replace('
', '
').split('
')
translation_lines = [BeautifulSoup(line, 'html.parser').get_text(strip=True) for line in translation_lines if line.strip()]
for line_number, (nom_line, translation_line) in enumerate(zip(nom_lines, translation_lines), start=1):
nom_line = BeautifulSoup(nom_line, 'html.parser').get_text(strip=True)
data["Poem_Title"].append(title)
data["Line_Number"].append(line_number)
data["Nom_Text"].append(nom_line)
data["Translation"].append(translation_line)
print(f"Đã ghi vào: {title} - Dòng {line_number}")
driver.back()
time.sleep(1)
pagination = driver.find_element(By.CLASS_NAME, "pagination")
page_links = pagination.find_elements(By.TAG_NAME, "a")
if page + 1 < len(page_links):
next_page_button = page_links[page + 1]
next_page_button.click()
time.sleep(2)
else:
print("Không có trang tiếp theo.")
break
driver.quit()
df = pd.DataFrame(data)
df_cleaned = df.dropna()
df_cleaned.to_csv("ho_xuan_huong_poems_full.csv", index=False, encoding="utf-8-sig")
print("Dữ liệu đã được lưu thành công vào file ho_xuan_huong_poems_full.csv!")