|
import aiohttp |
|
import asyncio |
|
import re |
|
|
|
import pandas as pd |
|
|
|
from pathlib import Path |
|
from aiolimiter import AsyncLimiter |
|
from typing import Dict, List |
|
from bs4 import BeautifulSoup |
|
from bs4.element import Tag |
|
|
|
LIST_COMICS_500_URL = ( |
|
"https://www.explainxkcd.com/wiki/index.php/List_of_all_comics_(1-500)" |
|
) |
|
LIST_COMICS_FULL_URL = ( |
|
"https://www.explainxkcd.com/wiki/index.php/List_of_all_comics_(full)" |
|
) |
|
|
|
|
|
def walk_tag(initial_tag: Tag, end_tag_name: str) -> str: |
|
""" |
|
Walk the HTML tree and aggregates all text between an |
|
initial tag and an end tag. |
|
|
|
Parameters |
|
---------- |
|
initial_tag: BeautifulSoup |
|
end_tag_name: str |
|
|
|
Returns |
|
------- |
|
aggregated_text: str |
|
""" |
|
result = [] |
|
current_tag = initial_tag |
|
|
|
|
|
while True: |
|
if current_tag.name in ["p", "dl"]: |
|
result.append(current_tag.get_text(separator=" ", strip=True)) |
|
elif current_tag.name == end_tag_name: |
|
|
|
break |
|
current_tag = current_tag.next_sibling |
|
return "\n".join(result) |
|
|
|
|
|
async def parse_url_html( |
|
session: aiohttp.ClientSession, url: str, throttler: AsyncLimiter |
|
) -> BeautifulSoup: |
|
""" |
|
Parse the HTML content of a given URL. |
|
The request is sent asynchronously and using a provided request throttler. |
|
If the request fails, we retry up to 5 times. |
|
|
|
Parameters |
|
---------- |
|
session: aiohttp.ClientSession |
|
url: str |
|
throttler: AsyncLimiter |
|
|
|
Returns |
|
------- |
|
BeautifulSoup |
|
""" |
|
for _ in range(5): |
|
try: |
|
|
|
async with throttler: |
|
async with session.get(url, raise_for_status=True) as resp: |
|
html = await resp.text() |
|
return BeautifulSoup(html, "html.parser") |
|
|
|
except aiohttp.ClientError: |
|
continue |
|
|
|
|
|
async def scrap_comic( |
|
session: aiohttp.ClientSession, explained_url: str, throttler: AsyncLimiter |
|
) -> Dict[str, str]: |
|
""" |
|
Try to scrap all information for a given XKCD comic using its `explainxkcd.com` URL |
|
|
|
Parameters |
|
---------- |
|
session: aiohttp.ClientSession |
|
explained_url: str |
|
throttler: AsyncLimiter |
|
|
|
Returns |
|
------- |
|
Dict[str, str] |
|
""" |
|
soup = await parse_url_html(session, explained_url, throttler) |
|
|
|
|
|
title_splits = soup.find("h1").text.split(":") |
|
if len(title_splits) > 1: |
|
id = title_splits[0] |
|
title = "".join(title_splits[1:]).strip() |
|
else: |
|
id = None |
|
title = "".join(title_splits).strip() |
|
|
|
|
|
explanation_soup = soup.find("span", {"id": "Explanation"}) |
|
try: |
|
explanation = walk_tag(explanation_soup.parent, "span") |
|
except: |
|
explanation = None |
|
|
|
|
|
transcript_soup = soup.find("span", {"id": "Transcript"}) |
|
try: |
|
transcript = walk_tag(transcript_soup.parent, "span") |
|
except: |
|
transcript = None |
|
|
|
xkcd_url = f"https://www.xkcd.com/{id}" |
|
xkcd_soup = await parse_url_html(session, xkcd_url, throttler) |
|
|
|
|
|
try: |
|
image = xkcd_soup.find("div", {"id": "comic"}).img |
|
if title in image: |
|
image_title = image["title"] |
|
else: |
|
image_title = image["alt"] |
|
except: |
|
image_title = None |
|
|
|
|
|
try: |
|
image_url = xkcd_soup.find(text=re.compile("https://imgs.xkcd.com")) |
|
except: |
|
image_url = None |
|
|
|
return dict( |
|
id=id, |
|
title=title, |
|
image_title=image_title, |
|
url=xkcd_url, |
|
image_url=image_url, |
|
explained_url=explained_url, |
|
transcript=transcript, |
|
explanation=explanation, |
|
) |
|
|
|
|
|
async def scap_comic_urls( |
|
session: aiohttp.ClientSession, comic_list_url: str |
|
) -> List[str]: |
|
""" |
|
Scrap all XKCD comic URLs from the `explainxkcd.com` website. |
|
|
|
Parameters |
|
---------- |
|
session: aiohttp.ClientSession |
|
comic_list_url: str |
|
|
|
Returns |
|
------- |
|
urls: List[str] |
|
""" |
|
async with session.get(comic_list_url) as resp: |
|
html = await resp.text() |
|
soup = BeautifulSoup(html, "html.parser") |
|
|
|
|
|
create_spans = soup.find_all("span", {"class": "create"}) |
|
return [ |
|
"https://www.explainxkcd.com" + span.parent.a["href"] for span in create_spans |
|
] |
|
|
|
|
|
async def main(): |
|
""" |
|
Scrap XKCD dataset |
|
""" |
|
|
|
throttler = AsyncLimiter(max_rate=10, time_period=1) |
|
async with aiohttp.ClientSession() as session: |
|
|
|
comic_urls = await scap_comic_urls( |
|
session, LIST_COMICS_500_URL |
|
) + await scap_comic_urls(session, LIST_COMICS_FULL_URL) |
|
|
|
|
|
data = await asyncio.gather( |
|
*[scrap_comic(session, url, throttler) for url in comic_urls] |
|
) |
|
|
|
df = ( |
|
pd.DataFrame.from_records(data) |
|
.dropna(subset=["id"]) |
|
.astype({"id": "int32"}) |
|
.sort_values("id") |
|
) |
|
df.to_json(Path(__file__).parent / "dataset.jsonl", orient="records", lines=True) |
|
|
|
|
|
if __name__ == "__main__": |
|
asyncio.run(main()) |
|
|