Use DnD API

This commit is contained in:
Romain Quinet 2023-10-07 08:32:07 +02:00
parent c47ff3d9ed
commit bf3fd878ac

View File

@ -1,6 +1,7 @@
import requests import requests
import os import os
from bs4 import BeautifulSoup from bs4 import BeautifulSoup
import json
folder_path = "transcripts" folder_path = "transcripts"
@ -8,21 +9,22 @@ if not os.path.exists(folder_path):
os.makedirs(folder_path) os.makedirs(folder_path)
for i in range(1, 139): for i in range(1, 139):
url = f"https://darknetdiaries.com/transcript/{i}" try:
r = requests.get(url) url = f"https://darknetdiaries.com/transcript/{i}"
soup = BeautifulSoup(r.text, 'html.parser') r = requests.get(url)
soup = BeautifulSoup(r.text, 'html.parser')
transcript = soup.find('pre').get_text() transcript = soup.find('pre').get_text()
title_section = soup.find('h1').get_text()
url = f"https://darknetdiaries.com/episode/{i}" url = f"https://api.darknetdiaries.com/{i}.json"
r = requests.get(url) r = requests.get(url)
soup = BeautifulSoup(r.text, 'html.parser') parsed_json = json.loads(r.text)
downloads = soup.find(id='downloads').get_text() title = parsed_json["episode_name"]
number = parsed_json["episode_number"]
downloads = parsed_json["total_downloads"]
ep, title = title_section.split(":", 1) with open(f"{folder_path}/episode_{number}.txt", "w") as f:
ep = ep.strip() f.write(f"{title}\n{downloads}\n{transcript}")
title = title.strip() print(f"{number} {title}")
with open(f"{folder_path}/episode_{i}.txt", "w") as f: except Exception:
f.write(f"{title}\n{downloads}\n{transcript}") print(f"Failed scraping episode {i}")
print(f"{ep} {title}")