2023-10-06 21:35:53 +02:00
|
|
|
import requests
|
2023-10-07 00:30:46 +02:00
|
|
|
import os
|
2023-10-06 21:35:53 +02:00
|
|
|
from bs4 import BeautifulSoup
|
|
|
|
|
2023-10-07 00:30:46 +02:00
|
|
|
folder_path = "transcripts"
|
|
|
|
|
|
|
|
if not os.path.exists(folder_path):
|
|
|
|
os.makedirs(folder_path)
|
|
|
|
|
2023-10-06 21:35:53 +02:00
|
|
|
for i in range(1, 139):
|
|
|
|
url = f"https://darknetdiaries.com/transcript/{i}"
|
|
|
|
r = requests.get(url)
|
|
|
|
soup = BeautifulSoup(r.text, 'html.parser')
|
|
|
|
pre_section = soup.find('pre')
|
2023-10-06 22:43:28 +02:00
|
|
|
title_section = soup.find('h1')
|
2023-10-06 21:35:53 +02:00
|
|
|
|
|
|
|
if pre_section:
|
2023-10-07 00:30:46 +02:00
|
|
|
transcript = pre_section.get_text()
|
|
|
|
ep, title = title_section.get_text().split(":", 1)
|
|
|
|
ep = ep.strip()
|
|
|
|
title = title.strip()
|
|
|
|
with open(f"{folder_path}/episode_{i}.txt", "w") as f:
|
|
|
|
f.write(f"{title}\n{transcript}")
|
|
|
|
print(f"{ep} {title}")
|