1
0
Fork 0

scrape-djtracklists

master
Adrien Abraham 2 years ago
parent 559646bc11
commit 415af8de98

@ -1,7 +1,9 @@
#!/usr/bin/env python3 #!/usr/bin/env python3
"""Download tracklists from djtracklists.com.""" """Download tracklists from djtracklists.com as CSV."""
import argparse import argparse
import csv
import time
from dataclasses import dataclass from dataclasses import dataclass
from pathlib import Path from pathlib import Path
from typing import Optional from typing import Optional
@ -12,31 +14,71 @@ from bs4 import BeautifulSoup
@dataclass @dataclass
class Track: class Track:
"""One track, parsed from a tracklist page."""
title: str title: str
artists: list[str] artists: list[str]
mix: Optional[str] mix: Optional[str]
mix_artists: Optional[list[str]] mix_artists: Optional[list[str]]
timestamp: str timestamp: str
def format_artists(self):
return " & ".join(self.artists)
def format_mix_artists(self):
return " & ".join(self.mix_artists) if self.mix_artists else ""
def main(): def main():
parser = argparse.ArgumentParser() parser = argparse.ArgumentParser()
parser.add_argument("-s", "--series", help="download this series") parser.add_argument("-s", "--series",
parser.add_argument("-t", "--tracklist", help="download this tracklist") help="download this series (provide URL of 1st page)")
parser.add_argument("-t", "--tracklist",
help="download this tracklist (provide URL)")
parser.add_argument("--pretty", help="pretty print a CSV file.")
args = parser.parse_args() args = parser.parse_args()
if tracklist_url := args.tracklist: if csv_file_name := args.pretty:
tracklist = download_tracklist(tracklist_url) pretty_print_csv(csv_file_name)
name = tracklist_url.rstrip("/").rsplit("/", maxsplit=1)[-1] elif series_url := args.series:
file_name = Path.cwd() / (name + ".txt") download_series(series_url)
save_tracklist(tracklist, file_name) elif tracklist_url := args.tracklist:
download_tracklist(tracklist_url)
def download_series(series_url: str):
while series_url:
print("Processing series URL", series_url)
response = requests.get(series_url, timeout=10)
response.raise_for_status()
soup = BeautifulSoup(response.text, "html.parser")
# Download all tracklists.
for tracklist_link in soup.find_all("a", class_="mix"):
tracklist_url = tracklist_link["href"]
print("Processing tracklist URL", tracklist_url)
download_tracklist(tracklist_url)
time.sleep(1) # throttle
# Look for the next page button.
for page_link in soup.find_all("a", class_="pagenumber"):
if "Next" in page_link.string:
series_url = page_link["href"]
break
else:
series_url = None
def download_tracklist(url: str):
tracklist = get_tracklist_from_url(url)
name = url.rstrip("/").rsplit("/", maxsplit=1)[-1]
file_name = Path.cwd() / (name + ".csv")
save_tracklist_as_csv(tracklist, file_name)
def is_track_row(css_class: str) -> bool: def is_track_row(css_class: str) -> bool:
return css_class in ("on", "off") return css_class in ("on", "off")
def download_tracklist(url: str) -> list[Track]: def get_tracklist_from_url(url: str) -> list[Track]:
"""Get tracklist from the Web and parse it into a list of Track objects."""
response = requests.get(url, timeout=10) response = requests.get(url, timeout=10)
response.raise_for_status() response.raise_for_status()
soup = BeautifulSoup(response.text, "html.parser") soup = BeautifulSoup(response.text, "html.parser")
@ -44,13 +86,22 @@ def download_tracklist(url: str) -> list[Track]:
for row in soup.find_all("div", class_=is_track_row): for row in soup.find_all("div", class_=is_track_row):
artists = [] artists = []
mix_artists = [] mix_artists = []
try: try:
title = row.find("a", class_="track").string title = row.find("a", class_="track").string
mix = row.find("a", class_="release").string mix = row.find("a", class_="release").string
except AttributeError: except AttributeError:
title = row.find("b").string try:
title = row.find("b").string
except AttributeError:
title = "(unknown title)"
mix = None mix = None
timestamp = row.find("span", class_="index_time").string
try:
timestamp = row.find("span", class_="index_time").string
except AttributeError:
timestamp = "(unknown timestamp)"
for artist in row.find_all("a", class_="artist"): for artist in row.find_all("a", class_="artist"):
prev_tag = artist.previous_sibling.string prev_tag = artist.previous_sibling.string
if getattr(prev_tag, "string", "").strip() == "remixed by": if getattr(prev_tag, "string", "").strip() == "remixed by":
@ -69,21 +120,49 @@ def download_tracklist(url: str) -> list[Track]:
return tracklist return tracklist
def save_tracklist(tracklist: list[Track], file_name: Path): def save_tracklist_as_csv(tracklist: list[Track], file_name: Path):
try: try:
with open(file_name, "wt", encoding="utf8") as file: with open(file_name, "wt", encoding="utf8", newline="") as file:
writer = csv.writer(file)
writer.writerow(
["Timestamp", "Artists", "Title", "Mix", "Remix artists"]
)
for track in tracklist: for track in tracklist:
artists = " & ".join(track.artists) writer.writerow([
line = f"{track.timestamp}{artists}{track.title}" track.timestamp,
if track.mix: track.format_artists(),
line += f" ({track.mix})" track.title,
if track.mix_artists: track.mix or "",
mix_artists = " & ".join(track.mix_artists) track.format_mix_artists(),
line += f" remixed by {mix_artists}" ])
file.write(line + "\n")
except OSError as exc: except OSError as exc:
print(f"Can't save tracklist: {exc}") print(f"Can't save tracklist: {exc}")
def pretty_print_csv(csv_file_name: str):
try:
with open(csv_file_name, "rt", encoding="utf8", newline="") as file:
reader = csv.reader(file)
first_line_skipped = False
for row in reader:
if not first_line_skipped:
first_line_skipped = True
continue
ts, artists, title, mix, remix_artists = row
if ":" in ts:
ts_min, ts_sec = ts.split(":")
ts_min = int(ts_min)
ts_h, ts_min = ts_min // 60, ts_min % 60
ts = f"{ts_h:02}:{ts_min:02}:{ts_sec}"
print(f"{ts} {artists or '(unknown)'}{title}", end="")
if mix:
print(f" ({mix})", end="")
if remix_artists:
print(f" by {remix_artists}", end="")
print()
except OSError as exc:
print(f"Can't read CSV: {exc}")
if __name__ == "__main__": if __name__ == "__main__":
main() main()

Loading…
Cancel
Save