|
import html |
|
import multiprocessing |
|
import pathlib |
|
from typing import Callable |
|
|
|
import msgspec |
|
import typer |
|
import zstandard |
|
from loguru import logger |
|
|
|
from RedditModels import ( |
|
RedditAuthor, |
|
RedditComment, |
|
RedditFlair, |
|
RedditSubmit, |
|
RedditSubreddit, |
|
) |
|
|
|
root_app = typer.Typer() |
|
|
|
|
|
def read_and_decode( |
|
reader, chunk_size, max_window_size, previous_chunk=None, bytes_read=0 |
|
): |
|
chunk = reader.read(chunk_size) |
|
bytes_read += chunk_size |
|
if previous_chunk is not None: |
|
chunk = previous_chunk + chunk |
|
try: |
|
return chunk.decode() |
|
except UnicodeDecodeError: |
|
if bytes_read > max_window_size: |
|
raise UnicodeError( |
|
f"Unable to decode frame after reading {bytes_read:,} bytes" |
|
) |
|
logger.debug(f"Decoding error with {bytes_read:,} bytes, reading another chunk") |
|
return read_and_decode(reader, chunk_size, max_window_size, chunk, bytes_read) |
|
|
|
|
|
GB = 2**30 |
|
|
|
|
|
def read_lines_jsonl(file_name, chunk_size=GB // 2): |
|
with open(file_name, "rb") as file_handle: |
|
buffer = b"" |
|
while True: |
|
chunk = file_handle.read(chunk_size) |
|
|
|
if not chunk: |
|
break |
|
lines = (buffer + chunk).split(b"\n") |
|
|
|
for line in lines[:-1]: |
|
yield line.strip(), file_handle.tell() |
|
|
|
buffer = lines[-1] |
|
|
|
|
|
def read_lines_zst(file_name, scale: float = 1): |
|
with open(file_name, "rb") as file_handle: |
|
buffer = "" |
|
reader = zstandard.ZstdDecompressor( |
|
max_window_size=int((2**31) * scale) |
|
).stream_reader(file_handle) |
|
while True: |
|
chunk = read_and_decode( |
|
reader, int((2**27) * scale), int((2**29) * 2 * scale) |
|
) |
|
|
|
if not chunk: |
|
break |
|
lines = (buffer + chunk).split("\n") |
|
|
|
for line in lines[:-1]: |
|
yield line.strip(), file_handle.tell() |
|
|
|
buffer = lines[-1] |
|
|
|
reader.close() |
|
|
|
|
|
def error_cb(err): |
|
logger.exception(err) |
|
|
|
|
|
def get_submission_flags(data: dict): |
|
flag_map = { |
|
"!": "spoiler", |
|
"#": "stickied", |
|
">": "pinned", |
|
"A": "archived", |
|
"C": "is_crosspostable", |
|
"c": "is_original_content", |
|
"E": "edited", |
|
"e": "is_meta", |
|
"G": "can_gild", |
|
"H": "hidden", |
|
"i": "is_robot_indexable", |
|
"L": "allow_live_comments", |
|
"l": "locked", |
|
"m": "is_reddit_media_domain", |
|
"M": "over_18", |
|
"O": "contest_mode", |
|
"q": "quarantine", |
|
"s": "is_self", |
|
"v": "is_video", |
|
} |
|
return "".join(flag for flag, key in flag_map.items() if data.get(key)) |
|
|
|
|
|
def get_comment_flags(data: dict): |
|
flag_map = { |
|
"#": "stickied", |
|
"A": "archived", |
|
"E": "edited", |
|
"G": "can_gild", |
|
"H": "hidden", |
|
"l": "locked", |
|
"=": "score_hidden", |
|
"P": "author_premium", |
|
"R": "send_replies", |
|
"O": "can_mod_post", |
|
"N": "no_follow", |
|
} |
|
return "".join(flag for flag, key in flag_map.items() if data.get(key)) |
|
|
|
|
|
def get_reddit_flair(data: dict, prefix: str): |
|
return RedditFlair( |
|
bg=data.get(f"{prefix}_flair_background_color"), |
|
css_cls=data.get(f"{prefix}_flair_css_class"), |
|
template=data.get(f"{prefix}_flair_template_id"), |
|
richtext=data.get(f"{prefix}_flair_richtext"), |
|
text=data.get(f"{prefix}_flair_text"), |
|
text_color=data.get(f"{prefix}_flair_text_color"), |
|
type=data.get(f"{prefix}_flair_type"), |
|
) |
|
|
|
|
|
def make_submission(data: dict, file_id: pathlib.Path): |
|
|
|
author = None |
|
if data.get("author_created_utc"): |
|
author_flair = get_reddit_flair(data, "author") |
|
author_fullname = data.get("author_fullname", "") |
|
afn = ( |
|
author_fullname[3:] |
|
if author_fullname.startswith("t2_") |
|
else author_fullname |
|
) |
|
author = RedditAuthor( |
|
name=data.get("author", ""), |
|
uid=afn, |
|
create=data.get("author_created_utc", -1), |
|
flair=author_flair if author_flair.is_flaired else None, |
|
patreon=data.get("author_patreon_flair", False), |
|
premium=data.get("author_premium", False), |
|
) |
|
|
|
|
|
subreddit_id = data.get("subreddit_id", "") |
|
if subreddit_id is not None: |
|
subid = subreddit_id[3:] if subreddit_id.startswith("t5_") else subreddit_id |
|
else: |
|
return None |
|
subreddit = RedditSubreddit( |
|
name=data.get("subreddit", "<?>"), |
|
id=subid, |
|
|
|
subs=data.get("subreddit_subscribers", None), |
|
type=data.get("subreddit_type", None), |
|
) |
|
link_flair = get_reddit_flair(data, "link") |
|
|
|
submission = RedditSubmit( |
|
sub=subreddit, |
|
author=author, |
|
id=data.get("id", None), |
|
score=data.get("score", 0) if data.get("score", 0) else 0, |
|
created=float(data.get("created_utc", 0.0)) |
|
if data.get("created_utc", 0.0) |
|
else -1.0, |
|
title=data.get("title", None), |
|
flags=get_submission_flags(data), |
|
link_flair=link_flair if link_flair.is_flaired else None, |
|
url=data.get("url"), |
|
text=data.get("selftext", None), |
|
) |
|
if submission.text == "[removed]": |
|
submission.removed = [ |
|
data.get("removal_reason"), |
|
data.get("removed_by"), |
|
data.get("removed_by_category"), |
|
] |
|
|
|
if data.get("crosspost_parent_list", []): |
|
submission.cross = [] |
|
for crosspost in data.get("crosspost_parent_list", []): |
|
post = make_submission(crosspost, file_id) |
|
if post is None: |
|
continue |
|
submission.cross.append(post) |
|
return submission |
|
|
|
|
|
def make_comment(data: dict): |
|
author = data.get("author", "") |
|
if author is None or author.lower() == "[deleted]": |
|
author = None |
|
else: |
|
author_flair = get_reddit_flair(data, "author") |
|
author_fullname = data.get("author_fullname", "") |
|
afn = ( |
|
author_fullname[3:] |
|
if author_fullname.startswith("t2_") |
|
else author_fullname |
|
) |
|
author = RedditAuthor( |
|
name=data.get("author", ""), |
|
uid=afn, |
|
create=data.get("author_created_utc", -1), |
|
flair=author_flair if author_flair.is_flaired else None, |
|
patreon=data.get("author_patreon_flair", False), |
|
premium=data.get("author_premium", False), |
|
) |
|
subreddit_id = data.get("subreddit_id", "") |
|
if subreddit_id is not None: |
|
subid = subreddit_id[3:] if subreddit_id.startswith("t5_") else subreddit_id |
|
else: |
|
return None |
|
subreddit = RedditSubreddit( |
|
name=data.get("subreddit", ""), |
|
id=subid, |
|
subs=data.get("subreddit_subscribers", -1), |
|
type=data.get("subreddit_type", ""), |
|
) |
|
text = html.unescape(data.get("body", "")).replace("\r\n", "\n") |
|
if author is None and text in ["[deleted]", "[removed]"]: |
|
text = None |
|
submission = RedditComment( |
|
sub=subreddit, |
|
author=author, |
|
id=data.get("id", ""), |
|
score=data.get("score", 0), |
|
created=data.get("created_utc", 0), |
|
thread_id=data.get("link_id", ""), |
|
parent_id=data.get("parent_id", ""), |
|
text=text, |
|
flags=get_comment_flags(data), |
|
) |
|
|
|
return submission |
|
|
|
|
|
|
|
|
|
|
|
def process_zst( |
|
input_file: pathlib.Path, output_file: pathlib.Path, processor_fn: Callable |
|
): |
|
decoder = msgspec.json.Decoder() |
|
encoder = msgspec.json.Encoder(decimal_format="number") |
|
write = output_file.with_stem(f"{output_file.stem}_{input_file.stem}") |
|
with open(write, "wb") as f: |
|
for lineidx, line in enumerate(read_lines_zst(input_file)): |
|
data, idx = line |
|
|
|
try: |
|
post = processor_fn(decoder.decode(data)) |
|
if post: |
|
f.write(encoder.encode(post) + b"\n") |
|
except msgspec.DecodeError: |
|
logger.warning("Decode error detected. Continuing...") |
|
if lineidx % 100_000 == 0: |
|
logger.info(f"{write} {lineidx} processed") |
|
logger.info(f"{write} done") |
|
|
|
|
|
|
|
|
|
|
|
@root_app.command(name="submissions") |
|
def process_submissions(zst_input: pathlib.Path, output_prefix: pathlib.Path): |
|
s = sorted(list(zst_input.iterdir())) |
|
|
|
with multiprocessing.Pool(processes=32) as pooled: |
|
results = [] |
|
for file in s: |
|
results.append( |
|
pooled.apply_async( |
|
process_zst, |
|
args=(file, output_prefix, make_submission), |
|
error_callback=error_cb, |
|
) |
|
) |
|
[result.wait() for result in results] |
|
|
|
|
|
@root_app.command(name="comments") |
|
def process_comments(zst_input: pathlib.Path, output_prefix: pathlib.Path): |
|
s = sorted(list(zst_input.iterdir())) |
|
|
|
with multiprocessing.Pool(processes=32) as pooled: |
|
results = [] |
|
for file in s: |
|
results.append( |
|
pooled.apply_async( |
|
process_zst, |
|
args=(file, output_prefix, make_submission), |
|
error_callback=error_cb, |
|
) |
|
) |
|
[result.wait() for result in results] |
|
|
|
|
|
|
|
|
|
if __name__ == "__main__": |
|
root_app() |
|
|