|
|
|
import json |
|
from pathlib import Path |
|
import pandas as pd |
|
import pyarrow as pa |
|
import pyarrow.parquet as pq |
|
|
|
def iter_commit_rows(snapshot_dir: Path, diff_dir: Path, repo: str): |
|
|
|
for snapshot_path in sorted(snapshot_dir.glob('pr-*.json')): |
|
pr = json.loads(snapshot_path.read_text(encoding='utf-8')) |
|
pr_number = pr.get('number') |
|
if pr_number is None: |
|
continue |
|
|
|
commits = pr.get('commits', {}).get('nodes', []) |
|
for node in commits: |
|
commit = node.get('commit', {}) |
|
oid = commit.get('oid') |
|
if oid: |
|
diff_file = diff_dir / f"{oid}.diff" |
|
commit['diff'] = diff_file.read_text(encoding='utf-8').strip() if diff_file.exists() else '' |
|
|
|
events = [] |
|
|
|
for c in pr.get('comments', {}).get('nodes', []): |
|
ts = c.get('createdAt') |
|
if ts: events.append(('comment', ts, c)) |
|
for rt in pr.get('reviewThreads', {}).get('nodes', []): |
|
for r in rt.get('comments', {}).get('nodes', []): |
|
ts = r.get('createdAt') |
|
if ts: events.append(('review', ts, r)) |
|
for node in commits: |
|
ts = node.get('commit', {}).get('committedDate') |
|
if ts: events.append(('commit', ts, node)) |
|
events.sort(key=lambda e: e[1]) |
|
|
|
history = [] |
|
for kind, ts, data in events: |
|
history.append((kind, ts, data)) |
|
if kind != 'commit': |
|
continue |
|
|
|
c = data['commit'] |
|
oid = c.get('oid') |
|
diff_text = c.get('diff', '') |
|
msg = c.get('message', '') |
|
if not oid or not diff_text.strip(): |
|
continue |
|
|
|
|
|
prompt_parts = [ |
|
f"Title: {pr.get('title', '')}", |
|
f"Body: {pr.get('body', '')}", |
|
] |
|
labels = pr.get('labels') or [] |
|
if labels: |
|
prompt_parts.append("Labels: " + ", ".join(labels)) |
|
|
|
|
|
last_idx = next((i for i in range(len(history)-2, -1, -1) if history[i][0] == 'commit'), None) |
|
seg = history[last_idx+1:-1] if last_idx is not None else history[:-1] |
|
if last_idx is not None: |
|
prev = history[last_idx][2]['commit'] |
|
prompt_parts.append( |
|
f"Last commit: {prev.get('message')}\nDiff:\n{prev.get('diff', '')}" |
|
) |
|
|
|
for ekind, _, edata in seg: |
|
if ekind == 'comment': |
|
body = edata.get('body', '').strip() |
|
if body: |
|
prompt_parts.append(f"Comment: {body}") |
|
elif ekind == 'review': |
|
path = edata.get('path', '') |
|
review_body = edata.get('body', '').strip() |
|
hunk = (edata.get('diffHunk') or '').strip() |
|
prompt_parts.append( |
|
f"Review on {path}: {review_body}\nDiff:\n{hunk}" |
|
) |
|
|
|
author = c.get('author', {}) or {} |
|
yield { |
|
'prompt': '\n'.join(prompt_parts), |
|
'completion': f"Diff:\n{diff_text}", |
|
'repo': repo, |
|
'pr_number': pr_number, |
|
'title': pr.get('title', ''), |
|
'body': pr.get('body', ''), |
|
'created_at': pr.get('createdAt', ''), |
|
'closed_at': pr.get('closedAt', ''), |
|
'merged_at': pr.get('mergedAt', ''), |
|
'author': author.get('login', ''), |
|
'state': pr.get('state', ''), |
|
'additions': pr.get('additions', 0), |
|
'deletions': pr.get('deletions', 0), |
|
'changed_files': pr.get('changedFiles', 0), |
|
'head_ref': pr.get('headRefName', ''), |
|
'labels': ", ".join(labels), |
|
'completion_commit': oid, |
|
} |
|
|
|
def main(): |
|
BASE_DIR = Path(__file__).resolve().parent |
|
snapshot_dir = BASE_DIR.parent / 'data' / 'raw-data' / 'prs' |
|
diff_dir = BASE_DIR.parent / 'data' / 'raw-data' / 'diffs' |
|
dataset_dir = BASE_DIR.parent / 'data' / 'dataset' |
|
dataset_dir.mkdir(parents=True, exist_ok=True) |
|
|
|
|
|
schema = pa.schema([ |
|
('prompt', pa.string()), |
|
('completion', pa.string()), |
|
('repo', pa.string()), |
|
('pr_number', pa.int64()), |
|
('title', pa.string()), |
|
('body', pa.string()), |
|
('created_at', pa.string()), |
|
('closed_at', pa.string()), |
|
('merged_at', pa.string()), |
|
('author', pa.string()), |
|
('state', pa.string()), |
|
('additions', pa.int64()), |
|
('deletions', pa.int64()), |
|
('changed_files', pa.int64()), |
|
('head_ref', pa.string()), |
|
('labels', pa.string()), |
|
('completion_commit', pa.string()), |
|
]) |
|
|
|
train_writer = pq.ParquetWriter(str(dataset_dir / 'train.parquet'), schema) |
|
test_writer = pq.ParquetWriter(str(dataset_dir / 'test.parquet'), schema) |
|
|
|
for row in iter_commit_rows(snapshot_dir, diff_dir, 'dotnet/runtime'): |
|
table = pa.Table.from_pydict({k: [v] for k, v in row.items()}, schema) |
|
|
|
if hash(row['completion_commit']) % 5 == 0: |
|
test_writer.write_table(table) |
|
else: |
|
train_writer.write_table(table) |
|
|
|
train_writer.close() |
|
test_writer.close() |
|
|
|
print(f"Wrote train.parquet and test.parquet to {dataset_dir}") |
|
|
|
if __name__ == '__main__': |
|
main() |