|
# Getting Started with WavePulse Radio Transcripts Dataset |
|
|
|
This tutorial will help you get started with using the WavePulse Radio Transcripts dataset from Hugging Face. |
|
|
|
## Prerequisites |
|
|
|
Before starting, make sure you have the required packages installed: |
|
|
|
```bash |
|
pip install datasets |
|
pip install huggingface-hub |
|
``` |
|
|
|
## Basic Setup |
|
|
|
First, let's set up our environment with some helpful configurations: |
|
|
|
```python |
|
from datasets import load_dataset |
|
import huggingface_hub |
|
|
|
# Increase timeout for large downloads |
|
huggingface_hub.constants.HF_HUB_DOWNLOAD_TIMEOUT = 60 |
|
|
|
# Set up cache directory (optional) |
|
cache_dir = "wavepulse_dataset" |
|
``` |
|
|
|
## Loading Strategies |
|
|
|
### 1. Loading a Specific State (Recommended for Beginners) |
|
Instead of loading the entire dataset, start with one state: |
|
|
|
```python |
|
# Load data for just New York |
|
ny_dataset = load_dataset("nyu-dice-lab/wavepulse-radio-raw-transcripts", |
|
"NY", |
|
cache_dir=cache_dir) |
|
``` |
|
|
|
### 2. Streaming Mode (Memory Efficient) |
|
If you're working with limited RAM: |
|
|
|
```python |
|
# Stream the dataset |
|
stream_dataset = load_dataset("nyu-dice-lab/wavepulse-radio-raw-transcripts", |
|
streaming=True, |
|
cache_dir=cache_dir) |
|
|
|
# Access data in a streaming fashion |
|
for example in stream_dataset["train"].take(5): |
|
print(example["text"]) |
|
``` |
|
|
|
## Common Tasks |
|
|
|
### 1. Filtering by Date Range |
|
|
|
```python |
|
# Filter for August 2024 |
|
filtered_ds = dataset.filter( |
|
lambda x: "2024-08-01" <= x['datetime'] <= "2024-08-31" |
|
) |
|
``` |
|
|
|
### 2. Finding Specific Stations |
|
|
|
```python |
|
# Get unique stations |
|
stations = set(dataset["train"]["station"]) |
|
|
|
# Filter for a specific station |
|
station_ds = dataset.filter(lambda x: x['station'] == 'KENI') |
|
``` |
|
|
|
### 3. Analyzing Transcripts |
|
|
|
```python |
|
# Get all segments from a specific transcript |
|
transcript_ds = dataset.filter( |
|
lambda x: x['transcript_id'] == 'AK_KAGV_2024_08_25_13_00' |
|
) |
|
|
|
# Sort segments by their index to maintain order |
|
sorted_segments = sorted(transcript_ds, key=lambda x: x['segment_index']) |
|
``` |
|
|
|
## Best Practices |
|
|
|
1. **Memory Management**: |
|
- Start with a single state or small sample |
|
- Use streaming mode for large-scale processing |
|
- Clear cache when needed: `from datasets import clear_cache; clear_cache()` |
|
|
|
2. **Disk Space**: |
|
- Ensure at least 75-80 GB free space for full dataset |
|
- Use state-specific loading to reduce space requirements |
|
- Regular cache cleanup |
|
|
|
3. **Error Handling**: |
|
- Always include timeout configurations |
|
- Implement retry logic for large downloads |
|
- Handle connection errors gracefully |
|
|
|
## Example Use Cases |
|
|
|
### 1. Basic Content Analysis |
|
|
|
```python |
|
# Count segments per station |
|
from collections import Counter |
|
|
|
station_counts = Counter(dataset["train"]["station"]) |
|
print("Most common stations:", station_counts.most_common(5)) |
|
``` |
|
|
|
### 2. Time-based Analysis |
|
|
|
```python |
|
# Get distribution of segments across hours |
|
import datetime |
|
|
|
hour_distribution = Counter( |
|
datetime.datetime.fromisoformat(dt).hour |
|
for dt in dataset["train"]["datetime"] |
|
) |
|
``` |
|
|
|
### 3. Speaker Analysis |
|
|
|
```python |
|
# Analyze speaker patterns in a transcript |
|
def analyze_speakers(transcript_id): |
|
segments = dataset.filter( |
|
lambda x: x['transcript_id'] == transcript_id |
|
) |
|
speakers = [seg['speaker'] for seg in segments] |
|
return Counter(speakers) |
|
``` |
|
|
|
## Common Issues and Solutions |
|
|
|
1. **Timeout Errors**: |
|
```python |
|
# Increase timeout duration |
|
huggingface_hub.constants.HF_HUB_DOWNLOAD_TIMEOUT = 120 |
|
``` |
|
|
|
2. **Memory Errors**: |
|
```python |
|
# Use streaming to process in chunks |
|
for batch in dataset.iter(batch_size=1000): |
|
process_batch(batch) |
|
``` |
|
|
|
3. **Disk Space Issues**: |
|
```python |
|
# Check available space before downloading |
|
import shutil |
|
total, used, free = shutil.disk_usage("/") |
|
print(f"Free disk space: {free // (2**30)} GB") |
|
``` |
|
|
|
## Need Help? |
|
|
|
- Dataset documentation: https://huggingface.co/datasets/nyu-dice-lab/wavepulse-radio-raw-transcripts |
|
- Project website: https://wave-pulse.io |
|
- Report issues: https://github.com/nyu-dice-lab/wavepulse/issues |
|
|
|
Remember to cite the dataset in your work: |
|
|
|
```bibtex |
|
@article{mittal2024wavepulse, |
|
title={WavePulse: Real-time Content Analytics of Radio Livestreams}, |
|
author={Mittal, Govind and Gupta, Sarthak and Wagle, Shruti and Chopra, Chirag |
|
and DeMattee, Anthony J and Memon, Nasir and Ahamad, Mustaque |
|
and Hegde, Chinmay}, |
|
journal={arXiv preprint arXiv:2412.17998}, |
|
year={2024} |
|
} |
|
``` |