|
|
--- |
|
|
license: apache-2.0 |
|
|
task_categories: |
|
|
- summarization |
|
|
tags: |
|
|
- video |
|
|
- article generation |
|
|
language: |
|
|
- en |
|
|
--- |
|
|
# Paper and Code |
|
|
Associated with the paper: WikiVideo (https://arxiv.org/abs/2504.00939) |
|
|
|
|
|
Associated with the github repository (https://github.com/alexmartin1722/wikivideo) |
|
|
|
|
|
|
|
|
### Download instructions |
|
|
The dataset can be found on [huggingface](https://huggingface.co/datasets/hltcoe/wikivideo). However, you can't use the datasets library to access the videos because everything is tarred. Instead you need to locally download the dataset and then untar the videos (and audios if you use those). |
|
|
|
|
|
#### Step 1: Install git-lfs |
|
|
The first thing you need to do is make sure that git-lfs is installed, otherwise you won't be able to pull the video and audio tar files. |
|
|
```bash |
|
|
git lfs install |
|
|
``` |
|
|
|
|
|
#### Step 2: Clone the dataset |
|
|
After enabling git-lfs, you can now pull the dataset from huggingface. |
|
|
```bash |
|
|
git clone https://huggingface.co/datasets/hltcoe/wikivideo |
|
|
``` |
|
|
I would also tmux this because it might take a while. |
|
|
|
|
|
#### Step 3: Untar the videos |
|
|
##### Videos from MultiVENT2.0 (WikiVideo 2024) |
|
|
These videos range from 2015-2024. |
|
|
```bash |
|
|
# untar the videos |
|
|
tar -xvzf videos.tar.gz -C data/videos |
|
|
# untar the audios |
|
|
tar -xvzf audios.tar.gz -C data/audios |
|
|
``` |
|
|
|
|
|
##### Videos from the year 2025 (WikiVideo25) |
|
|
These videos are for the year 2025. They are also the videos used in the MAGMaR shared task at ACL 2026. |
|
|
```bash |
|
|
# untar the videos |
|
|
tar -xvzf videos_2025.tar.gz -C data/videos |
|
|
# untar the audios |
|
|
tar -xvzf audios_2025.tar.gz -C data/audios |
|
|
``` |
|
|
|
|
|
#### Finish |
|
|
Now you should be done. You will see a `annotations` folder in the huggingface repo, but this also exists in the `data/` folder already in the `data/wikivideo` directory. |
|
|
|
|
|
|
|
|
### Dataset Format |
|
|
In the `data/wikivideo` directory, you will find the file `final_data.json` which has the articles. This file is formatted as a dictonary: |
|
|
```json |
|
|
{ |
|
|
"Wikipedia Title": { |
|
|
"claims": [["claim1", "claim2", ...], ...], |
|
|
"original_article": ["sent1", "sent2", ...], |
|
|
"claims_to_supporting_videos": { |
|
|
"claim1": { |
|
|
"supporting_videos": ["video_id1", "video_id2", ...], |
|
|
"videos_modalities": { |
|
|
"video_id1": { |
|
|
"video": true, |
|
|
"audio": false, |
|
|
"ocr": true |
|
|
}, |
|
|
... |
|
|
} |
|
|
} |
|
|
}, |
|
|
"article": "The article text", |
|
|
"query_id": "query id", |
|
|
"videos": { |
|
|
"video_id1": { |
|
|
"anon_scale_id": "XXX", |
|
|
"language": "english", |
|
|
"video_type": "Professional | Edited | Diet Raw | Raw", |
|
|
"relevance": 3 |
|
|
} |
|
|
}, |
|
|
... |
|
|
} |
|
|
``` |
|
|
In this json, you see that the top level key is the Wikipeda Article Title. Each other key is defined as follows: |
|
|
- `claims`: The claims from the original article |
|
|
- `original_article`: The original article split into sentences |
|
|
- `claims_to_supporting_videos`: A mapping from each claim to the videos that support it along with the modalities present in each video |
|
|
- `article`: This is the human written article on the topic using the video data. |
|
|
- `query_id`: This is the query id for the article from the MultiVENT 2.0 dataset. This will be helpful when doing RAG experiments. |
|
|
- `videos`: Metadata about each video used in the article |
|
|
|
|
|
# RAG Data |
|
|
The videos that we used as the distractor set (and also include these videos in videos.tar.gz) can be found here MultiVENT 2.0 (https://huggingface.co/datasets/hltcoe/MultiVENT2.0) |
|
|
|
|
|
The relevance judgments for this data can be found in `annotations/qrels.trec` |
|
|
|
|
|
|
|
|
--- |
|
|
license: apache-2.0 |
|
|
--- |