Commit
·
7b74c1d
1
Parent(s):
ab307c1
Host data files and update dates (#1)
Browse files- Add data file (0f26a0cc23ea8aa19afcb11a79b0eba4630a2f42)
- Update loading script (f18514cd1fd20dbda5b98f28a5304fb452709064)
- Update metadata num_examples from 1090591 to 2541575 (f0059d7d03589a431143b09336c3f8210680ccd5)
- Update script to iterate data files for all dates (5b609e65a8b2842d4f0b8dbab2c70585fde7cf6d)
- Update metadata num_examples from 2541575 to 3140158 (86fb2fd673767d9dbcc967010320ea1453928272)
- Update metadata dates in dataset card (09034ad1097e28c418049ff188807a75978cc6d1)
- Delete legacy dataset_infos.json (4b119b723b16ee8bb7b020a558d3c7513f68c07e)
- README.md +22 -10
- ar_cov19.py +5 -4
- dataset-all_tweets.zip +3 -0
- dataset_infos.json +0 -1
README.md
CHANGED
|
@@ -19,16 +19,16 @@ pretty_name: ArCOV19
|
|
| 19 |
tags:
|
| 20 |
- data-mining
|
| 21 |
dataset_info:
|
|
|
|
| 22 |
features:
|
| 23 |
- name: tweetID
|
| 24 |
dtype: int64
|
| 25 |
-
config_name: ar_cov19
|
| 26 |
splits:
|
| 27 |
- name: train
|
| 28 |
-
num_bytes:
|
| 29 |
-
num_examples:
|
| 30 |
-
download_size:
|
| 31 |
-
dataset_size:
|
| 32 |
---
|
| 33 |
|
| 34 |
# Dataset Card for ArCOV19
|
|
@@ -67,7 +67,15 @@ dataset_info:
|
|
| 67 |
|
| 68 |
### Dataset Summary
|
| 69 |
|
| 70 |
-
ArCOV-19 is an Arabic COVID-19 Twitter dataset that covers the period from 27th of January till
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 71 |
|
| 72 |
### Supported Tasks and Leaderboards
|
| 73 |
|
|
@@ -155,12 +163,16 @@ No annotation was provided with the dataset.
|
|
| 155 |
|
| 156 |
### Citation Information
|
| 157 |
|
|
|
|
| 158 |
@article{haouari2020arcov19,
|
| 159 |
-
|
| 160 |
-
|
| 161 |
-
|
| 162 |
-
|
|
|
|
|
|
|
| 163 |
}
|
|
|
|
| 164 |
|
| 165 |
### Contributions
|
| 166 |
|
|
|
|
| 19 |
tags:
|
| 20 |
- data-mining
|
| 21 |
dataset_info:
|
| 22 |
+
config_name: ar_cov19
|
| 23 |
features:
|
| 24 |
- name: tweetID
|
| 25 |
dtype: int64
|
|
|
|
| 26 |
splits:
|
| 27 |
- name: train
|
| 28 |
+
num_bytes: 25121264
|
| 29 |
+
num_examples: 3140158
|
| 30 |
+
download_size: 23678407
|
| 31 |
+
dataset_size: 25121264
|
| 32 |
---
|
| 33 |
|
| 34 |
# Dataset Card for ArCOV19
|
|
|
|
| 67 |
|
| 68 |
### Dataset Summary
|
| 69 |
|
| 70 |
+
ArCOV-19 is an Arabic COVID-19 Twitter dataset that covers the period from 27th of January till 5th of May 2021.
|
| 71 |
+
ArCOV-19 is the first publicly-available Arabic Twitter dataset covering COVID-19 pandemic that includes about 3.2M
|
| 72 |
+
tweets alongside the propagation networks of the most-popular subset of them (i.e., most-retweeted and-liked).
|
| 73 |
+
The propagation networks include both retweets and conversational threads (i.e., threads of replies).
|
| 74 |
+
ArCOV-19 is designed to enable research under several domains including natural language processing, information
|
| 75 |
+
retrieval, and social computing, among others. Preliminary analysis shows that ArCOV-19 captures rising discussions
|
| 76 |
+
associated with the first reported cases of the disease as they appeared in the Arab world. In addition to the source
|
| 77 |
+
tweets and the propagation networks, we also release the search queries and the language-independent crawler used to
|
| 78 |
+
collect the tweets to encourage the curation of similar datasets.
|
| 79 |
|
| 80 |
### Supported Tasks and Leaderboards
|
| 81 |
|
|
|
|
| 163 |
|
| 164 |
### Citation Information
|
| 165 |
|
| 166 |
+
```
|
| 167 |
@article{haouari2020arcov19,
|
| 168 |
+
title={ArCOV-19: The First Arabic COVID-19 Twitter Dataset with Propagation Networks},
|
| 169 |
+
author={Fatima Haouari and Maram Hasanain and Reem Suwaileh and Tamer Elsayed},
|
| 170 |
+
year={2021},
|
| 171 |
+
eprint={2004.05861},
|
| 172 |
+
archivePrefix={arXiv},
|
| 173 |
+
primaryClass={cs.CL}
|
| 174 |
}
|
| 175 |
+
```
|
| 176 |
|
| 177 |
### Contributions
|
| 178 |
|
ar_cov19.py
CHANGED
|
@@ -48,7 +48,7 @@ _HOMEPAGE = "https://gitlab.com/bigirqu/ArCOV-19"
|
|
| 48 |
# TODO: Add link to the official dataset URLs here
|
| 49 |
# The HuggingFace dataset library don't host the datasets but only point to the original files
|
| 50 |
# This can be an arbitrary nested dict/list of URLs (see below in `_split_generators` method)
|
| 51 |
-
_URL = "
|
| 52 |
# _URL="https://gitlab.com/bigirqu/ArCOV-19/-/archive/master/ArCOV-19-master.zip?path=dataset/all_tweets"
|
| 53 |
|
| 54 |
|
|
@@ -121,14 +121,15 @@ class ArCov19(datasets.GeneratorBasedBuilder):
|
|
| 121 |
# It can accept any type or nested list/dict and will give back the same structure with the url replaced with path to local files.
|
| 122 |
# By default the archives will be extracted and a path to a cached folder where they are extracted is returned instead of the archive
|
| 123 |
data_dir = dl_manager.download_and_extract(_URL)
|
| 124 |
-
|
|
|
|
| 125 |
|
| 126 |
-
def _generate_examples(self,
|
| 127 |
"""Yields examples."""
|
| 128 |
# TODO: This method will receive as arguments the `gen_kwargs` defined in the previous `_split_generators` method.
|
| 129 |
# It is in charge of opening the given file and yielding (key, example) tuples from the dataset
|
| 130 |
# The key is not important, it's more here for legacy reason (legacy from tfds)
|
| 131 |
-
for fname in
|
| 132 |
|
| 133 |
df = pd.read_csv(fname, names=["tweetID"])
|
| 134 |
for id_, record in df.iterrows():
|
|
|
|
| 48 |
# TODO: Add link to the official dataset URLs here
|
| 49 |
# The HuggingFace dataset library don't host the datasets but only point to the original files
|
| 50 |
# This can be an arbitrary nested dict/list of URLs (see below in `_split_generators` method)
|
| 51 |
+
_URL = "dataset-all_tweets.zip"
|
| 52 |
# _URL="https://gitlab.com/bigirqu/ArCOV-19/-/archive/master/ArCOV-19-master.zip?path=dataset/all_tweets"
|
| 53 |
|
| 54 |
|
|
|
|
| 121 |
# It can accept any type or nested list/dict and will give back the same structure with the url replaced with path to local files.
|
| 122 |
# By default the archives will be extracted and a path to a cached folder where they are extracted is returned instead of the archive
|
| 123 |
data_dir = dl_manager.download_and_extract(_URL)
|
| 124 |
+
data_files = dl_manager.iter_files(data_dir)
|
| 125 |
+
return [datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={"data_files": data_files})]
|
| 126 |
|
| 127 |
+
def _generate_examples(self, data_files):
|
| 128 |
"""Yields examples."""
|
| 129 |
# TODO: This method will receive as arguments the `gen_kwargs` defined in the previous `_split_generators` method.
|
| 130 |
# It is in charge of opening the given file and yielding (key, example) tuples from the dataset
|
| 131 |
# The key is not important, it's more here for legacy reason (legacy from tfds)
|
| 132 |
+
for fname in data_files:
|
| 133 |
|
| 134 |
df = pd.read_csv(fname, names=["tweetID"])
|
| 135 |
for id_, record in df.iterrows():
|
dataset-all_tweets.zip
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2c503fcf94f768847e59323aca91052bf22c3f102a01ece358ea421cf3abcbde
|
| 3 |
+
size 23678407
|
dataset_infos.json
DELETED
|
@@ -1 +0,0 @@
|
|
| 1 |
-
{"ar_cov19": {"description": "ArCOV-19 is an Arabic COVID-19 Twitter dataset that covers the period from 27th of January till 30th of April 2020. ArCOV-19 is designed to enable research under several domains including natural language processing, information retrieval, and social computing, among others\n", "citation": "@article{haouari2020arcov19,\n title={ArCOV-19: The First Arabic COVID-19 Twitter Dataset with Propagation Networks},\n author={Fatima Haouari and Maram Hasanain and Reem Suwaileh and Tamer Elsayed},\n journal={arXiv preprint arXiv:2004.05861},\n year={2020}\n", "homepage": "https://gitlab.com/bigirqu/ArCOV-19", "license": "", "features": {"tweetID": {"dtype": "int64", "id": null, "_type": "Value"}}, "post_processed": null, "supervised_keys": null, "builder_name": "ar_cov19", "config_name": "ar_cov19", "version": {"version_str": "1.0.0", "description": "", "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 8724728, "num_examples": 1090591, "dataset_name": "ar_cov19"}}, "download_checksums": {"https://gitlab.com/bigirqu/ArCOV-19/-/archive/master/ArCOV-19-master.zip": {"num_bytes": 54902390, "checksum": "96211408035f8082c072a5eb4fbccf28ef4de00379abfc9523c552ab84646579"}}, "download_size": 54902390, "post_processing_size": null, "dataset_size": 8724728, "size_in_bytes": 63627118}}
|
|
|
|
|
|