url
stringlengths
58
61
repository_url
stringclasses
1 value
labels_url
stringlengths
72
75
comments_url
stringlengths
67
70
events_url
stringlengths
65
68
html_url
stringlengths
46
51
id
int64
599M
1B
node_id
stringlengths
18
32
number
int64
1
2.96k
title
stringlengths
1
268
user
dict
labels
list
state
stringclasses
2 values
locked
bool
1 class
assignee
dict
assignees
list
milestone
dict
comments
list
created_at
int64
1,587B
1,632B
updated_at
int64
1,587B
1,632B
closed_at
int64
1,587B
1,632B
author_association
stringclasses
4 values
active_lock_reason
null
pull_request
dict
body
stringlengths
0
228k
timeline_url
stringlengths
67
70
performed_via_github_app
null
is_pull_request
bool
2 classes
https://api.github.com/repos/huggingface/datasets/issues/422
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/422/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/422/comments
https://api.github.com/repos/huggingface/datasets/issues/422/events
https://github.com/huggingface/datasets/pull/422
663,028,497
MDExOlB1bGxSZXF1ZXN0NDU0NTE3MDU2
422
- Corrected encoding for IMDB.
{ "login": "ghazi-f", "id": 25091538, "node_id": "MDQ6VXNlcjI1MDkxNTM4", "avatar_url": "https://avatars.githubusercontent.com/u/25091538?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ghazi-f", "html_url": "https://github.com/ghazi-f", "followers_url": "https://api.github.com/users/ghazi-...
[]
closed
false
null
[]
null
[]
1,595,339,219,000
1,595,433,773,000
1,595,433,773,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/422", "html_url": "https://github.com/huggingface/datasets/pull/422", "diff_url": "https://github.com/huggingface/datasets/pull/422.diff", "patch_url": "https://github.com/huggingface/datasets/pull/422.patch" }
The preparation phase (after the download phase) crashed on windows because of charmap encoding not being able to decode certain characters. This change suggested in Issue #347 fixes it for the IMDB dataset.
https://api.github.com/repos/huggingface/datasets/issues/422/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/421
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/421/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/421/comments
https://api.github.com/repos/huggingface/datasets/issues/421/events
https://github.com/huggingface/datasets/pull/421
662,213,864
MDExOlB1bGxSZXF1ZXN0NDUzNzkzMzQ1
421
Style change
{ "login": "lordtt13", "id": 35500534, "node_id": "MDQ6VXNlcjM1NTAwNTM0", "avatar_url": "https://avatars.githubusercontent.com/u/35500534?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lordtt13", "html_url": "https://github.com/lordtt13", "followers_url": "https://api.github.com/users/lor...
[]
closed
false
null
[]
null
[ "What about the other PR #419 ?", "Oh this is the PR where I ran make quality and make style and some previous files from master were changed", "Oh right ! Let me fix the style myself if you don't mind" ]
1,595,275,709,000
1,595,434,120,000
1,595,434,119,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/421", "html_url": "https://github.com/huggingface/datasets/pull/421", "diff_url": "https://github.com/huggingface/datasets/pull/421.diff", "patch_url": "https://github.com/huggingface/datasets/pull/421.patch" }
make quality and make style ran on scripts
https://api.github.com/repos/huggingface/datasets/issues/421/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/420
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/420/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/420/comments
https://api.github.com/repos/huggingface/datasets/issues/420/events
https://github.com/huggingface/datasets/pull/420
662,029,782
MDExOlB1bGxSZXF1ZXN0NDUzNjI5OTk2
420
Better handle nested features
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,595,263,453,000
1,595,319,649,000
1,595,318,992,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/420", "html_url": "https://github.com/huggingface/datasets/pull/420", "diff_url": "https://github.com/huggingface/datasets/pull/420.diff", "patch_url": "https://github.com/huggingface/datasets/pull/420.patch" }
Changes: - added arrow schema to features conversion (it's going to be useful to fix #342 ) - make flatten handle deep features (useful for tfrecords conversion in #339 ) - add tests for flatten and features conversions - the reader now returns the kwargs to instantiate a Dataset (fix circular dependencies)
https://api.github.com/repos/huggingface/datasets/issues/420/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/419
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/419/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/419/comments
https://api.github.com/repos/huggingface/datasets/issues/419/events
https://github.com/huggingface/datasets/pull/419
661,974,747
MDExOlB1bGxSZXF1ZXN0NDUzNTgxNzQz
419
EmoContext dataset add
{ "login": "lordtt13", "id": 35500534, "node_id": "MDQ6VXNlcjM1NTAwNTM0", "avatar_url": "https://avatars.githubusercontent.com/u/35500534?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lordtt13", "html_url": "https://github.com/lordtt13", "followers_url": "https://api.github.com/users/lor...
[]
closed
false
null
[]
null
[]
1,595,260,125,000
1,595,578,921,000
1,595,578,920,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/419", "html_url": "https://github.com/huggingface/datasets/pull/419", "diff_url": "https://github.com/huggingface/datasets/pull/419.diff", "patch_url": "https://github.com/huggingface/datasets/pull/419.patch" }
EmoContext Dataset add Signed-off-by: lordtt13 <thakurtanmay72@yahoo.com>
https://api.github.com/repos/huggingface/datasets/issues/419/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/418
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/418/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/418/comments
https://api.github.com/repos/huggingface/datasets/issues/418/events
https://github.com/huggingface/datasets/issues/418
661,914,873
MDU6SXNzdWU2NjE5MTQ4NzM=
418
Addition of google drive links to dl_manager
{ "login": "lordtt13", "id": 35500534, "node_id": "MDQ6VXNlcjM1NTAwNTM0", "avatar_url": "https://avatars.githubusercontent.com/u/35500534?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lordtt13", "html_url": "https://github.com/lordtt13", "followers_url": "https://api.github.com/users/lor...
[]
closed
false
null
[]
null
[ "I think the problem is the way you wrote your urls. Try the following structure to see `https://drive.google.com/uc?export=download&id=your_file_id` . \r\n\r\n@lhoestq ", "Oh sorry, I think `_get_drive_url` is doing that. \r\n\r\nHave you tried to use `dl_manager.download_and_extract(_get_drive_url(_TRAIN_URL)`...
1,595,256,722,000
1,595,259,572,000
1,595,259,572,000
CONTRIBUTOR
null
null
Hello there, I followed the template to create a download script of my own, which works fine for me, although I had to shun the dl_manager because it was downloading nothing from the drive links and instead use gdown. This is the script for me: ```python class EmoConfig(nlp.BuilderConfig): """BuilderConfig ...
https://api.github.com/repos/huggingface/datasets/issues/418/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/417
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/417/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/417/comments
https://api.github.com/repos/huggingface/datasets/issues/417/events
https://github.com/huggingface/datasets/pull/417
661,804,054
MDExOlB1bGxSZXF1ZXN0NDUzNDMyODE5
417
Fix docstrins multiple metrics instances
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,595,250,539,000
1,595,411,460,000
1,595,411,459,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/417", "html_url": "https://github.com/huggingface/datasets/pull/417", "diff_url": "https://github.com/huggingface/datasets/pull/417.diff", "patch_url": "https://github.com/huggingface/datasets/pull/417.patch" }
We change the docstrings of `nlp.Metric.compute`, `nlp.Metric.add` and `nlp.Metric.add_batch` depending on which metric is instantiated. However we had issues when instantiating multiple metrics (docstrings were duplicated). This should fix #304
https://api.github.com/repos/huggingface/datasets/issues/417/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/416
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/416/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/416/comments
https://api.github.com/repos/huggingface/datasets/issues/416/events
https://github.com/huggingface/datasets/pull/416
661,635,393
MDExOlB1bGxSZXF1ZXN0NDUzMjg1NTM4
416
Fix xtreme panx directory
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[ "great, I think I did not download the data the way you do, but yours is more reasonable." ]
1,595,239,757,000
1,595,319,346,000
1,595,319,344,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/416", "html_url": "https://github.com/huggingface/datasets/pull/416", "diff_url": "https://github.com/huggingface/datasets/pull/416.diff", "patch_url": "https://github.com/huggingface/datasets/pull/416.patch" }
Fix #412
https://api.github.com/repos/huggingface/datasets/issues/416/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/415
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/415/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/415/comments
https://api.github.com/repos/huggingface/datasets/issues/415/events
https://github.com/huggingface/datasets/issues/415
660,687,076
MDU6SXNzdWU2NjA2ODcwNzY=
415
Something is wrong with WMT 19 kk-en dataset
{ "login": "ChenghaoMou", "id": 32014649, "node_id": "MDQ6VXNlcjMyMDE0NjQ5", "avatar_url": "https://avatars.githubusercontent.com/u/32014649?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ChenghaoMou", "html_url": "https://github.com/ChenghaoMou", "followers_url": "https://api.github.com/...
[ { "id": 2067388877, "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug", "name": "dataset bug", "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library" } ]
open
false
null
[]
null
[]
1,595,146,731,000
1,595,238,866,000
null
NONE
null
null
The translation in the `train` set does not look right: ``` >>>import nlp >>>from nlp import load_dataset >>>dataset = load_dataset('wmt19', 'kk-en') >>>dataset["train"]["translation"][0] {'kk': 'Trumpian Uncertainty', 'en': 'Трамптық белгісіздік'} >>>dataset["validation"]["translation"][0] {'kk': 'Ақша-несие...
https://api.github.com/repos/huggingface/datasets/issues/415/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/414
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/414/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/414/comments
https://api.github.com/repos/huggingface/datasets/issues/414/events
https://github.com/huggingface/datasets/issues/414
660,654,013
MDU6SXNzdWU2NjA2NTQwMTM=
414
from_dict delete?
{ "login": "hackerxiaobai", "id": 22817243, "node_id": "MDQ6VXNlcjIyODE3MjQz", "avatar_url": "https://avatars.githubusercontent.com/u/22817243?v=4", "gravatar_id": "", "url": "https://api.github.com/users/hackerxiaobai", "html_url": "https://github.com/hackerxiaobai", "followers_url": "https://api.githu...
[]
closed
false
null
[]
null
[ "`from_dict` was added in #350 that was unfortunately not included in the 0.3.0 release. It's going to be included in the next release that will be out pretty soon though.\r\nRight now if you want to use `from_dict` you have to install the package from the master branch\r\n```\r\npip install git+https://github.com/...
1,595,142,516,000
1,595,298,077,000
1,595,298,077,000
NONE
null
null
AttributeError: type object 'Dataset' has no attribute 'from_dict'
https://api.github.com/repos/huggingface/datasets/issues/414/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/413
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/413/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/413/comments
https://api.github.com/repos/huggingface/datasets/issues/413/events
https://github.com/huggingface/datasets/issues/413
660,063,655
MDU6SXNzdWU2NjAwNjM2NTU=
413
Is there a way to download only NQ dev?
{ "login": "tholor", "id": 1563902, "node_id": "MDQ6VXNlcjE1NjM5MDI=", "avatar_url": "https://avatars.githubusercontent.com/u/1563902?v=4", "gravatar_id": "", "url": "https://api.github.com/users/tholor", "html_url": "https://github.com/tholor", "followers_url": "https://api.github.com/users/tholor/foll...
[]
open
false
null
[]
null
[ "Unfortunately it's not possible to download only the dev set of NQ.\r\n\r\nI think we could add a way to download only the test set by adding a custom configuration to the processing script though.", "Ok, got it. I think this could be a valuable feature - especially for large datasets like NQ, but potentially al...
1,595,068,103,000
1,596,027,980,000
null
NONE
null
null
Maybe I missed that in the docs, but is there a way to only download the dev set of natural questions (~1 GB)? As we want to benchmark QA models on different datasets, I would like to avoid downloading the 41GB of training data. I tried ``` dataset = nlp.load_dataset('natural_questions', split="validation", bea...
https://api.github.com/repos/huggingface/datasets/issues/413/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/412
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/412/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/412/comments
https://api.github.com/repos/huggingface/datasets/issues/412/events
https://github.com/huggingface/datasets/issues/412
660,047,139
MDU6SXNzdWU2NjAwNDcxMzk=
412
Unable to load XTREME dataset from disk
{ "login": "lewtun", "id": 26859204, "node_id": "MDQ6VXNlcjI2ODU5MjA0", "avatar_url": "https://avatars.githubusercontent.com/u/26859204?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lewtun", "html_url": "https://github.com/lewtun", "followers_url": "https://api.github.com/users/lewtun/fo...
[]
closed
false
null
[]
null
[ "Hi @lewtun, you have to provide the full path to the downloaded file for example `/home/lewtum/..`", "I was able to repro. Opening a PR to fix that.\r\nThanks for reporting this issue !", "Thanks for the rapid fix @lhoestq!" ]
1,595,066,100,000
1,595,319,344,000
1,595,319,344,000
MEMBER
null
null
Hi 🤗 team! ## Description of the problem Following the [docs](https://huggingface.co/nlp/loading_datasets.html?highlight=xtreme#manually-downloading-files) I'm trying to load the `PAN-X.fr` dataset from the [XTREME](https://github.com/google-research/xtreme) benchmark. I have manually downloaded the `AmazonPho...
https://api.github.com/repos/huggingface/datasets/issues/412/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/411
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/411/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/411/comments
https://api.github.com/repos/huggingface/datasets/issues/411/events
https://github.com/huggingface/datasets/pull/411
659,393,398
MDExOlB1bGxSZXF1ZXN0NDUxMjQxOTQy
411
Sbf
{ "login": "mariamabarham", "id": 38249783, "node_id": "MDQ6VXNlcjM4MjQ5Nzgz", "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariamabarham", "html_url": "https://github.com/mariamabarham", "followers_url": "https://api.githu...
[]
closed
false
null
[]
null
[]
1,595,002,785,000
1,595,322,826,000
1,595,322,825,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/411", "html_url": "https://github.com/huggingface/datasets/pull/411", "diff_url": "https://github.com/huggingface/datasets/pull/411.diff", "patch_url": "https://github.com/huggingface/datasets/pull/411.patch" }
This PR adds the Social Bias Frames Dataset (ACL 2020) . dataset homepage: https://homes.cs.washington.edu/~msap/social-bias-frames/
https://api.github.com/repos/huggingface/datasets/issues/411/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/410
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/410/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/410/comments
https://api.github.com/repos/huggingface/datasets/issues/410/events
https://github.com/huggingface/datasets/pull/410
659,242,871
MDExOlB1bGxSZXF1ZXN0NDUxMTEzMTI3
410
20newsgroup
{ "login": "mariamabarham", "id": 38249783, "node_id": "MDQ6VXNlcjM4MjQ5Nzgz", "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariamabarham", "html_url": "https://github.com/mariamabarham", "followers_url": "https://api.githu...
[]
closed
false
null
[]
null
[]
1,594,991,277,000
1,595,228,729,000
1,595,228,728,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/410", "html_url": "https://github.com/huggingface/datasets/pull/410", "diff_url": "https://github.com/huggingface/datasets/pull/410.diff", "patch_url": "https://github.com/huggingface/datasets/pull/410.patch" }
Add 20Newsgroup dataset. #353
https://api.github.com/repos/huggingface/datasets/issues/410/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/409
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/409/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/409/comments
https://api.github.com/repos/huggingface/datasets/issues/409/events
https://github.com/huggingface/datasets/issues/409
659,128,611
MDU6SXNzdWU2NTkxMjg2MTE=
409
train_test_split error: 'dict' object has no attribute 'deepcopy'
{ "login": "morganmcg1", "id": 20516801, "node_id": "MDQ6VXNlcjIwNTE2ODAx", "avatar_url": "https://avatars.githubusercontent.com/u/20516801?v=4", "gravatar_id": "", "url": "https://api.github.com/users/morganmcg1", "html_url": "https://github.com/morganmcg1", "followers_url": "https://api.github.com/use...
[]
closed
false
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[ { "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.git...
null
[ "It was fixed in 2ddd18d139d3047c9c3abe96e1e7d05bb360132c.\r\nCould you pull the latest changes from master @morganmcg1 ?", "Thanks @lhoestq, works fine now!" ]
1,594,982,188,000
1,595,342,092,000
1,595,342,092,000
NONE
null
null
`train_test_split` is giving me an error when I try and call it: `'dict' object has no attribute 'deepcopy'` ## To reproduce ``` dataset = load_dataset('glue', 'mrpc', split='train') dataset = dataset.train_test_split(test_size=0.2) ``` ## Full Stacktrace ``` -------------------------------------------...
https://api.github.com/repos/huggingface/datasets/issues/409/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/408
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/408/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/408/comments
https://api.github.com/repos/huggingface/datasets/issues/408/events
https://github.com/huggingface/datasets/pull/408
659,064,144
MDExOlB1bGxSZXF1ZXN0NDUwOTU1MTE0
408
Add tests datasets gcp
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,594,977,807,000
1,594,978,017,000
1,594,978,016,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/408", "html_url": "https://github.com/huggingface/datasets/pull/408", "diff_url": "https://github.com/huggingface/datasets/pull/408.diff", "patch_url": "https://github.com/huggingface/datasets/pull/408.patch" }
Some datasets are available on our google cloud storage in arrow format, so that the users don't need to process the data. These tests make sure that they're always available. It also makes sure that their scripts are in sync between S3 and the repo. This should avoid future issues like #407
https://api.github.com/repos/huggingface/datasets/issues/408/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/407
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/407/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/407/comments
https://api.github.com/repos/huggingface/datasets/issues/407/events
https://github.com/huggingface/datasets/issues/407
658,672,736
MDU6SXNzdWU2NTg2NzI3MzY=
407
MissingBeamOptions for Wikipedia 20200501.en
{ "login": "mitchellgordon95", "id": 7490438, "node_id": "MDQ6VXNlcjc0OTA0Mzg=", "avatar_url": "https://avatars.githubusercontent.com/u/7490438?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mitchellgordon95", "html_url": "https://github.com/mitchellgordon95", "followers_url": "https://ap...
[]
closed
false
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[ { "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.git...
null
[ "Fixed. Could you try again @mitchellgordon95 ?\r\nIt was due a file not being updated on S3.\r\n\r\nWe need to make sure all the datasets scripts get updated properly @julien-c ", "Works for me! Thanks.", "I found the same issue with almost any language other than English. (For English, it works). Will someone...
1,594,943,283,000
1,610,451,676,000
1,594,995,868,000
CONTRIBUTOR
null
null
There may or may not be a regression for the pre-processed Wikipedia dataset. This was working fine 10 commits ago (without having Apache Beam available): ``` nlp.load_dataset('wikipedia', "20200501.en", split='train') ``` And now, having pulled master, I get: ``` Downloading and preparing dataset wikipedia...
https://api.github.com/repos/huggingface/datasets/issues/407/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/406
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/406/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/406/comments
https://api.github.com/repos/huggingface/datasets/issues/406/events
https://github.com/huggingface/datasets/issues/406
658,581,764
MDU6SXNzdWU2NTg1ODE3NjQ=
406
Faster Shuffling?
{ "login": "mitchellgordon95", "id": 7490438, "node_id": "MDQ6VXNlcjc0OTA0Mzg=", "avatar_url": "https://avatars.githubusercontent.com/u/7490438?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mitchellgordon95", "html_url": "https://github.com/mitchellgordon95", "followers_url": "https://ap...
[]
closed
false
null
[]
null
[ "I think the slowness here probably come from the fact that we are copying from and to python.\r\n\r\n@lhoestq for all the `select`-based methods I think we should stay in Arrow format and update the writer so that it can accept Arrow tables or batches as well. What do you think?", "> @lhoestq for all the `select...
1,594,934,513,000
1,599,489,926,000
1,599,489,925,000
CONTRIBUTOR
null
null
Consider shuffling bookcorpus: ``` dataset = nlp.load_dataset('bookcorpus', split='train') dataset.shuffle() ``` According to tqdm, this will take around 2.5 hours on my machine to complete (even with the faster version of select from #405). I've also tried with `keep_in_memory=True` and `writer_batch_size=1000`...
https://api.github.com/repos/huggingface/datasets/issues/406/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/405
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/405/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/405/comments
https://api.github.com/repos/huggingface/datasets/issues/405/events
https://github.com/huggingface/datasets/pull/405
658,580,192
MDExOlB1bGxSZXF1ZXN0NDUwNTI1MTc3
405
Make select() faster by batching reads
{ "login": "mitchellgordon95", "id": 7490438, "node_id": "MDQ6VXNlcjc0OTA0Mzg=", "avatar_url": "https://avatars.githubusercontent.com/u/7490438?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mitchellgordon95", "html_url": "https://github.com/mitchellgordon95", "followers_url": "https://ap...
[]
closed
false
null
[]
null
[]
1,594,934,385,000
1,595,005,544,000
1,595,004,686,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/405", "html_url": "https://github.com/huggingface/datasets/pull/405", "diff_url": "https://github.com/huggingface/datasets/pull/405.diff", "patch_url": "https://github.com/huggingface/datasets/pull/405.patch" }
Here's a benchmark: ``` dataset = nlp.load_dataset('bookcorpus', split='train') start = time.time() dataset.select(np.arange(1000), reader_batch_size=1, load_from_cache_file=False) end = time.time() print(f'{end - start}') start = time.time() dataset.select(np.arange(1000), reader_batch_size=1000, load_fr...
https://api.github.com/repos/huggingface/datasets/issues/405/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/404
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/404/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/404/comments
https://api.github.com/repos/huggingface/datasets/issues/404/events
https://github.com/huggingface/datasets/pull/404
658,400,987
MDExOlB1bGxSZXF1ZXN0NDUwMzY4Mjg4
404
Add seed in metrics
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,594,920,425,000
1,595,239,955,000
1,595,239,954,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/404", "html_url": "https://github.com/huggingface/datasets/pull/404", "diff_url": "https://github.com/huggingface/datasets/pull/404.diff", "patch_url": "https://github.com/huggingface/datasets/pull/404.patch" }
With #361 we noticed that some metrics were not deterministic. In this PR I allow the user to specify numpy's seed when instantiating a metric with `load_metric`. The seed is set only when `compute` is called, and reset afterwards. Moreover when calling `compute` with the same metric instance (i.e. same experiment...
https://api.github.com/repos/huggingface/datasets/issues/404/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/403
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/403/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/403/comments
https://api.github.com/repos/huggingface/datasets/issues/403/events
https://github.com/huggingface/datasets/pull/403
658,325,756
MDExOlB1bGxSZXF1ZXN0NDUwMzAzNjI2
403
return python objects instead of arrays by default
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,594,914,712,000
1,594,985,821,000
1,594,985,820,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/403", "html_url": "https://github.com/huggingface/datasets/pull/403", "diff_url": "https://github.com/huggingface/datasets/pull/403.diff", "patch_url": "https://github.com/huggingface/datasets/pull/403.patch" }
We were using to_pandas() to convert from arrow types, however it returns numpy arrays instead of python lists. I fixed it by using to_pydict/to_pylist instead. Fix #387 It was mentioned in https://github.com/huggingface/transformers/issues/5729
https://api.github.com/repos/huggingface/datasets/issues/403/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/402
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/402/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/402/comments
https://api.github.com/repos/huggingface/datasets/issues/402/events
https://github.com/huggingface/datasets/pull/402
658,001,288
MDExOlB1bGxSZXF1ZXN0NDUwMDI2NTE0
402
Search qa
{ "login": "mariamabarham", "id": 38249783, "node_id": "MDQ6VXNlcjM4MjQ5Nzgz", "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariamabarham", "html_url": "https://github.com/mariamabarham", "followers_url": "https://api.githu...
[]
closed
false
null
[]
null
[]
1,594,890,010,000
1,594,909,620,000
1,594,909,619,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/402", "html_url": "https://github.com/huggingface/datasets/pull/402", "diff_url": "https://github.com/huggingface/datasets/pull/402.diff", "patch_url": "https://github.com/huggingface/datasets/pull/402.patch" }
add SearchQA dataset #336
https://api.github.com/repos/huggingface/datasets/issues/402/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/401
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/401/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/401/comments
https://api.github.com/repos/huggingface/datasets/issues/401/events
https://github.com/huggingface/datasets/pull/401
657,996,252
MDExOlB1bGxSZXF1ZXN0NDUwMDIyNTc0
401
add web_questions
{ "login": "mariamabarham", "id": 38249783, "node_id": "MDQ6VXNlcjM4MjQ5Nzgz", "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariamabarham", "html_url": "https://github.com/mariamabarham", "followers_url": "https://api.githu...
[]
closed
false
null
[]
null
[ "What does the `nlp-cli dummy_data` command returns ?", "`test.json` -> `test` \r\nand \r\n`train.json` -> `train`\r\n\r\nas shown by the `nlp-cli dummy_data` command ;-)", "LGTM for merge @lhoestq - I let you merge if you want to." ]
1,594,889,699,000
1,596,694,580,000
1,596,694,579,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/401", "html_url": "https://github.com/huggingface/datasets/pull/401", "diff_url": "https://github.com/huggingface/datasets/pull/401.diff", "patch_url": "https://github.com/huggingface/datasets/pull/401.patch" }
add Web Question dataset #336 Maybe @patrickvonplaten you can help with the dummy_data structure? it still broken
https://api.github.com/repos/huggingface/datasets/issues/401/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/400
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/400/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/400/comments
https://api.github.com/repos/huggingface/datasets/issues/400/events
https://github.com/huggingface/datasets/pull/400
657,975,600
MDExOlB1bGxSZXF1ZXN0NDUwMDA1MDU5
400
Web questions
{ "login": "mariamabarham", "id": 38249783, "node_id": "MDQ6VXNlcjM4MjQ5Nzgz", "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariamabarham", "html_url": "https://github.com/mariamabarham", "followers_url": "https://api.githu...
[]
closed
false
null
[]
null
[]
1,594,888,109,000
1,594,889,451,000
1,594,888,974,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/400", "html_url": "https://github.com/huggingface/datasets/pull/400", "diff_url": "https://github.com/huggingface/datasets/pull/400.diff", "patch_url": "https://github.com/huggingface/datasets/pull/400.patch" }
add the WebQuestion dataset #336
https://api.github.com/repos/huggingface/datasets/issues/400/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/399
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/399/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/399/comments
https://api.github.com/repos/huggingface/datasets/issues/399/events
https://github.com/huggingface/datasets/pull/399
657,841,433
MDExOlB1bGxSZXF1ZXN0NDQ5ODkxNTEy
399
Spelling mistake
{ "login": "BlancRay", "id": 9410067, "node_id": "MDQ6VXNlcjk0MTAwNjc=", "avatar_url": "https://avatars.githubusercontent.com/u/9410067?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BlancRay", "html_url": "https://github.com/BlancRay", "followers_url": "https://api.github.com/users/Blanc...
[]
closed
false
null
[]
null
[ "Thanks!" ]
1,594,874,278,000
1,594,882,188,000
1,594,882,177,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/399", "html_url": "https://github.com/huggingface/datasets/pull/399", "diff_url": "https://github.com/huggingface/datasets/pull/399.diff", "patch_url": "https://github.com/huggingface/datasets/pull/399.patch" }
In "Formatting the dataset" part, "The two toehr modifications..." should be "The two other modifications..." ,the word "other" wrong spelled as "toehr".
https://api.github.com/repos/huggingface/datasets/issues/399/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/398
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/398/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/398/comments
https://api.github.com/repos/huggingface/datasets/issues/398/events
https://github.com/huggingface/datasets/pull/398
657,511,962
MDExOlB1bGxSZXF1ZXN0NDQ5NjE1OTk1
398
Add inline links
{ "login": "Bharat123rox", "id": 13381361, "node_id": "MDQ6VXNlcjEzMzgxMzYx", "avatar_url": "https://avatars.githubusercontent.com/u/13381361?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Bharat123rox", "html_url": "https://github.com/Bharat123rox", "followers_url": "https://api.github.c...
[]
closed
false
null
[]
null
[ "Do you mind adding a link to the much more extended pages on adding and sharing a dataset in the new documentation?", "Sure, I will do that too" ]
1,594,832,644,000
1,595,412,862,000
1,595,412,862,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/398", "html_url": "https://github.com/huggingface/datasets/pull/398", "diff_url": "https://github.com/huggingface/datasets/pull/398.diff", "patch_url": "https://github.com/huggingface/datasets/pull/398.patch" }
Add inline links to `Contributing.md`
https://api.github.com/repos/huggingface/datasets/issues/398/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/397
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/397/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/397/comments
https://api.github.com/repos/huggingface/datasets/issues/397/events
https://github.com/huggingface/datasets/pull/397
657,510,856
MDExOlB1bGxSZXF1ZXN0NDQ5NjE1MDA4
397
Add contiguous sharding
{ "login": "jarednielsen", "id": 4564897, "node_id": "MDQ6VXNlcjQ1NjQ4OTc=", "avatar_url": "https://avatars.githubusercontent.com/u/4564897?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jarednielsen", "html_url": "https://github.com/jarednielsen", "followers_url": "https://api.github.com...
[]
closed
false
null
[]
null
[]
1,594,832,578,000
1,595,005,171,000
1,595,005,171,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/397", "html_url": "https://github.com/huggingface/datasets/pull/397", "diff_url": "https://github.com/huggingface/datasets/pull/397.diff", "patch_url": "https://github.com/huggingface/datasets/pull/397.patch" }
This makes dset.shard() play nice with nlp.concatenate_datasets(). When I originally wrote the shard() method, I was thinking about a distributed training scenario, but https://github.com/huggingface/nlp/pull/389 also uses it for splitting the dataset for distributed preprocessing. Usage: ``` nlp.concatenate_datas...
https://api.github.com/repos/huggingface/datasets/issues/397/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/396
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/396/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/396/comments
https://api.github.com/repos/huggingface/datasets/issues/396/events
https://github.com/huggingface/datasets/pull/396
657,477,952
MDExOlB1bGxSZXF1ZXN0NDQ5NTg3MDQ4
396
Fix memory issue when doing select
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,594,829,704,000
1,594,886,852,000
1,594,886,851,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/396", "html_url": "https://github.com/huggingface/datasets/pull/396", "diff_url": "https://github.com/huggingface/datasets/pull/396.diff", "patch_url": "https://github.com/huggingface/datasets/pull/396.patch" }
We were passing the `nlp.Dataset` object to get the hash for the new dataset's file name. Fix #395
https://api.github.com/repos/huggingface/datasets/issues/396/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/395
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/395/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/395/comments
https://api.github.com/repos/huggingface/datasets/issues/395/events
https://github.com/huggingface/datasets/issues/395
657,454,983
MDU6SXNzdWU2NTc0NTQ5ODM=
395
Memory issue when doing select
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[ { "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.git...
null
[]
1,594,827,818,000
1,594,886,851,000
1,594,886,851,000
MEMBER
null
null
As noticed in #389, the following code loads the entire wikipedia in memory. ```python import nlp w = nlp.load_dataset("wikipedia", "20200501.en", split="train") w.select([0]) ``` This is caused by [this line](https://github.com/huggingface/nlp/blob/master/src/nlp/arrow_dataset.py#L626) for some reason, that ...
https://api.github.com/repos/huggingface/datasets/issues/395/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/394
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/394/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/394/comments
https://api.github.com/repos/huggingface/datasets/issues/394/events
https://github.com/huggingface/datasets/pull/394
657,425,548
MDExOlB1bGxSZXF1ZXN0NDQ5NTQzNTE0
394
Remove remaining nested dict
{ "login": "mariamabarham", "id": 38249783, "node_id": "MDQ6VXNlcjM4MjQ5Nzgz", "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariamabarham", "html_url": "https://github.com/mariamabarham", "followers_url": "https://api.githu...
[]
closed
false
null
[]
null
[]
1,594,825,552,000
1,594,885,192,000
1,594,885,191,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/394", "html_url": "https://github.com/huggingface/datasets/pull/394", "diff_url": "https://github.com/huggingface/datasets/pull/394.diff", "patch_url": "https://github.com/huggingface/datasets/pull/394.patch" }
This PR deletes the remaining unnecessary nested dict #378
https://api.github.com/repos/huggingface/datasets/issues/394/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/393
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/393/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/393/comments
https://api.github.com/repos/huggingface/datasets/issues/393/events
https://github.com/huggingface/datasets/pull/393
657,330,911
MDExOlB1bGxSZXF1ZXN0NDQ5NDY1MTAz
393
Fix extracted files directory for the DownloadManager
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,594,817,995,000
1,595,005,336,000
1,595,005,334,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/393", "html_url": "https://github.com/huggingface/datasets/pull/393", "diff_url": "https://github.com/huggingface/datasets/pull/393.diff", "patch_url": "https://github.com/huggingface/datasets/pull/393.patch" }
The cache dir was often cluttered by extracted files because of the download manager. For downloaded files, we are using the `downloads` directory to make things easier to navigate, but extracted files were still placed at the root of the cache directory. To fix that I changed the directory for extracted files to ca...
https://api.github.com/repos/huggingface/datasets/issues/393/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/392
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/392/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/392/comments
https://api.github.com/repos/huggingface/datasets/issues/392/events
https://github.com/huggingface/datasets/pull/392
657,313,738
MDExOlB1bGxSZXF1ZXN0NDQ5NDUwOTkx
392
Style change detection
{ "login": "ghomasHudson", "id": 13795113, "node_id": "MDQ6VXNlcjEzNzk1MTEz", "avatar_url": "https://avatars.githubusercontent.com/u/13795113?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ghomasHudson", "html_url": "https://github.com/ghomasHudson", "followers_url": "https://api.github.c...
[]
closed
false
null
[]
null
[]
1,594,816,334,000
1,595,337,516,000
1,595,006,003,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/392", "html_url": "https://github.com/huggingface/datasets/pull/392", "diff_url": "https://github.com/huggingface/datasets/pull/392.diff", "patch_url": "https://github.com/huggingface/datasets/pull/392.patch" }
Another [PAN task](https://pan.webis.de/clef20/pan20-web/style-change-detection.html). This time about identifying when the style/author changes in documents. - There's the possibility of adding the [PAN19](https://zenodo.org/record/3577602) and PAN18 style change detection tasks too (these are datasets whose labels...
https://api.github.com/repos/huggingface/datasets/issues/392/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/391
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/391/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/391/comments
https://api.github.com/repos/huggingface/datasets/issues/391/events
https://github.com/huggingface/datasets/issues/391
656,991,432
MDU6SXNzdWU2NTY5OTE0MzI=
391
🌟 [Metric Request] WOOD score
{ "login": "astariul", "id": 43774355, "node_id": "MDQ6VXNlcjQzNzc0MzU1", "avatar_url": "https://avatars.githubusercontent.com/u/43774355?v=4", "gravatar_id": "", "url": "https://api.github.com/users/astariul", "html_url": "https://github.com/astariul", "followers_url": "https://api.github.com/users/ast...
[ { "id": 2459308248, "node_id": "MDU6TGFiZWwyNDU5MzA4MjQ4", "url": "https://api.github.com/repos/huggingface/datasets/labels/metric%20request", "name": "metric request", "color": "d4c5f9", "default": false, "description": "Requesting to add a new metric" } ]
open
false
null
[]
null
[]
1,594,775,797,000
1,603,813,408,000
null
NONE
null
null
WOOD score paper : https://arxiv.org/pdf/2007.06898.pdf Abstract : >Models that surpass human performance on several popular benchmarks display significant degradation in performance on exposure to Out of Distribution (OOD) data. Recent research has shown that models overfit to spurious biases and ‘hack’ datasets...
https://api.github.com/repos/huggingface/datasets/issues/391/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/390
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/390/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/390/comments
https://api.github.com/repos/huggingface/datasets/issues/390/events
https://github.com/huggingface/datasets/pull/390
656,956,384
MDExOlB1bGxSZXF1ZXN0NDQ5MTYxMzY3
390
Concatenate datasets
{ "login": "jarednielsen", "id": 4564897, "node_id": "MDQ6VXNlcjQ1NjQ4OTc=", "avatar_url": "https://avatars.githubusercontent.com/u/4564897?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jarednielsen", "html_url": "https://github.com/jarednielsen", "followers_url": "https://api.github.com...
[]
closed
false
null
[]
null
[ "Looks cool :)\r\n\r\nI feel like \r\n```python\r\nconcatenated_dataset = dataset1.concatenate(dataset2)\r\n```\r\ncould be more natural. What do you think ?\r\n\r\nAlso could you also concatenate the `nlp.Dataset._data_files` ?\r\n```python\r\nreturn cls(table, info=info, split=split, data_files=self._data_files +...
1,594,769,077,000
1,595,411,398,000
1,595,411,398,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/390", "html_url": "https://github.com/huggingface/datasets/pull/390", "diff_url": "https://github.com/huggingface/datasets/pull/390.diff", "patch_url": "https://github.com/huggingface/datasets/pull/390.patch" }
I'm constructing the "WikiBooks" dataset, which is a concatenation of Wikipedia & BookCorpus. So I implemented the `Dataset.from_concat()` method, which concatenates two datasets with the same schema. This would also be useful if someone wants to pretrain on a large generic dataset + their own custom dataset. Not in...
https://api.github.com/repos/huggingface/datasets/issues/390/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/389
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/389/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/389/comments
https://api.github.com/repos/huggingface/datasets/issues/389/events
https://github.com/huggingface/datasets/pull/389
656,921,768
MDExOlB1bGxSZXF1ZXN0NDQ5MTMyOTU5
389
Fix pickling of SplitDict
{ "login": "mitchellgordon95", "id": 7490438, "node_id": "MDQ6VXNlcjc0OTA0Mzg=", "avatar_url": "https://avatars.githubusercontent.com/u/7490438?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mitchellgordon95", "html_url": "https://github.com/mitchellgordon95", "followers_url": "https://ap...
[]
closed
false
null
[]
null
[ "By the way, the reason this is an issue for me is because I want to be able to \"save\" changes made to a dataset by writing something to disk. In this case, I would like to pre-process my dataset once, and then train multiple models on the dataset later without having to re-process the data. \r\n\r\nIs pickling/u...
1,594,763,619,000
1,596,551,890,000
1,596,551,890,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/389", "html_url": "https://github.com/huggingface/datasets/pull/389", "diff_url": "https://github.com/huggingface/datasets/pull/389.diff", "patch_url": "https://github.com/huggingface/datasets/pull/389.patch" }
It would be nice to pickle and unpickle Datasets, as done in [this tutorial](https://github.com/patil-suraj/exploring-T5/blob/master/T5_on_TPU.ipynb). Example: ``` wiki = nlp.load_dataset('wikipedia', split='train') def sentencize(examples): ... wiki = wiki.map(sentencize, batched=True) torch.save(wiki, '...
https://api.github.com/repos/huggingface/datasets/issues/389/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/388
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/388/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/388/comments
https://api.github.com/repos/huggingface/datasets/issues/388/events
https://github.com/huggingface/datasets/issues/388
656,707,497
MDU6SXNzdWU2NTY3MDc0OTc=
388
🐛 [Dataset] Cannot download wmt14, wmt15 and wmt17
{ "login": "SamuelCahyawijaya", "id": 2826602, "node_id": "MDQ6VXNlcjI4MjY2MDI=", "avatar_url": "https://avatars.githubusercontent.com/u/2826602?v=4", "gravatar_id": "", "url": "https://api.github.com/users/SamuelCahyawijaya", "html_url": "https://github.com/SamuelCahyawijaya", "followers_url": "https:/...
[ { "id": 2067388877, "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug", "name": "dataset bug", "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library" } ]
open
false
{ "login": "patrickvonplaten", "id": 23423619, "node_id": "MDQ6VXNlcjIzNDIzNjE5", "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "gravatar_id": "", "url": "https://api.github.com/users/patrickvonplaten", "html_url": "https://github.com/patrickvonplaten", "followers_url": "https://...
[ { "login": "patrickvonplaten", "id": 23423619, "node_id": "MDQ6VXNlcjIzNDIzNjE5", "avatar_url": "https://avatars.githubusercontent.com/u/23423619?v=4", "gravatar_id": "", "url": "https://api.github.com/users/patrickvonplaten", "html_url": "https://github.com/patrickvonplaten", "follo...
null
[ "similar slow download speed here for nlp.load_dataset('wmt14', 'fr-en')\r\n`\r\nDownloading: 100%|██████████████████████████████████████████████████████████| 658M/658M [1:00:42<00:00, 181kB/s]\r\nDownloading: 100%|██████████████████████████████████████████████████████████| 918M/918M [1:39:38<00:00, 154kB/s]\r\nDow...
1,594,741,001,000
1,596,639,392,000
null
NONE
null
null
1. I try downloading `wmt14`, `wmt15`, `wmt17`, `wmt19` with the following code: ``` nlp.load_dataset('wmt14','de-en') nlp.load_dataset('wmt15','de-en') nlp.load_dataset('wmt17','de-en') nlp.load_dataset('wmt19','de-en') ``` The code runs but the download speed is **extremely slow**, the same behaviour is not ob...
https://api.github.com/repos/huggingface/datasets/issues/388/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/387
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/387/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/387/comments
https://api.github.com/repos/huggingface/datasets/issues/387/events
https://github.com/huggingface/datasets/issues/387
656,361,357
MDU6SXNzdWU2NTYzNjEzNTc=
387
Conversion through to_pandas output numpy arrays for lists instead of python objects
{ "login": "thomwolf", "id": 7353373, "node_id": "MDQ6VXNlcjczNTMzNzM=", "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "gravatar_id": "", "url": "https://api.github.com/users/thomwolf", "html_url": "https://github.com/thomwolf", "followers_url": "https://api.github.com/users/thomw...
[]
closed
false
null
[]
null
[ "To convert from arrow type we have three options: to_numpy, to_pandas and to_pydict/to_pylist.\r\n\r\n- to_numpy and to_pandas return numpy arrays instead of lists but are very fast.\r\n- to_pydict/to_pylist can be 100x slower and become the bottleneck for reading data, but at least they return lists.\r\n\r\nMaybe...
1,594,707,841,000
1,594,985,820,000
1,594,985,820,000
MEMBER
null
null
In a related question, the conversion through to_pandas output numpy arrays for the lists instead of python objects. Here is an example: ```python >>> dataset._data.slice(key, 1).to_pandas().to_dict("list") {'sentence1': ['Amrozi accused his brother , whom he called " the witness " , of deliberately distorting hi...
https://api.github.com/repos/huggingface/datasets/issues/387/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/386
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/386/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/386/comments
https://api.github.com/repos/huggingface/datasets/issues/386/events
https://github.com/huggingface/datasets/pull/386
655,839,067
MDExOlB1bGxSZXF1ZXN0NDQ4MjQ1NDI4
386
Update dataset loading and features - Add TREC dataset
{ "login": "thomwolf", "id": 7353373, "node_id": "MDQ6VXNlcjczNTMzNzM=", "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "gravatar_id": "", "url": "https://api.github.com/users/thomwolf", "html_url": "https://github.com/thomwolf", "followers_url": "https://api.github.com/users/thomw...
[]
closed
false
null
[]
null
[ "I just copied the files that are on google storage to follow the new `_relative_data_dir ` format. It should be good to merge now :)\r\n\r\nWell actually it seems there are some merge conflicts to fix first" ]
1,594,645,818,000
1,594,887,478,000
1,594,887,478,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/386", "html_url": "https://github.com/huggingface/datasets/pull/386", "diff_url": "https://github.com/huggingface/datasets/pull/386.diff", "patch_url": "https://github.com/huggingface/datasets/pull/386.patch" }
This PR: - add a template for a new dataset script - update the caching structure so that the path to the cached data files is also a function of the dataset loading script hash. This way when you update a loading script the data will be automatically updated instead of falling back to the previous version (which is ...
https://api.github.com/repos/huggingface/datasets/issues/386/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/385
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/385/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/385/comments
https://api.github.com/repos/huggingface/datasets/issues/385/events
https://github.com/huggingface/datasets/pull/385
655,663,997
MDExOlB1bGxSZXF1ZXN0NDQ4MTAzMjY5
385
Remove unnecessary nested dict
{ "login": "mariamabarham", "id": 38249783, "node_id": "MDQ6VXNlcjM4MjQ5Nzgz", "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariamabarham", "html_url": "https://github.com/mariamabarham", "followers_url": "https://api.githu...
[]
closed
false
null
[]
null
[ "We can probably scan the dataset scripts with a regexpr to try to identify this pattern cc @patrickvonplaten maybe", "@mariamabarham This script should work. I tested it for a couple of datasets. There might be exceptions where the script breaks - did not test everything.\r\n\r\n```python\r\n#!/usr/bin/env pytho...
1,594,629,983,000
1,594,812,458,000
1,594,807,433,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/385", "html_url": "https://github.com/huggingface/datasets/pull/385", "diff_url": "https://github.com/huggingface/datasets/pull/385.diff", "patch_url": "https://github.com/huggingface/datasets/pull/385.patch" }
This PR is removing unnecessary nested dictionary used in some datasets. For now the following datasets are updated: - MLQA - RACE Will be adding more if necessary. #378
https://api.github.com/repos/huggingface/datasets/issues/385/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/383
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/383/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/383/comments
https://api.github.com/repos/huggingface/datasets/issues/383/events
https://github.com/huggingface/datasets/pull/383
655,291,201
MDExOlB1bGxSZXF1ZXN0NDQ3ODI0OTky
383
Adding the Linguistic Code-switching Evaluation (LinCE) benchmark
{ "login": "gaguilar", "id": 5833357, "node_id": "MDQ6VXNlcjU4MzMzNTc=", "avatar_url": "https://avatars.githubusercontent.com/u/5833357?v=4", "gravatar_id": "", "url": "https://api.github.com/users/gaguilar", "html_url": "https://github.com/gaguilar", "followers_url": "https://api.github.com/users/gagui...
[]
closed
false
null
[]
null
[ "I am checking the details of the CI log for the failed test, but I don't see how the error relates to the code I added; the error is coming from a config builder different than the `LinceConfig`, and it crashes when `self.config.data_files` because is self.config is None. I would appreciate if someone could help m...
1,594,506,920,000
1,594,916,386,000
1,594,916,386,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/383", "html_url": "https://github.com/huggingface/datasets/pull/383", "diff_url": "https://github.com/huggingface/datasets/pull/383.diff", "patch_url": "https://github.com/huggingface/datasets/pull/383.patch" }
Hi, First of all, this library is really cool! Thanks for putting all of this together! This PR contains the [Linguistic Code-switching Evaluation (LinCE) benchmark](https://ritual.uh.edu/lince). As described in the official website (FAQ): > 1. Why do we need LinCE? >LinCE brings 10 code-switching datasets t...
https://api.github.com/repos/huggingface/datasets/issues/383/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/382
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/382/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/382/comments
https://api.github.com/repos/huggingface/datasets/issues/382/events
https://github.com/huggingface/datasets/issues/382
655,290,482
MDU6SXNzdWU2NTUyOTA0ODI=
382
1080
{ "login": "saq194", "id": 60942503, "node_id": "MDQ6VXNlcjYwOTQyNTAz", "avatar_url": "https://avatars.githubusercontent.com/u/60942503?v=4", "gravatar_id": "", "url": "https://api.github.com/users/saq194", "html_url": "https://github.com/saq194", "followers_url": "https://api.github.com/users/saq194/fo...
[]
closed
false
null
[]
null
[]
1,594,506,547,000
1,594,507,778,000
1,594,507,778,000
NONE
null
null
https://api.github.com/repos/huggingface/datasets/issues/382/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/381
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/381/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/381/comments
https://api.github.com/repos/huggingface/datasets/issues/381/events
https://github.com/huggingface/datasets/issues/381
655,277,119
MDU6SXNzdWU2NTUyNzcxMTk=
381
NLp
{ "login": "Spartanthor", "id": 68147610, "node_id": "MDQ6VXNlcjY4MTQ3NjEw", "avatar_url": "https://avatars.githubusercontent.com/u/68147610?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Spartanthor", "html_url": "https://github.com/Spartanthor", "followers_url": "https://api.github.com/...
[]
closed
false
null
[]
null
[]
1,594,500,614,000
1,594,500,639,000
1,594,500,639,000
NONE
null
null
https://api.github.com/repos/huggingface/datasets/issues/381/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/378
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/378/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/378/comments
https://api.github.com/repos/huggingface/datasets/issues/378/events
https://github.com/huggingface/datasets/issues/378
655,226,316
MDU6SXNzdWU2NTUyMjYzMTY=
378
[dataset] Structure of MLQA seems unecessary nested
{ "login": "thomwolf", "id": 7353373, "node_id": "MDQ6VXNlcjczNTMzNzM=", "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "gravatar_id": "", "url": "https://api.github.com/users/thomwolf", "html_url": "https://github.com/thomwolf", "followers_url": "https://api.github.com/users/thomw...
[]
closed
false
null
[]
null
[ "Same for the RACE dataset: https://github.com/huggingface/nlp/blob/master/datasets/race/race.py\r\n\r\nShould we scan all the datasets to remove this pattern of un-necessary nesting?", "You're right, I think we don't need to use the nested dictionary. \r\n" ]
1,594,480,568,000
1,594,829,840,000
1,594,829,840,000
MEMBER
null
null
The features of the MLQA dataset comprise several nested dictionaries with a single element inside (for `questions` and `ids`): https://github.com/huggingface/nlp/blob/master/datasets/mlqa/mlqa.py#L90-L97 Should we keep this @mariamabarham @patrickvonplaten? Was this added for compatibility with tfds? ```python ...
https://api.github.com/repos/huggingface/datasets/issues/378/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/377
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/377/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/377/comments
https://api.github.com/repos/huggingface/datasets/issues/377/events
https://github.com/huggingface/datasets/issues/377
655,215,790
MDU6SXNzdWU2NTUyMTU3OTA=
377
Iyy!!!
{ "login": "ajinomoh", "id": 68154535, "node_id": "MDQ6VXNlcjY4MTU0NTM1", "avatar_url": "https://avatars.githubusercontent.com/u/68154535?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ajinomoh", "html_url": "https://github.com/ajinomoh", "followers_url": "https://api.github.com/users/aji...
[]
closed
false
null
[]
null
[]
1,594,476,667,000
1,594,477,851,000
1,594,477,851,000
NONE
null
null
https://api.github.com/repos/huggingface/datasets/issues/377/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/376
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/376/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/376/comments
https://api.github.com/repos/huggingface/datasets/issues/376/events
https://github.com/huggingface/datasets/issues/376
655,047,826
MDU6SXNzdWU2NTUwNDc4MjY=
376
to_pandas conversion doesn't always work
{ "login": "thomwolf", "id": 7353373, "node_id": "MDQ6VXNlcjczNTMzNzM=", "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "gravatar_id": "", "url": "https://api.github.com/users/thomwolf", "html_url": "https://github.com/thomwolf", "followers_url": "https://api.github.com/users/thomw...
[]
open
false
null
[]
null
[ "**Edit**: other topic previously in this message moved to a new issue: https://github.com/huggingface/nlp/issues/387", "Could you try to update pyarrow to >=0.17.0 ? It should fix the `to_pandas` bug\r\n\r\nAlso I'm not sure that structures like list<struct> are fully supported in the lib (none of the datasets u...
1,594,416,811,000
1,595,239,845,000
null
MEMBER
null
null
For some complex nested types, the conversion from Arrow to python dict through pandas doesn't seem to be possible. Here is an example using the official SQUAD v2 JSON file. This example was found while investigating #373. ```python >>> squad = load_dataset('json', data_files={nlp.Split.TRAIN: ["./train-v2.0....
https://api.github.com/repos/huggingface/datasets/issues/376/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/375
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/375/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/375/comments
https://api.github.com/repos/huggingface/datasets/issues/375/events
https://github.com/huggingface/datasets/issues/375
655,023,307
MDU6SXNzdWU2NTUwMjMzMDc=
375
TypeError when computing bertscore
{ "login": "willywsm1013", "id": 13269577, "node_id": "MDQ6VXNlcjEzMjY5NTc3", "avatar_url": "https://avatars.githubusercontent.com/u/13269577?v=4", "gravatar_id": "", "url": "https://api.github.com/users/willywsm1013", "html_url": "https://github.com/willywsm1013", "followers_url": "https://api.github.c...
[]
open
false
null
[]
null
[ "I am not able to reproduce this issue on my side.\r\nCould you give us more details about the inputs you used ?\r\n\r\nI do get another error though:\r\n```\r\n~/.virtualenvs/hf-datasets/lib/python3.7/site-packages/bert_score/utils.py in bert_cos_score_idf(model, refs, hyps, tokenizer, idf_dict, verbose, batch_siz...
1,594,413,464,000
1,599,490,212,000
null
NONE
null
null
Hi, I installed nlp 0.3.0 via pip, and my python version is 3.7. When I tried to compute bertscore with the code: ``` import nlp bertscore = nlp.load_metric('bertscore') # load hyps and refs ... print (bertscore.compute(hyps, refs, lang='en')) ``` I got the following error. ``` Traceback (most rece...
https://api.github.com/repos/huggingface/datasets/issues/375/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/374
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/374/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/374/comments
https://api.github.com/repos/huggingface/datasets/issues/374/events
https://github.com/huggingface/datasets/pull/374
654,895,066
MDExOlB1bGxSZXF1ZXN0NDQ3NTMxMzUy
374
Add dataset post processing for faiss indexes
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[ "I changed the `wiki_dpr` script to ignore the last 24 examples for now. Hopefully we'll have the full version soon.\r\nThe datasets_infos.json and the data on GCS are updated.\r\n\r\nAnd I also added a check to make sure we don't have post processing resources in sub-directories.", "I added a dummy config that c...
1,594,398,359,000
1,594,647,843,000
1,594,647,841,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/374", "html_url": "https://github.com/huggingface/datasets/pull/374", "diff_url": "https://github.com/huggingface/datasets/pull/374.diff", "patch_url": "https://github.com/huggingface/datasets/pull/374.patch" }
# Post processing of datasets for faiss indexes Now that we can have datasets with embeddings (see `wiki_pr` for example), we can allow users to load the dataset + get the Faiss index that comes with it to do nearest neighbors queries. ## Implementation proposition - Faiss indexes have to be added to the `nlp....
https://api.github.com/repos/huggingface/datasets/issues/374/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/373
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/373/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/373/comments
https://api.github.com/repos/huggingface/datasets/issues/373/events
https://github.com/huggingface/datasets/issues/373
654,845,133
MDU6SXNzdWU2NTQ4NDUxMzM=
373
Segmentation fault when loading local JSON dataset as of #372
{ "login": "vegarab", "id": 24683907, "node_id": "MDQ6VXNlcjI0NjgzOTA3", "avatar_url": "https://avatars.githubusercontent.com/u/24683907?v=4", "gravatar_id": "", "url": "https://api.github.com/users/vegarab", "html_url": "https://github.com/vegarab", "followers_url": "https://api.github.com/users/vegara...
[]
open
false
null
[]
null
[ "I've seen this sort of thing before -- it might help to delete the directory -- I've also noticed that there is an error with the json Dataloader for any data I've tried to load. I've replaced it with this, which skips over the data feature population step:\r\n\r\n\r\n```python\r\nimport os\r\n\r\nimport pyarrow.j...
1,594,393,465,000
1,608,017,240,000
null
CONTRIBUTOR
null
null
The last issue was closed (#369) once the #372 update was merged. However, I'm still not able to load a SQuAD formatted JSON file. Instead of the previously recorded pyarrow error, I now get a segmentation fault. ``` dataset = nlp.load_dataset('json', data_files={nlp.Split.TRAIN: ["./datasets/train-v2.0.json"]}, f...
https://api.github.com/repos/huggingface/datasets/issues/373/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/372
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/372/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/372/comments
https://api.github.com/repos/huggingface/datasets/issues/372/events
https://github.com/huggingface/datasets/pull/372
654,774,420
MDExOlB1bGxSZXF1ZXN0NDQ3NDMzNTA4
372
Make the json script more flexible
{ "login": "thomwolf", "id": 7353373, "node_id": "MDQ6VXNlcjczNTMzNzM=", "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "gravatar_id": "", "url": "https://api.github.com/users/thomwolf", "html_url": "https://github.com/thomwolf", "followers_url": "https://api.github.com/users/thomw...
[]
closed
false
null
[]
null
[]
1,594,386,915,000
1,594,392,727,000
1,594,392,726,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/372", "html_url": "https://github.com/huggingface/datasets/pull/372", "diff_url": "https://github.com/huggingface/datasets/pull/372.diff", "patch_url": "https://github.com/huggingface/datasets/pull/372.patch" }
Fix https://github.com/huggingface/nlp/issues/359 Fix https://github.com/huggingface/nlp/issues/369 JSON script now can accept JSON files containing a single dict with the records as a list in one attribute to the dict (previously it only accepted JSON files containing records as rows of dicts in the file). In t...
https://api.github.com/repos/huggingface/datasets/issues/372/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/371
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/371/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/371/comments
https://api.github.com/repos/huggingface/datasets/issues/371/events
https://github.com/huggingface/datasets/pull/371
654,668,242
MDExOlB1bGxSZXF1ZXN0NDQ3MzQ4NDgw
371
Fix cached file path for metrics with different config names
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[ "Thanks for the fast fix!" ]
1,594,375,344,000
1,594,388,722,000
1,594,388,720,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/371", "html_url": "https://github.com/huggingface/datasets/pull/371", "diff_url": "https://github.com/huggingface/datasets/pull/371.diff", "patch_url": "https://github.com/huggingface/datasets/pull/371.patch" }
The config name was not taken into account to build the cached file path. It should fix #368
https://api.github.com/repos/huggingface/datasets/issues/371/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/370
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/370/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/370/comments
https://api.github.com/repos/huggingface/datasets/issues/370/events
https://github.com/huggingface/datasets/pull/370
654,304,193
MDExOlB1bGxSZXF1ZXN0NDQ3MDU3NTIw
370
Allow indexing Dataset via np.ndarray
{ "login": "jarednielsen", "id": 4564897, "node_id": "MDQ6VXNlcjQ1NjQ4OTc=", "avatar_url": "https://avatars.githubusercontent.com/u/4564897?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jarednielsen", "html_url": "https://github.com/jarednielsen", "followers_url": "https://api.github.com...
[]
closed
false
null
[]
null
[ "Looks like a flaky CI, failed download from S3." ]
1,594,323,795,000
1,594,389,944,000
1,594,389,943,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/370", "html_url": "https://github.com/huggingface/datasets/pull/370", "diff_url": "https://github.com/huggingface/datasets/pull/370.diff", "patch_url": "https://github.com/huggingface/datasets/pull/370.patch" }
https://api.github.com/repos/huggingface/datasets/issues/370/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/369
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/369/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/369/comments
https://api.github.com/repos/huggingface/datasets/issues/369/events
https://github.com/huggingface/datasets/issues/369
654,186,890
MDU6SXNzdWU2NTQxODY4OTA=
369
can't load local dataset: pyarrow.lib.ArrowInvalid: straddling object straddles two block boundaries
{ "login": "vegarab", "id": 24683907, "node_id": "MDQ6VXNlcjI0NjgzOTA3", "avatar_url": "https://avatars.githubusercontent.com/u/24683907?v=4", "gravatar_id": "", "url": "https://api.github.com/users/vegarab", "html_url": "https://github.com/vegarab", "followers_url": "https://api.github.com/users/vegara...
[ { "id": 2067388877, "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug", "name": "dataset bug", "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library" } ]
closed
false
null
[]
null
[ "I am able to reproduce this with the official SQuAD `train-v2.0.json` file downloaded directly from https://rajpurkar.github.io/SQuAD-explorer/", "I am facing this issue in transformers library 3.0.2 while reading a csv using datasets.\r\nIs this fixed in latest version? \r\nI updated the latest version 4.0.1 bu...
1,594,311,413,000
1,608,073,642,000
1,594,392,726,000
CONTRIBUTOR
null
null
Trying to load a local SQuAD-formatted dataset (from a JSON file, about 60MB): ``` dataset = nlp.load_dataset(path='json', data_files={nlp.Split.TRAIN: ["./path/to/file.json"]}) ``` causes ``` Traceback (most recent call last): File "dataloader.py", line 9, in <module> ["./path/to/file.json"]}) File "/...
https://api.github.com/repos/huggingface/datasets/issues/369/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/368
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/368/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/368/comments
https://api.github.com/repos/huggingface/datasets/issues/368/events
https://github.com/huggingface/datasets/issues/368
654,087,251
MDU6SXNzdWU2NTQwODcyNTE=
368
load_metric can't acquire lock anymore
{ "login": "ydshieh", "id": 2521628, "node_id": "MDQ6VXNlcjI1MjE2Mjg=", "avatar_url": "https://avatars.githubusercontent.com/u/2521628?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ydshieh", "html_url": "https://github.com/ydshieh", "followers_url": "https://api.github.com/users/ydshieh/...
[]
closed
false
null
[]
null
[ "I found that, in the same process (or the same interactive session), if I do\r\n\r\nimport nlp\r\n\r\nm1 = nlp.load_metric('glue', 'mrpc')\r\nm2 = nlp.load_metric('glue', 'sst2')\r\n\r\nI will get the same error `ValueError: Cannot acquire lock, caching file might be used by another process, you should setup a uni...
1,594,303,449,000
1,594,388,720,000
1,594,388,720,000
NONE
null
null
I can't load metric (glue) anymore after an error in a previous run. I even removed the whole cache folder `/home/XXX/.cache/huggingface/`, and the issue persisted. What are the steps to fix this? Traceback (most recent call last): File "/home/XXX/miniconda3/envs/ML-DL-py-3.7/lib/python3.7/site-packages/n...
https://api.github.com/repos/huggingface/datasets/issues/368/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/367
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/367/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/367/comments
https://api.github.com/repos/huggingface/datasets/issues/367/events
https://github.com/huggingface/datasets/pull/367
654,012,984
MDExOlB1bGxSZXF1ZXN0NDQ2ODIxNTAz
367
Update Xtreme to add PAWS-X es
{ "login": "mariamabarham", "id": 38249783, "node_id": "MDQ6VXNlcjM4MjQ5Nzgz", "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariamabarham", "html_url": "https://github.com/mariamabarham", "followers_url": "https://api.githu...
[]
closed
false
null
[]
null
[]
1,594,296,877,000
1,594,298,231,000
1,594,298,230,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/367", "html_url": "https://github.com/huggingface/datasets/pull/367", "diff_url": "https://github.com/huggingface/datasets/pull/367.diff", "patch_url": "https://github.com/huggingface/datasets/pull/367.patch" }
This PR adds the `PAWS-X.es` in the Xtreme dataset #362
https://api.github.com/repos/huggingface/datasets/issues/367/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/366
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/366/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/366/comments
https://api.github.com/repos/huggingface/datasets/issues/366/events
https://github.com/huggingface/datasets/pull/366
653,954,896
MDExOlB1bGxSZXF1ZXN0NDQ2NzcyODE2
366
Add quora dataset
{ "login": "ghomasHudson", "id": 13795113, "node_id": "MDQ6VXNlcjEzNzk1MTEz", "avatar_url": "https://avatars.githubusercontent.com/u/13795113?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ghomasHudson", "html_url": "https://github.com/ghomasHudson", "followers_url": "https://api.github.c...
[]
closed
false
null
[]
null
[ "Tests seem to be failing because of pandas", "Kaggle needs authentification to download datasets. We don't have a way to handle that in the lib for now" ]
1,594,290,862,000
1,594,661,721,000
1,594,661,721,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/366", "html_url": "https://github.com/huggingface/datasets/pull/366", "diff_url": "https://github.com/huggingface/datasets/pull/366.diff", "patch_url": "https://github.com/huggingface/datasets/pull/366.patch" }
Added the [Quora question pairs dataset](https://www.quora.com/q/quoradata/First-Quora-Dataset-Release-Question-Pairs). Implementation Notes: - I used the original version provided on the quora website. There's also a [Kaggle competition](https://www.kaggle.com/c/quora-question-pairs) which has a nice train/test sp...
https://api.github.com/repos/huggingface/datasets/issues/366/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/365
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/365/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/365/comments
https://api.github.com/repos/huggingface/datasets/issues/365/events
https://github.com/huggingface/datasets/issues/365
653,845,964
MDU6SXNzdWU2NTM4NDU5NjQ=
365
How to augment data ?
{ "login": "astariul", "id": 43774355, "node_id": "MDQ6VXNlcjQzNzc0MzU1", "avatar_url": "https://avatars.githubusercontent.com/u/43774355?v=4", "gravatar_id": "", "url": "https://api.github.com/users/astariul", "html_url": "https://github.com/astariul", "followers_url": "https://api.github.com/users/ast...
[]
closed
false
null
[]
null
[ "Using batched map is probably the easiest way at the moment.\r\nWhat kind of augmentation would you like to do ?", "Some samples in the dataset are too long, I want to divide them in several samples.", "Using batched map is the way to go then.\r\nWe'll make it clearer in the docs that map could be used for aug...
1,594,281,157,000
1,594,372,327,000
1,594,369,335,000
NONE
null
null
Is there any clean way to augment data ? For now my work-around is to use batched map, like this : ```python def aug(samples): # Simply copy the existing data to have x2 amount of data for k, v in samples.items(): samples[k].extend(v) return samples dataset = dataset.map(aug, batched=T...
https://api.github.com/repos/huggingface/datasets/issues/365/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/364
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/364/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/364/comments
https://api.github.com/repos/huggingface/datasets/issues/364/events
https://github.com/huggingface/datasets/pull/364
653,821,597
MDExOlB1bGxSZXF1ZXN0NDQ2NjY0NzM5
364
add MS MARCO dataset
{ "login": "mariamabarham", "id": 38249783, "node_id": "MDQ6VXNlcjM4MjQ5Nzgz", "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariamabarham", "html_url": "https://github.com/mariamabarham", "followers_url": "https://api.githu...
[]
closed
false
null
[]
null
[ "The dummy data for v2.1 is missing as far as I can see. I think running the dummy data command should work correctly here. ", "Also, it might be that the structure of the dummy data is wrong - looking at `generate_examples` the structure does not look too easy.", "The fact that the dummy data for v2.1 is miss...
1,594,278,679,000
1,596,694,549,000
1,596,694,548,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/364", "html_url": "https://github.com/huggingface/datasets/pull/364", "diff_url": "https://github.com/huggingface/datasets/pull/364.diff", "patch_url": "https://github.com/huggingface/datasets/pull/364.patch" }
This PR adds the MS MARCO dataset as requested in this issue #336. MS mARCO has multiple task including: - Passage and Document Retrieval - Keyphrase Extraction - QA and NLG This PR only adds the 2 versions of the QA and NLG task dataset which was realeased with the original paper here https://arxiv.org/pd...
https://api.github.com/repos/huggingface/datasets/issues/364/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/363
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/363/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/363/comments
https://api.github.com/repos/huggingface/datasets/issues/363/events
https://github.com/huggingface/datasets/pull/363
653,821,172
MDExOlB1bGxSZXF1ZXN0NDQ2NjY0NDIy
363
Adding support for generic multi dimensional tensors and auxillary image data for multimodal datasets
{ "login": "eltoto1219", "id": 14030663, "node_id": "MDQ6VXNlcjE0MDMwNjYz", "avatar_url": "https://avatars.githubusercontent.com/u/14030663?v=4", "gravatar_id": "", "url": "https://api.github.com/users/eltoto1219", "html_url": "https://github.com/eltoto1219", "followers_url": "https://api.github.com/use...
[]
closed
false
null
[]
null
[ "Thank you! I just marked this as a draft PR. It probably would be better to create specific Array2D and Array3D classes as needed instead of a generic MultiArray for now, it should simplify the code a lot too so, I'll update it as such. Also i was meaning to reply earlier, but I wanted to thank you for the testing...
1,594,278,630,000
1,598,263,175,000
1,598,263,175,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/363", "html_url": "https://github.com/huggingface/datasets/pull/363", "diff_url": "https://github.com/huggingface/datasets/pull/363.diff", "patch_url": "https://github.com/huggingface/datasets/pull/363.patch" }
nlp/features.py: The main factory class is MultiArray, every single time this class is called, a corresponding pyarrow extension array and type class is generated (and added to the list of globals for future use) for a given root data type and set of dimensions/shape. I provide examples on working with this in datas...
https://api.github.com/repos/huggingface/datasets/issues/363/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/362
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/362/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/362/comments
https://api.github.com/repos/huggingface/datasets/issues/362/events
https://github.com/huggingface/datasets/issues/362
653,766,245
MDU6SXNzdWU2NTM3NjYyNDU=
362
[dateset subset missing] xtreme paws-x
{ "login": "jerryIsHere", "id": 50871412, "node_id": "MDQ6VXNlcjUwODcxNDEy", "avatar_url": "https://avatars.githubusercontent.com/u/50871412?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jerryIsHere", "html_url": "https://github.com/jerryIsHere", "followers_url": "https://api.github.com/...
[]
closed
false
null
[]
null
[ "You're right, thanks for pointing it out. We will update it " ]
1,594,271,094,000
1,594,298,322,000
1,594,298,322,000
CONTRIBUTOR
null
null
I tried nlp.load_dataset('xtreme', 'PAWS-X.es') but get the value error It turns out that the subset for Spanish is missing https://github.com/google-research-datasets/paws/tree/master/pawsx
https://api.github.com/repos/huggingface/datasets/issues/362/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/361
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/361/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/361/comments
https://api.github.com/repos/huggingface/datasets/issues/361/events
https://github.com/huggingface/datasets/issues/361
653,757,376
MDU6SXNzdWU2NTM3NTczNzY=
361
🐛 [Metrics] ROUGE is non-deterministic
{ "login": "astariul", "id": 43774355, "node_id": "MDQ6VXNlcjQzNzc0MzU1", "avatar_url": "https://avatars.githubusercontent.com/u/43774355?v=4", "gravatar_id": "", "url": "https://api.github.com/users/astariul", "html_url": "https://github.com/astariul", "followers_url": "https://api.github.com/users/ast...
[]
closed
false
null
[]
null
[ "Hi, can you give a full self-contained example to reproduce this behavior?", "> Hi, can you give a full self-contained example to reproduce this behavior?\r\n\r\nThere is a notebook in the post ;)", "> If I run the ROUGE metric 2 times, with same predictions / references, the scores are slightly different.\r\n...
1,594,269,577,000
1,595,288,917,000
1,595,288,917,000
NONE
null
null
If I run the ROUGE metric 2 times, with same predictions / references, the scores are slightly different. Refer to [this Colab notebook](https://colab.research.google.com/drive/1wRssNXgb9ldcp4ulwj-hMJn0ywhDOiDy?usp=sharing) for reproducing the problem. Example of F-score for ROUGE-1, ROUGE-2, ROUGE-L in 2 differe...
https://api.github.com/repos/huggingface/datasets/issues/361/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/360
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/360/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/360/comments
https://api.github.com/repos/huggingface/datasets/issues/360/events
https://github.com/huggingface/datasets/issues/360
653,687,176
MDU6SXNzdWU2NTM2ODcxNzY=
360
[Feature request] Add dataset.ragged_map() function for many-to-many transformations
{ "login": "jarednielsen", "id": 4564897, "node_id": "MDQ6VXNlcjQ1NjQ4OTc=", "avatar_url": "https://avatars.githubusercontent.com/u/4564897?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jarednielsen", "html_url": "https://github.com/jarednielsen", "followers_url": "https://api.github.com...
[]
closed
false
null
[]
null
[ "Actually `map(batched=True)` can already change the size of the dataset.\r\nIt can accept examples of length `N` and returns a batch of length `M` (can be null or greater than `N`).\r\n\r\nI'll make that explicit in the doc that I'm currently writing.", "You're two steps ahead of me :) In my testing, it also wor...
1,594,256,683,000
1,594,323,111,000
1,594,323,111,000
CONTRIBUTOR
null
null
`dataset.map()` enables one-to-one transformations. Input one example and output one example. This is helpful for tokenizing and cleaning individual lines. `dataset.filter()` enables one-to-(one-or-none) transformations. Input one example and output either zero/one example. This is helpful for removing portions from t...
https://api.github.com/repos/huggingface/datasets/issues/360/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/359
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/359/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/359/comments
https://api.github.com/repos/huggingface/datasets/issues/359/events
https://github.com/huggingface/datasets/issues/359
653,656,279
MDU6SXNzdWU2NTM2NTYyNzk=
359
ArrowBasedBuilder _prepare_split parse_schema breaks on nested structures
{ "login": "timothyjlaurent", "id": 2000204, "node_id": "MDQ6VXNlcjIwMDAyMDQ=", "avatar_url": "https://avatars.githubusercontent.com/u/2000204?v=4", "gravatar_id": "", "url": "https://api.github.com/users/timothyjlaurent", "html_url": "https://github.com/timothyjlaurent", "followers_url": "https://api.g...
[]
closed
false
null
[]
null
[ "Hi, it depends on what it is in your `dataset_builder.py` file. Can you share it?\r\n\r\nIf you are just loading `json` files, you can also directly use the `json` script (which will find the schema/features from your JSON structure):\r\n\r\n```python\r\nfrom nlp import load_dataset\r\nds = load_dataset(\"json\", ...
1,594,250,645,000
1,594,392,726,000
1,594,392,726,000
NONE
null
null
I tried using the Json dataloader to load some JSON lines files. but get an exception in the parse_schema function. ``` --------------------------------------------------------------------------- ValueError Traceback (most recent call last) <ipython-input-23-9aecfbee53bd> in <mo...
https://api.github.com/repos/huggingface/datasets/issues/359/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/358
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/358/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/358/comments
https://api.github.com/repos/huggingface/datasets/issues/358/events
https://github.com/huggingface/datasets/pull/358
653,645,121
MDExOlB1bGxSZXF1ZXN0NDQ2NTI0NjQ5
358
Starting to add some real doc
{ "login": "thomwolf", "id": 7353373, "node_id": "MDQ6VXNlcjczNTMzNzM=", "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "gravatar_id": "", "url": "https://api.github.com/users/thomwolf", "html_url": "https://github.com/thomwolf", "followers_url": "https://api.github.com/users/thomw...
[]
closed
false
null
[]
null
[ "Ok this is starting to be really big so it's probably good to merge this first version of the doc and continue in another PR :)\r\n\r\nThis first version of the doc can be explored here: https://2219-250213286-gh.circle-artifacts.com/0/docs/_build/html/index.html" ]
1,594,248,783,000
1,594,720,697,000
1,594,720,695,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/358", "html_url": "https://github.com/huggingface/datasets/pull/358", "diff_url": "https://github.com/huggingface/datasets/pull/358.diff", "patch_url": "https://github.com/huggingface/datasets/pull/358.patch" }
Adding a lot of documentation for: - load a dataset - explore the dataset object - process data with the dataset - add a new dataset script - share a dataset script - full package reference This version of the doc can be explored here: https://2219-250213286-gh.circle-artifacts.com/0/docs/_build/html/index.htm...
https://api.github.com/repos/huggingface/datasets/issues/358/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/357
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/357/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/357/comments
https://api.github.com/repos/huggingface/datasets/issues/357/events
https://github.com/huggingface/datasets/pull/357
653,642,292
MDExOlB1bGxSZXF1ZXN0NDQ2NTIyMzU2
357
Add hashes to cnn_dailymail
{ "login": "jbragg", "id": 2238344, "node_id": "MDQ6VXNlcjIyMzgzNDQ=", "avatar_url": "https://avatars.githubusercontent.com/u/2238344?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jbragg", "html_url": "https://github.com/jbragg", "followers_url": "https://api.github.com/users/jbragg/foll...
[]
closed
false
null
[]
null
[ "Looks you to me :)\r\n\r\nCould you also update the json file that goes with the dataset script by doing \r\n```\r\nnlp-cli test ./datasets/cnn_dailymail --save_infos --all_configs\r\n```\r\nIt will update the features metadata and the size of the dataset with your changes.", "@lhoestq I ran that command.\r\n\r\...
1,594,248,321,000
1,594,649,798,000
1,594,649,798,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/357", "html_url": "https://github.com/huggingface/datasets/pull/357", "diff_url": "https://github.com/huggingface/datasets/pull/357.diff", "patch_url": "https://github.com/huggingface/datasets/pull/357.patch" }
The URL hashes are helpful for comparing results from other sources.
https://api.github.com/repos/huggingface/datasets/issues/357/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/356
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/356/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/356/comments
https://api.github.com/repos/huggingface/datasets/issues/356/events
https://github.com/huggingface/datasets/pull/356
653,537,388
MDExOlB1bGxSZXF1ZXN0NDQ2NDM3MDQ5
356
Add text dataset
{ "login": "jarednielsen", "id": 4564897, "node_id": "MDQ6VXNlcjQ1NjQ4OTc=", "avatar_url": "https://avatars.githubusercontent.com/u/4564897?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jarednielsen", "html_url": "https://github.com/jarednielsen", "followers_url": "https://api.github.com...
[]
closed
false
null
[]
null
[]
1,594,236,113,000
1,594,390,743,000
1,594,390,743,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/356", "html_url": "https://github.com/huggingface/datasets/pull/356", "diff_url": "https://github.com/huggingface/datasets/pull/356.diff", "patch_url": "https://github.com/huggingface/datasets/pull/356.patch" }
Usage: ```python from nlp import load_dataset dset = load_dataset("text", data_files="/path/to/file.txt")["train"] ``` I created a dummy_data.zip which contains three files: `train.txt`, `test.txt`, `dev.txt`. Each of these contains two lines. It passes ```bash RUN_SLOW=1 pytest tests/test_dataset_common...
https://api.github.com/repos/huggingface/datasets/issues/356/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/355
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/355/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/355/comments
https://api.github.com/repos/huggingface/datasets/issues/355/events
https://github.com/huggingface/datasets/issues/355
653,451,013
MDU6SXNzdWU2NTM0NTEwMTM=
355
can't load SNLI dataset
{ "login": "jxmorris12", "id": 13238952, "node_id": "MDQ6VXNlcjEzMjM4OTUy", "avatar_url": "https://avatars.githubusercontent.com/u/13238952?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jxmorris12", "html_url": "https://github.com/jxmorris12", "followers_url": "https://api.github.com/use...
[]
closed
false
null
[]
null
[ "I just added the processed files of `snli` on our google storage, so that when you do `load_dataset` it can download the processed files from there :)\r\n\r\nWe are thinking about having available those processed files for more datasets in the future, because sometimes files aren't available (like for `snli`), or ...
1,594,227,254,000
1,595,049,357,000
1,594,799,941,000
CONTRIBUTOR
null
null
`nlp` seems to load `snli` from some URL based on nlp.stanford.edu. This subdomain is frequently down -- including right now, when I'd like to load `snli` in a Colab notebook, but can't. Is there a plan to move these datasets to huggingface servers for a more stable solution? Btw, here's the stack trace: ``` ...
https://api.github.com/repos/huggingface/datasets/issues/355/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/354
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/354/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/354/comments
https://api.github.com/repos/huggingface/datasets/issues/354/events
https://github.com/huggingface/datasets/pull/354
653,357,617
MDExOlB1bGxSZXF1ZXN0NDQ2MjkyMTc4
354
More faiss control
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[ "> Ok, so we're getting rid of the `FaissGpuOptions`?\r\n\r\nWe support `device=...` because it's simple, but faiss GPU options can be used in so many ways (you can set different gpu options for the different parts of your index for example) that it's probably better to let the user create and configure its index a...
1,594,219,520,000
1,594,288,494,000
1,594,288,491,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/354", "html_url": "https://github.com/huggingface/datasets/pull/354", "diff_url": "https://github.com/huggingface/datasets/pull/354.diff", "patch_url": "https://github.com/huggingface/datasets/pull/354.patch" }
Allow users to specify a faiss index they created themselves, as sometimes indexes can be composite for examples
https://api.github.com/repos/huggingface/datasets/issues/354/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/353
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/353/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/353/comments
https://api.github.com/repos/huggingface/datasets/issues/353/events
https://github.com/huggingface/datasets/issues/353
653,250,611
MDU6SXNzdWU2NTMyNTA2MTE=
353
[Dataset requests] New datasets for Text Classification
{ "login": "thomwolf", "id": 7353373, "node_id": "MDQ6VXNlcjczNTMzNzM=", "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "gravatar_id": "", "url": "https://api.github.com/users/thomwolf", "html_url": "https://github.com/thomwolf", "followers_url": "https://api.github.com/users/thomw...
[ { "id": 1935892884, "node_id": "MDU6TGFiZWwxOTM1ODkyODg0", "url": "https://api.github.com/repos/huggingface/datasets/labels/help%20wanted", "name": "help wanted", "color": "008672", "default": true, "description": "Extra attention is needed" }, { "id": 2067376369, "node_id": ...
open
false
null
[]
null
[ "Pinging @mariamabarham as well", "- `nlp` has MR! It's called `rotten_tomatoes`\r\n- SST is part of GLUE, or is that just SST-2?\r\n- `nlp` also has `ag_news`, a popular news classification dataset\r\n\r\nI'd also like to see:\r\n- the Yahoo Answers topic classification dataset\r\n- the Kaggle Fake News classifi...
1,594,210,678,000
1,603,165,283,000
null
MEMBER
null
null
We are missing a few datasets for Text Classification which is an important field. Namely, it would be really nice to add: - TREC-6 dataset (see here for instance: https://pytorchnlp.readthedocs.io/en/latest/source/torchnlp.datasets.html#torchnlp.datasets.trec_dataset) **[done]** - Yelp-5 - Movie review (Movie R...
https://api.github.com/repos/huggingface/datasets/issues/353/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/352
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/352/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/352/comments
https://api.github.com/repos/huggingface/datasets/issues/352/events
https://github.com/huggingface/datasets/pull/352
653,128,883
MDExOlB1bGxSZXF1ZXN0NDQ2MTA1Mjky
352
🐛[BugFix]fix seqeval
{ "login": "AlongWY", "id": 20281571, "node_id": "MDQ6VXNlcjIwMjgxNTcx", "avatar_url": "https://avatars.githubusercontent.com/u/20281571?v=4", "gravatar_id": "", "url": "https://api.github.com/users/AlongWY", "html_url": "https://github.com/AlongWY", "followers_url": "https://api.github.com/users/AlongW...
[]
closed
false
null
[]
null
[ "I think this is good but can you detail a bit the behavior before and after your fix?", "examples:\r\n\r\ninput: `['B', 'I', 'I', 'O', 'B', 'I']`\r\nbefore: `[('B', 0, 0), ('I', 1, 2), ('B', 4, 4), ('I', 5, 5)]`\r\nafter: `[('_', 0, 2), ('_', 4, 5)]`\r\n\r\ninput: `['B-ARGM-LOC', 'I-ARGM-LOC', 'I-ARGM-LOC', 'O',...
1,594,199,532,000
1,594,888,006,000
1,594,888,006,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/352", "html_url": "https://github.com/huggingface/datasets/pull/352", "diff_url": "https://github.com/huggingface/datasets/pull/352.diff", "patch_url": "https://github.com/huggingface/datasets/pull/352.patch" }
Fix seqeval process labels such as 'B', 'B-ARGM-LOC'
https://api.github.com/repos/huggingface/datasets/issues/352/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/351
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/351/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/351/comments
https://api.github.com/repos/huggingface/datasets/issues/351/events
https://github.com/huggingface/datasets/pull/351
652,424,048
MDExOlB1bGxSZXF1ZXN0NDQ1NDk0NTE4
351
add pandas dataset
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,594,136,287,000
1,594,217,716,000
1,594,217,715,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/351", "html_url": "https://github.com/huggingface/datasets/pull/351", "diff_url": "https://github.com/huggingface/datasets/pull/351.diff", "patch_url": "https://github.com/huggingface/datasets/pull/351.patch" }
Create a dataset from serialized pandas dataframes. Usage: ```python from nlp import load_dataset dset = load_dataset("pandas", data_files="df.pkl")["train"] ```
https://api.github.com/repos/huggingface/datasets/issues/351/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/350
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/350/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/350/comments
https://api.github.com/repos/huggingface/datasets/issues/350/events
https://github.com/huggingface/datasets/pull/350
652,398,691
MDExOlB1bGxSZXF1ZXN0NDQ1NDczODYz
350
add from_pandas and from_dict
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,594,134,233,000
1,594,217,673,000
1,594,217,672,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/350", "html_url": "https://github.com/huggingface/datasets/pull/350", "diff_url": "https://github.com/huggingface/datasets/pull/350.diff", "patch_url": "https://github.com/huggingface/datasets/pull/350.patch" }
I added two new methods to the `Dataset` class: - `from_pandas()` to create a dataset from a pandas dataframe - `from_dict()` to create a dataset from a dictionary (keys = columns) It uses the `pa.Table.from_pandas` and `pa.Table.from_pydict` funcitons to do so. It is also possible to specify the features types v...
https://api.github.com/repos/huggingface/datasets/issues/350/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/349
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/349/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/349/comments
https://api.github.com/repos/huggingface/datasets/issues/349/events
https://github.com/huggingface/datasets/pull/349
652,231,571
MDExOlB1bGxSZXF1ZXN0NDQ1MzQwMTQ1
349
Hyperpartisan news detection
{ "login": "ghomasHudson", "id": 13795113, "node_id": "MDQ6VXNlcjEzNzk1MTEz", "avatar_url": "https://avatars.githubusercontent.com/u/13795113?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ghomasHudson", "html_url": "https://github.com/ghomasHudson", "followers_url": "https://api.github.c...
[]
closed
false
null
[]
null
[ "Thank you so much for working on this! This is awesome!\r\n\r\nHow much would it help you if we would remove the manual request?\r\n\r\nWe are naturally interested in getting some broad idea of how many people and who are using our dataset. But if you consider hosting the dataset yourself, I would rather remove th...
1,594,119,997,000
1,594,154,847,000
1,594,133,831,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/349", "html_url": "https://github.com/huggingface/datasets/pull/349", "diff_url": "https://github.com/huggingface/datasets/pull/349.diff", "patch_url": "https://github.com/huggingface/datasets/pull/349.patch" }
Adding the hyperpartisan news detection dataset from PAN. This contains news article text, labelled with whether they're hyper-partisan and why kinds of biases they display. Implementation notes: - As with many PAN tasks, the data is hosted on [Zenodo](https://zenodo.org/record/1489920) and must be requested before...
https://api.github.com/repos/huggingface/datasets/issues/349/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/348
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/348/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/348/comments
https://api.github.com/repos/huggingface/datasets/issues/348/events
https://github.com/huggingface/datasets/pull/348
652,158,308
MDExOlB1bGxSZXF1ZXN0NDQ1MjgwNjk3
348
Add OSCAR dataset
{ "login": "pjox", "id": 635220, "node_id": "MDQ6VXNlcjYzNTIyMA==", "avatar_url": "https://avatars.githubusercontent.com/u/635220?v=4", "gravatar_id": "", "url": "https://api.github.com/users/pjox", "html_url": "https://github.com/pjox", "followers_url": "https://api.github.com/users/pjox/followers", ...
[]
closed
false
null
[]
null
[ "@pjox I think the tests don't pass because you haven't provided any dummy data (`dummy_data.zip`).\r\n\r\n ", "> @pjox I think the tests don't pass because you haven't provided any dummy data (`dummy_data.zip`).\r\n\r\nBut can I do the dummy data without running `python nlp-cli test datasets/<your-dataset-folder...
1,594,113,727,000
1,620,079,628,000
1,612,865,959,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/348", "html_url": "https://github.com/huggingface/datasets/pull/348", "diff_url": "https://github.com/huggingface/datasets/pull/348.diff", "patch_url": "https://github.com/huggingface/datasets/pull/348.patch" }
I don't know if tests pass, when I run them it tries to download the whole corpus which is around 3.5TB compressed and I don't have that kind of space. I'll really need some help with it 😅 Thanks!
https://api.github.com/repos/huggingface/datasets/issues/348/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/347
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/347/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/347/comments
https://api.github.com/repos/huggingface/datasets/issues/347/events
https://github.com/huggingface/datasets/issues/347
652,106,567
MDU6SXNzdWU2NTIxMDY1Njc=
347
'cp950' codec error from load_dataset('xtreme', 'tydiqa')
{ "login": "jerryIsHere", "id": 50871412, "node_id": "MDQ6VXNlcjUwODcxNDEy", "avatar_url": "https://avatars.githubusercontent.com/u/50871412?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jerryIsHere", "html_url": "https://github.com/jerryIsHere", "followers_url": "https://api.github.com/...
[ { "id": 2067388877, "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug", "name": "dataset bug", "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library" } ]
closed
false
null
[]
null
[ "This is probably a Windows issue, we need to specify the encoding when `load_dataset()` reads the original CSV file.\r\nTry to find the `open()` statement called by `load_dataset()` and add an `encoding='utf-8'` parameter.\r\nSee issues #242 and #307 ", "It should be in `xtreme.py:L755`:\r\n```python\r\n ...
1,594,109,663,000
1,599,490,305,000
1,599,490,305,000
CONTRIBUTOR
null
null
![image](https://user-images.githubusercontent.com/50871412/86744744-67481680-c06c-11ea-8612-b77eba92a392.png) I guess the error is related to python source encoding issue that my PC is trying to decode the source code with wrong encoding-decoding tools, perhaps : https://www.python.org/dev/peps/pep-0263/ I gues...
https://api.github.com/repos/huggingface/datasets/issues/347/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/346
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/346/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/346/comments
https://api.github.com/repos/huggingface/datasets/issues/346/events
https://github.com/huggingface/datasets/pull/346
652,044,151
MDExOlB1bGxSZXF1ZXN0NDQ1MTg4MTUz
346
Add emotion dataset
{ "login": "lewtun", "id": 26859204, "node_id": "MDQ6VXNlcjI2ODU5MjA0", "avatar_url": "https://avatars.githubusercontent.com/u/26859204?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lewtun", "html_url": "https://github.com/lewtun", "followers_url": "https://api.github.com/users/lewtun/fo...
[]
closed
false
null
[]
null
[ "I've tried it and am getting the same error as you.\r\n\r\nYou could use the text files rather than the pickle:\r\n```\r\nhttps://www.dropbox.com/s/ikkqxfdbdec3fuj/test.txt\r\nhttps://www.dropbox.com/s/1pzkadrvffbqw6o/train.txt\r\nhttps://www.dropbox.com/s/2mzialpsgf9k5l3/val.txt\r\n```\r\n\r\nThen you would get a...
1,594,103,741,000
1,619,162,023,000
1,594,651,178,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/346", "html_url": "https://github.com/huggingface/datasets/pull/346", "diff_url": "https://github.com/huggingface/datasets/pull/346.diff", "patch_url": "https://github.com/huggingface/datasets/pull/346.patch" }
Hello 🤗 team! I am trying to add an emotion classification dataset ([link](https://github.com/dair-ai/emotion_dataset)) to `nlp` but I am a bit stuck about what I should do when the URL for the dataset is not a ZIP file, but just a pickled `pandas.DataFrame` (see [here](https://www.dropbox.com/s/607ptdakxuh5i4s/me...
https://api.github.com/repos/huggingface/datasets/issues/346/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/345
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/345/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/345/comments
https://api.github.com/repos/huggingface/datasets/issues/345/events
https://github.com/huggingface/datasets/issues/345
651,761,201
MDU6SXNzdWU2NTE3NjEyMDE=
345
Supporting documents in ELI5
{ "login": "saverymax", "id": 29262273, "node_id": "MDQ6VXNlcjI5MjYyMjcz", "avatar_url": "https://avatars.githubusercontent.com/u/29262273?v=4", "gravatar_id": "", "url": "https://api.github.com/users/saverymax", "html_url": "https://github.com/saverymax", "followers_url": "https://api.github.com/users/...
[]
closed
false
null
[]
null
[ "Hi @saverymax ! For licensing reasons, the original team was unable to release pre-processed CommonCrawl documents. Instead, they provided a script to re-create them from a CommonCrawl dump, but it unfortunately requires access to a medium-large size cluster:\r\nhttps://github.com/facebookresearch/ELI5#downloading...
1,594,062,853,000
1,603,813,125,000
1,603,813,125,000
NONE
null
null
I was attempting to use the ELI5 dataset, when I realized that huggingface does not provide the supporting documents (the source documents from the common crawl). Without the supporting documents, this makes the dataset about as useful for my project as a block of cheese, or some other more apt metaphor. According to ...
https://api.github.com/repos/huggingface/datasets/issues/345/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/344
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/344/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/344/comments
https://api.github.com/repos/huggingface/datasets/issues/344/events
https://github.com/huggingface/datasets/pull/344
651,495,246
MDExOlB1bGxSZXF1ZXN0NDQ0NzQwMTIw
344
Search qa
{ "login": "mariamabarham", "id": 38249783, "node_id": "MDQ6VXNlcjM4MjQ5Nzgz", "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariamabarham", "html_url": "https://github.com/mariamabarham", "followers_url": "https://api.githu...
[]
closed
false
null
[]
null
[ "Could you rebase from master just to make sure we won't break anything for `fever` pls @mariamabarham ?" ]
1,594,038,196,000
1,594,889,896,000
1,594,889,896,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/344", "html_url": "https://github.com/huggingface/datasets/pull/344", "diff_url": "https://github.com/huggingface/datasets/pull/344.diff", "patch_url": "https://github.com/huggingface/datasets/pull/344.patch" }
This PR adds the Search QA dataset used in **SearchQA: A New Q&A Dataset Augmented with Context from a Search Engine**. The dataset has the following config name: - raw_jeopardy: raw data - train_test_val: which is the splitted version #336
https://api.github.com/repos/huggingface/datasets/issues/344/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/343
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/343/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/343/comments
https://api.github.com/repos/huggingface/datasets/issues/343/events
https://github.com/huggingface/datasets/pull/343
651,419,630
MDExOlB1bGxSZXF1ZXN0NDQ0Njc4NDEw
343
Fix nested tensorflow format
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,594,030,425,000
1,594,041,112,000
1,594,041,111,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/343", "html_url": "https://github.com/huggingface/datasets/pull/343", "diff_url": "https://github.com/huggingface/datasets/pull/343.diff", "patch_url": "https://github.com/huggingface/datasets/pull/343.patch" }
In #339 and #337 we are thinking about adding a way to export datasets to tfrecords. However I noticed that it was not possible to do `dset.set_format("tensorflow")` on datasets with nested features like `squad`. I fixed that using a nested map operations to convert features to `tf.ragged.constant`. I also added ...
https://api.github.com/repos/huggingface/datasets/issues/343/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/342
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/342/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/342/comments
https://api.github.com/repos/huggingface/datasets/issues/342/events
https://github.com/huggingface/datasets/issues/342
651,333,194
MDU6SXNzdWU2NTEzMzMxOTQ=
342
Features should be updated when `map()` changes schema
{ "login": "thomwolf", "id": 7353373, "node_id": "MDQ6VXNlcjczNTMzNzM=", "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "gravatar_id": "", "url": "https://api.github.com/users/thomwolf", "html_url": "https://github.com/thomwolf", "followers_url": "https://api.github.com/users/thomw...
[]
closed
false
null
[]
null
[ "`dataset.column_names` are being updated but `dataset.features` aren't indeed..." ]
1,594,022,603,000
1,595,499,316,000
1,595,499,316,000
MEMBER
null
null
`dataset.map()` can change the schema and column names. We should update the features in this case (with what is possible to infer).
https://api.github.com/repos/huggingface/datasets/issues/342/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/341
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/341/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/341/comments
https://api.github.com/repos/huggingface/datasets/issues/341/events
https://github.com/huggingface/datasets/pull/341
650,611,969
MDExOlB1bGxSZXF1ZXN0NDQ0MDcwMjEx
341
add fever dataset
{ "login": "mariamabarham", "id": 38249783, "node_id": "MDQ6VXNlcjM4MjQ5Nzgz", "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariamabarham", "html_url": "https://github.com/mariamabarham", "followers_url": "https://api.githu...
[]
closed
false
null
[]
null
[]
1,593,784,387,000
1,594,040,628,000
1,594,040,627,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/341", "html_url": "https://github.com/huggingface/datasets/pull/341", "diff_url": "https://github.com/huggingface/datasets/pull/341.diff", "patch_url": "https://github.com/huggingface/datasets/pull/341.patch" }
This PR add the FEVER dataset https://fever.ai/ used in with the paper: FEVER: a large-scale dataset for Fact Extraction and VERification (https://arxiv.org/pdf/1803.05355.pdf). #336
https://api.github.com/repos/huggingface/datasets/issues/341/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/340
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/340/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/340/comments
https://api.github.com/repos/huggingface/datasets/issues/340/events
https://github.com/huggingface/datasets/pull/340
650,533,920
MDExOlB1bGxSZXF1ZXN0NDQ0MDA2Nzcy
340
Update cfq.py
{ "login": "brainshawn", "id": 4437290, "node_id": "MDQ6VXNlcjQ0MzcyOTA=", "avatar_url": "https://avatars.githubusercontent.com/u/4437290?v=4", "gravatar_id": "", "url": "https://api.github.com/users/brainshawn", "html_url": "https://github.com/brainshawn", "followers_url": "https://api.github.com/users...
[]
closed
false
null
[]
null
[ "Thanks @brainshawn for this update" ]
1,593,775,399,000
1,593,779,630,000
1,593,779,630,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/340", "html_url": "https://github.com/huggingface/datasets/pull/340", "diff_url": "https://github.com/huggingface/datasets/pull/340.diff", "patch_url": "https://github.com/huggingface/datasets/pull/340.patch" }
Make the dataset name consistent with in the paper: Compositional Freebase Question => Compositional Freebase Questions.
https://api.github.com/repos/huggingface/datasets/issues/340/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/339
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/339/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/339/comments
https://api.github.com/repos/huggingface/datasets/issues/339/events
https://github.com/huggingface/datasets/pull/339
650,156,468
MDExOlB1bGxSZXF1ZXN0NDQzNzAyNTcw
339
Add dataset.export() to TFRecords
{ "login": "jarednielsen", "id": 4564897, "node_id": "MDQ6VXNlcjQ1NjQ4OTc=", "avatar_url": "https://avatars.githubusercontent.com/u/4564897?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jarednielsen", "html_url": "https://github.com/jarednielsen", "followers_url": "https://api.github.com...
[]
closed
false
null
[]
null
[ "Really cool @jarednielsen !\r\nDo you think we can make it work with dataset with nested features like `squad` ?\r\n\r\nI just did a PR to fix `.set_format` for datasets with nested features, but as soon as it's merged we could try to make the conversion work on a dataset like `squad`.", "For datasets with neste...
1,593,717,987,000
1,595,409,372,000
1,595,409,372,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/339", "html_url": "https://github.com/huggingface/datasets/pull/339", "diff_url": "https://github.com/huggingface/datasets/pull/339.diff", "patch_url": "https://github.com/huggingface/datasets/pull/339.patch" }
Fixes https://github.com/huggingface/nlp/issues/337 Some design decisions: - Simplified the function API to not handle sharding. It writes the entire dataset as a single TFRecord file. This simplifies the function logic and users can use other functions (`select`, `shard`, etc) to handle custom sharding or splitt...
https://api.github.com/repos/huggingface/datasets/issues/339/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/338
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/338/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/338/comments
https://api.github.com/repos/huggingface/datasets/issues/338/events
https://github.com/huggingface/datasets/pull/338
650,057,253
MDExOlB1bGxSZXF1ZXN0NDQzNjIxMTEx
338
Run `make style`
{ "login": "jarednielsen", "id": 4564897, "node_id": "MDQ6VXNlcjQ1NjQ4OTc=", "avatar_url": "https://avatars.githubusercontent.com/u/4564897?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jarednielsen", "html_url": "https://github.com/jarednielsen", "followers_url": "https://api.github.com...
[]
closed
false
null
[]
null
[]
1,593,706,787,000
1,593,712,990,000
1,593,712,990,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/338", "html_url": "https://github.com/huggingface/datasets/pull/338", "diff_url": "https://github.com/huggingface/datasets/pull/338.diff", "patch_url": "https://github.com/huggingface/datasets/pull/338.patch" }
These files get changed when I run `make style` on an unrelated PR. Upstreaming these changes so development on a different branch can be easier.
https://api.github.com/repos/huggingface/datasets/issues/338/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/337
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/337/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/337/comments
https://api.github.com/repos/huggingface/datasets/issues/337/events
https://github.com/huggingface/datasets/issues/337
650,035,887
MDU6SXNzdWU2NTAwMzU4ODc=
337
[Feature request] Export Arrow dataset to TFRecords
{ "login": "jarednielsen", "id": 4564897, "node_id": "MDQ6VXNlcjQ1NjQ4OTc=", "avatar_url": "https://avatars.githubusercontent.com/u/4564897?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jarednielsen", "html_url": "https://github.com/jarednielsen", "followers_url": "https://api.github.com...
[]
closed
false
null
[]
null
[]
1,593,704,832,000
1,595,409,372,000
1,595,409,372,000
CONTRIBUTOR
null
null
The TFRecord generation process is error-prone and requires complex separate Python scripts to download and preprocess the data. I propose to combine the user-friendly features of `nlp` with the speed and efficiency of TFRecords. Sample API: ```python # use these existing methods ds = load_dataset("wikitext", "wik...
https://api.github.com/repos/huggingface/datasets/issues/337/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/336
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/336/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/336/comments
https://api.github.com/repos/huggingface/datasets/issues/336/events
https://github.com/huggingface/datasets/issues/336
649,914,203
MDU6SXNzdWU2NDk5MTQyMDM=
336
[Dataset requests] New datasets for Open Question Answering
{ "login": "thomwolf", "id": 7353373, "node_id": "MDQ6VXNlcjczNTMzNzM=", "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "gravatar_id": "", "url": "https://api.github.com/users/thomwolf", "html_url": "https://github.com/thomwolf", "followers_url": "https://api.github.com/users/thomw...
[ { "id": 1935892884, "node_id": "MDU6TGFiZWwxOTM1ODkyODg0", "url": "https://api.github.com/repos/huggingface/datasets/labels/help%20wanted", "name": "help wanted", "color": "008672", "default": true, "description": "Extra attention is needed" }, { "id": 2067376369, "node_id": ...
closed
false
{ "login": "mariamabarham", "id": 38249783, "node_id": "MDQ6VXNlcjM4MjQ5Nzgz", "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariamabarham", "html_url": "https://github.com/mariamabarham", "followers_url": "https://api.githu...
[ { "login": "mariamabarham", "id": 38249783, "node_id": "MDQ6VXNlcjM4MjQ5Nzgz", "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariamabarham", "html_url": "https://github.com/mariamabarham", "followers_url"...
null
[]
1,593,694,983,000
1,594,890,262,000
1,594,890,262,000
MEMBER
null
null
We are still a few datasets missing for Open-Question Answering which is currently a field in strong development. Namely, it would be really nice to add: - WebQuestions (Berant et al., 2013) [done] - CuratedTrec (Baudis et al. 2015) [not open-source] - MS-MARCO (NGuyen et al. 2016) [done] - SearchQA (Dunn et al....
https://api.github.com/repos/huggingface/datasets/issues/336/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/335
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/335/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/335/comments
https://api.github.com/repos/huggingface/datasets/issues/335/events
https://github.com/huggingface/datasets/pull/335
649,765,179
MDExOlB1bGxSZXF1ZXN0NDQzMzgwMjI1
335
BioMRC Dataset presented in BioNLP 2020 ACL Workshop
{ "login": "PetrosStav", "id": 15162021, "node_id": "MDQ6VXNlcjE1MTYyMDIx", "avatar_url": "https://avatars.githubusercontent.com/u/15162021?v=4", "gravatar_id": "", "url": "https://api.github.com/users/PetrosStav", "html_url": "https://github.com/PetrosStav", "followers_url": "https://api.github.com/use...
[]
closed
false
null
[]
null
[ "I fixed the issues that you pointed out, re-run all the test and pushed the fixed code :-)", "```\r\n=================================== FAILURES ===================================\r\n___________________ AWSDatasetTest.test_load_dataset_pandas ____________________\r\n\r\nself = <tests.test_dataset_common.AWSDat...
1,593,680,621,000
1,594,800,127,000
1,594,800,127,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/335", "html_url": "https://github.com/huggingface/datasets/pull/335", "diff_url": "https://github.com/huggingface/datasets/pull/335.diff", "patch_url": "https://github.com/huggingface/datasets/pull/335.patch" }
https://api.github.com/repos/huggingface/datasets/issues/335/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/334
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/334/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/334/comments
https://api.github.com/repos/huggingface/datasets/issues/334/events
https://github.com/huggingface/datasets/pull/334
649,661,791
MDExOlB1bGxSZXF1ZXN0NDQzMjk1NjQ0
334
Add dataset.shard() method
{ "login": "jarednielsen", "id": 4564897, "node_id": "MDQ6VXNlcjQ1NjQ4OTc=", "avatar_url": "https://avatars.githubusercontent.com/u/4564897?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jarednielsen", "html_url": "https://github.com/jarednielsen", "followers_url": "https://api.github.com...
[]
closed
false
null
[]
null
[ "Great, done!" ]
1,593,669,919,000
1,594,038,936,000
1,594,038,936,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/334", "html_url": "https://github.com/huggingface/datasets/pull/334", "diff_url": "https://github.com/huggingface/datasets/pull/334.diff", "patch_url": "https://github.com/huggingface/datasets/pull/334.patch" }
Fixes https://github.com/huggingface/nlp/issues/312
https://api.github.com/repos/huggingface/datasets/issues/334/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/333
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/333/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/333/comments
https://api.github.com/repos/huggingface/datasets/issues/333/events
https://github.com/huggingface/datasets/pull/333
649,236,516
MDExOlB1bGxSZXF1ZXN0NDQyOTE1NDQ0
333
fix variable name typo
{ "login": "stas00", "id": 10676103, "node_id": "MDQ6VXNlcjEwNjc2MTAz", "avatar_url": "https://avatars.githubusercontent.com/u/10676103?v=4", "gravatar_id": "", "url": "https://api.github.com/users/stas00", "html_url": "https://github.com/stas00", "followers_url": "https://api.github.com/users/stas00/fo...
[]
closed
false
null
[]
null
[ "Good catch :)\r\nI think there is another occurence that needs to be fixed in the second gist (line 4924 of the notebook file):\r\n```python\r\nbleu = nlp.load_metric(...)\r\n```", "Was fixed in e16f79b5f7fc12a6a30c777722be46897a272e6f\r\nClosing it." ]
1,593,630,830,000
1,595,605,411,000
1,595,579,536,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/333", "html_url": "https://github.com/huggingface/datasets/pull/333", "diff_url": "https://github.com/huggingface/datasets/pull/333.diff", "patch_url": "https://github.com/huggingface/datasets/pull/333.patch" }
https://api.github.com/repos/huggingface/datasets/issues/333/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/332
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/332/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/332/comments
https://api.github.com/repos/huggingface/datasets/issues/332/events
https://github.com/huggingface/datasets/pull/332
649,140,135
MDExOlB1bGxSZXF1ZXN0NDQyODMwMzMz
332
Add wiki_dpr
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[ "The two configurations don't have the same sizes, I may change that so that they both have 21015300 examples for convenience, even though it's supposed to have 21015324 examples in total.\r\n\r\nOne configuration only has 21015300 examples because it seems that the embeddings of the last 24 examples are missing.",...
1,593,623,520,000
1,594,038,077,000
1,594,038,076,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/332", "html_url": "https://github.com/huggingface/datasets/pull/332", "diff_url": "https://github.com/huggingface/datasets/pull/332.diff", "patch_url": "https://github.com/huggingface/datasets/pull/332.patch" }
Presented in the [Dense Passage Retrieval paper](https://arxiv.org/pdf/2004.04906.pdf), this dataset consists in 21M passages from the english wikipedia along with their 768-dim embeddings computed using DPR's context encoder. Note on the implementation: - There are two configs: with and without the embeddings (73G...
https://api.github.com/repos/huggingface/datasets/issues/332/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/331
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/331/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/331/comments
https://api.github.com/repos/huggingface/datasets/issues/331/events
https://github.com/huggingface/datasets/issues/331
648,533,199
MDU6SXNzdWU2NDg1MzMxOTk=
331
Loading CNN/Daily Mail dataset produces `nlp.utils.info_utils.NonMatchingSplitsSizesError`
{ "login": "jxmorris12", "id": 13238952, "node_id": "MDQ6VXNlcjEzMjM4OTUy", "avatar_url": "https://avatars.githubusercontent.com/u/13238952?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jxmorris12", "html_url": "https://github.com/jxmorris12", "followers_url": "https://api.github.com/use...
[ { "id": 2067388877, "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug", "name": "dataset bug", "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library" } ]
closed
false
null
[]
null
[ "I couldn't reproduce on my side.\r\nIt looks like you were not able to generate all the examples, and you have the problem for each split train-test-validation.\r\nCould you try to enable logging, try again and send the logs ?\r\n```python\r\nimport logging\r\nlogging.basicConfig(level=logging.INFO)\r\n```", "he...
1,593,555,693,000
1,594,299,820,000
1,594,299,820,000
CONTRIBUTOR
null
null
``` >>> import nlp >>> nlp.load_dataset('cnn_dailymail', '3.0.0') Downloading and preparing dataset cnn_dailymail/3.0.0 (download: 558.32 MiB, generated: 1.26 GiB, total: 1.81 GiB) to /u/jm8wx/.cache/huggingface/datasets/cnn_dailymail/3.0.0/3.0.0... Traceback (most recent call last): File "<stdin>", line 1, in...
https://api.github.com/repos/huggingface/datasets/issues/331/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/330
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/330/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/330/comments
https://api.github.com/repos/huggingface/datasets/issues/330/events
https://github.com/huggingface/datasets/pull/330
648,525,720
MDExOlB1bGxSZXF1ZXN0NDQyMzIxMjEw
330
Doc red
{ "login": "ghomasHudson", "id": 13795113, "node_id": "MDQ6VXNlcjEzNzk1MTEz", "avatar_url": "https://avatars.githubusercontent.com/u/13795113?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ghomasHudson", "html_url": "https://github.com/ghomasHudson", "followers_url": "https://api.github.c...
[]
closed
false
null
[]
null
[]
1,593,554,731,000
1,594,037,439,000
1,593,952,049,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/330", "html_url": "https://github.com/huggingface/datasets/pull/330", "diff_url": "https://github.com/huggingface/datasets/pull/330.diff", "patch_url": "https://github.com/huggingface/datasets/pull/330.patch" }
Adding [DocRED](https://github.com/thunlp/DocRED) - a relation extraction dataset which tests document-level RE. A few implementation notes: - There are 2 separate versions of the training set - *annotated* and *distant*. Instead of `nlp.Split.Train` I've used the splits `"train_annotated"` and `"train_distant"` to ...
https://api.github.com/repos/huggingface/datasets/issues/330/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/329
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/329/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/329/comments
https://api.github.com/repos/huggingface/datasets/issues/329/events
https://github.com/huggingface/datasets/issues/329
648,446,979
MDU6SXNzdWU2NDg0NDY5Nzk=
329
[Bug] FileLock dependency incompatible with filesystem
{ "login": "jarednielsen", "id": 4564897, "node_id": "MDQ6VXNlcjQ1NjQ4OTc=", "avatar_url": "https://avatars.githubusercontent.com/u/4564897?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jarednielsen", "html_url": "https://github.com/jarednielsen", "followers_url": "https://api.github.com...
[]
closed
false
null
[]
null
[ "Hi, can you give details on your environment/os/packages versions/etc?", "Environment is Ubuntu 18.04, Python 3.7.5, nlp==0.3.0, filelock=3.0.12.\r\n\r\nThe external volume is Amazon FSx for Lustre, and it by default creates files with limited permissions. My working theory is that FileLock creates a lockfile th...
1,593,546,331,000
1,593,586,558,000
1,593,552,786,000
CONTRIBUTOR
null
null
I'm downloading a dataset successfully with `load_dataset("wikitext", "wikitext-2-raw-v1")` But when I attempt to cache it on an external volume, it hangs indefinitely: `load_dataset("wikitext", "wikitext-2-raw-v1", cache_dir="/fsx") # /fsx is an external volume mount` The filesystem when hanging looks like thi...
https://api.github.com/repos/huggingface/datasets/issues/329/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/328
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/328/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/328/comments
https://api.github.com/repos/huggingface/datasets/issues/328/events
https://github.com/huggingface/datasets/issues/328
648,326,841
MDU6SXNzdWU2NDgzMjY4NDE=
328
Fork dataset
{ "login": "timothyjlaurent", "id": 2000204, "node_id": "MDQ6VXNlcjIwMDAyMDQ=", "avatar_url": "https://avatars.githubusercontent.com/u/2000204?v=4", "gravatar_id": "", "url": "https://api.github.com/users/timothyjlaurent", "html_url": "https://github.com/timothyjlaurent", "followers_url": "https://api.g...
[]
closed
false
null
[]
null
[ "To be able to generate the Arrow dataset you need to either use our csv or json utilities `load_dataset(\"json\", data_files=my_json_files)` OR write your own custom dataset script (you can find some inspiration from the [squad](https://github.com/huggingface/nlp/blob/master/datasets/squad/squad.py) script for exa...
1,593,535,373,000
1,594,071,839,000
1,594,071,839,000
NONE
null
null
We have a multi-task learning model training I'm trying to convert to using the Arrow-based nlp dataset. We're currently training a custom TensorFlow model but the nlp paradigm should be a bridge for us to be able to use the wealth of pre-trained models in Transformers. Our preprocessing flow parses raw text and...
https://api.github.com/repos/huggingface/datasets/issues/328/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/327
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/327/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/327/comments
https://api.github.com/repos/huggingface/datasets/issues/327/events
https://github.com/huggingface/datasets/pull/327
648,312,858
MDExOlB1bGxSZXF1ZXN0NDQyMTQyOTQw
327
set seed for suffling tests
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,593,534,094,000
1,593,678,845,000
1,593,678,844,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/327", "html_url": "https://github.com/huggingface/datasets/pull/327", "diff_url": "https://github.com/huggingface/datasets/pull/327.diff", "patch_url": "https://github.com/huggingface/datasets/pull/327.patch" }
Some tests were randomly failing because of a missing seed in a test for `train_test_split(shuffle=True)`
https://api.github.com/repos/huggingface/datasets/issues/327/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/326
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/326/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/326/comments
https://api.github.com/repos/huggingface/datasets/issues/326/events
https://github.com/huggingface/datasets/issues/326
648,126,103
MDU6SXNzdWU2NDgxMjYxMDM=
326
Large dataset in Squad2-format
{ "login": "flozi00", "id": 47894090, "node_id": "MDQ6VXNlcjQ3ODk0MDkw", "avatar_url": "https://avatars.githubusercontent.com/u/47894090?v=4", "gravatar_id": "", "url": "https://api.github.com/users/flozi00", "html_url": "https://github.com/flozi00", "followers_url": "https://api.github.com/users/flozi0...
[]
closed
false
null
[]
null
[ "I'm pretty sure you can get some inspiration from the squad_v2 script. It looks like the dataset is quite big so it will take some time for the users to generate it, but it should be reasonable.\r\n\r\nAlso you are saying that you are still making the dataset grow in size right ?\r\nIt's probably good practice to ...
1,593,519,539,000
1,594,285,310,000
1,594,285,310,000
NONE
null
null
At the moment we are building an large question answering dataset and think about sharing it with the huggingface community. Caused the computing power we splitted it into multiple tiles, but they are all in the same format. Right now the most important facts about are this: - Contexts: 1.047.671 - questions: 1.677...
https://api.github.com/repos/huggingface/datasets/issues/326/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/325
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/325/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/325/comments
https://api.github.com/repos/huggingface/datasets/issues/325/events
https://github.com/huggingface/datasets/pull/325
647,601,592
MDExOlB1bGxSZXF1ZXN0NDQxNTk3NTgw
325
Add SQuADShifts dataset
{ "login": "millerjohnp", "id": 8953195, "node_id": "MDQ6VXNlcjg5NTMxOTU=", "avatar_url": "https://avatars.githubusercontent.com/u/8953195?v=4", "gravatar_id": "", "url": "https://api.github.com/users/millerjohnp", "html_url": "https://github.com/millerjohnp", "followers_url": "https://api.github.com/us...
[]
closed
false
null
[]
null
[ "Very cool to have this dataset, thank you for adding it :)" ]
1,593,457,876,000
1,593,536,851,000
1,593,536,851,000
CONTRIBUTOR
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/325", "html_url": "https://github.com/huggingface/datasets/pull/325", "diff_url": "https://github.com/huggingface/datasets/pull/325.diff", "patch_url": "https://github.com/huggingface/datasets/pull/325.patch" }
This PR adds the four new variants of the SQuAD dataset used in [The Effect of Natural Distribution Shift on Question Answering Models](https://arxiv.org/abs/2004.14444) to facilitate evaluating model robustness to distribution shift.
https://api.github.com/repos/huggingface/datasets/issues/325/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/324
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/324/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/324/comments
https://api.github.com/repos/huggingface/datasets/issues/324/events
https://github.com/huggingface/datasets/issues/324
647,525,725
MDU6SXNzdWU2NDc1MjU3MjU=
324
Error when calculating glue score
{ "login": "D-i-l-r-u-k-s-h-i", "id": 47185867, "node_id": "MDQ6VXNlcjQ3MTg1ODY3", "avatar_url": "https://avatars.githubusercontent.com/u/47185867?v=4", "gravatar_id": "", "url": "https://api.github.com/users/D-i-l-r-u-k-s-h-i", "html_url": "https://github.com/D-i-l-r-u-k-s-h-i", "followers_url": "https...
[]
closed
false
null
[]
null
[ "The glue metric for cola is a metric for classification. It expects label ids as integers as inputs.", "I want to evaluate a sentence pair whether they are semantically equivalent, so I used MRPC and it gives the same error, does that mean we have to encode the sentences and parse as input?\r\n\r\nusing BertToke...
1,593,449,628,000
1,594,286,014,000
1,594,286,014,000
NONE
null
null
I was trying glue score along with other metrics here. But glue gives me this error; ``` import nlp glue_metric = nlp.load_metric('glue',name="cola") glue_score = glue_metric.compute(predictions, references) ``` ``` --------------------------------------------------------------------------- --------------...
https://api.github.com/repos/huggingface/datasets/issues/324/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/323
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/323/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/323/comments
https://api.github.com/repos/huggingface/datasets/issues/323/events
https://github.com/huggingface/datasets/pull/323
647,521,308
MDExOlB1bGxSZXF1ZXN0NDQxNTMxOTY3
323
Add package path to sys when downloading package as github archive
{ "login": "yjernite", "id": 10469459, "node_id": "MDQ6VXNlcjEwNDY5NDU5", "avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4", "gravatar_id": "", "url": "https://api.github.com/users/yjernite", "html_url": "https://github.com/yjernite", "followers_url": "https://api.github.com/users/yje...
[]
closed
false
null
[]
null
[ "Sorry for the long diff, everything after the imports comes from `black` for code quality :/ ", " I think it's fine and I can't think of another way to make the import work anyways.\r\n\r\nMaybe we can have the `sys.path` behavior inside `prepare_module` instead ? Currently it seems to come out of nowhere in the...
1,593,449,161,000
1,596,117,623,000
1,596,117,623,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/323", "html_url": "https://github.com/huggingface/datasets/pull/323", "diff_url": "https://github.com/huggingface/datasets/pull/323.diff", "patch_url": "https://github.com/huggingface/datasets/pull/323.patch" }
This fixes the `coval.py` metric so that imports within the downloaded module work correctly. We can use a similar trick to add the BLEURT metric (@ankparikh) @thomwolf not sure how you feel about adding to the `PYTHONPATH` from the script. This is the only way I could make it work with my understanding of `importli...
https://api.github.com/repos/huggingface/datasets/issues/323/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/322
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/322/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/322/comments
https://api.github.com/repos/huggingface/datasets/issues/322/events
https://github.com/huggingface/datasets/pull/322
647,483,850
MDExOlB1bGxSZXF1ZXN0NDQxNTAyMjc2
322
output nested dict in get_nearest_examples
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,593,445,667,000
1,593,678,813,000
1,593,678,812,000
MEMBER
null
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/322", "html_url": "https://github.com/huggingface/datasets/pull/322", "diff_url": "https://github.com/huggingface/datasets/pull/322.diff", "patch_url": "https://github.com/huggingface/datasets/pull/322.patch" }
As we are using a columnar format like arrow as the backend for datasets, we expect to have a dictionary of columns when we slice a dataset like in this example: ```python my_examples = dataset[0:10] print(type(my_examples)) # >>> dict print(my_examples["my_column"][0] # >>> this is the first element of the colum...
https://api.github.com/repos/huggingface/datasets/issues/322/timeline
null
true
https://api.github.com/repos/huggingface/datasets/issues/321
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/321/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/321/comments
https://api.github.com/repos/huggingface/datasets/issues/321/events
https://github.com/huggingface/datasets/issues/321
647,271,526
MDU6SXNzdWU2NDcyNzE1MjY=
321
ERROR:root:mwparserfromhell
{ "login": "Shiro-LK", "id": 26505641, "node_id": "MDQ6VXNlcjI2NTA1NjQx", "avatar_url": "https://avatars.githubusercontent.com/u/26505641?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Shiro-LK", "html_url": "https://github.com/Shiro-LK", "followers_url": "https://api.github.com/users/Shi...
[ { "id": 2067388877, "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug", "name": "dataset bug", "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library" } ]
open
false
null
[]
null
[ "It looks like it comes from `mwparserfromhell`.\r\n\r\nWould it be possible to get the bad `section` that causes this issue ? The `section` string is from `datasets/wikipedia.py:L548` ? You could just add a `try` statement and print the section if the line `section_text.append(section.strip_code().strip())` crashe...
1,593,429,043,000
1,595,521,714,000
null
NONE
null
null
Hi, I am trying to download some wikipedia data but I got this error for spanish "es" (but there are maybe some others languages which have the same error I haven't tried all of them ). `ERROR:root:mwparserfromhell ParseError: This is a bug and should be reported. Info: C tokenizer exited with non-empty token sta...
https://api.github.com/repos/huggingface/datasets/issues/321/timeline
null
false
https://api.github.com/repos/huggingface/datasets/issues/320
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/320/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/320/comments
https://api.github.com/repos/huggingface/datasets/issues/320/events
https://github.com/huggingface/datasets/issues/320
647,188,167
MDU6SXNzdWU2NDcxODgxNjc=
320
Blog Authorship Corpus, Non Matching Splits Sizes Error, nlp viewer
{ "login": "mariamabarham", "id": 38249783, "node_id": "MDQ6VXNlcjM4MjQ5Nzgz", "avatar_url": "https://avatars.githubusercontent.com/u/38249783?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariamabarham", "html_url": "https://github.com/mariamabarham", "followers_url": "https://api.githu...
[ { "id": 2107841032, "node_id": "MDU6TGFiZWwyMTA3ODQxMDMy", "url": "https://api.github.com/repos/huggingface/datasets/labels/nlp-viewer", "name": "nlp-viewer", "color": "94203D", "default": false, "description": "" } ]
closed
false
null
[]
null
[ "I wonder if this means downloading failed? That corpus has a really slow server.", "This dataset seems to have a decoding problem that results in inconsistencies in the number of generated examples.\r\nSee #215.\r\nThat's why we end up with a `NonMatchingSplitsSizesError `." ]
1,593,416,195,000
1,593,441,882,000
1,593,441,882,000
CONTRIBUTOR
null
null
Selecting `blog_authorship_corpus` in the nlp viewer throws the following error: ``` NonMatchingSplitsSizesError: [{'expected': SplitInfo(name='train', num_bytes=610252351, num_examples=532812, dataset_name='blog_authorship_corpus'), 'recorded': SplitInfo(name='train', num_bytes=614706451, num_examples=535568, dat...
https://api.github.com/repos/huggingface/datasets/issues/320/timeline
null
false