Dataset Viewer
Auto-converted to Parquet Duplicate
url
stringlengths
61
61
repository_url
stringclasses
1 value
labels_url
stringlengths
75
75
comments_url
stringlengths
70
70
events_url
stringlengths
68
68
html_url
stringlengths
49
51
id
int64
2.35B
2.43B
node_id
stringlengths
18
19
number
int64
6.97k
7.07k
title
stringlengths
6
95
user
dict
labels
listlengths
0
2
state
stringclasses
2 values
locked
bool
1 class
assignee
dict
assignees
listlengths
0
1
milestone
dict
comments
sequencelengths
0
21
created_at
timestamp[s]
updated_at
timestamp[s]
closed_at
timestamp[s]
author_association
stringclasses
3 values
active_lock_reason
null
draft
bool
2 classes
pull_request
dict
body
stringlengths
10
17.1k
reactions
dict
timeline_url
stringlengths
70
70
performed_via_github_app
null
state_reason
stringclasses
2 values
is_pull_request
bool
2 classes
https://api.github.com/repos/huggingface/datasets/issues/7068
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7068/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7068/comments
https://api.github.com/repos/huggingface/datasets/issues/7068/events
https://github.com/huggingface/datasets/pull/7068
2,426,657,434
PR_kwDODunzps52SwXS
7,068
Fix prepare_single_hop_path_and_storage_options
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[]
open
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7068). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2024-07-24T05:52:34
2024-07-24T08:54:32
null
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7068", "html_url": "https://github.com/huggingface/datasets/pull/7068", "diff_url": "https://github.com/huggingface/datasets/pull/7068.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7068.patch", "merged_at": null }
Fix `_prepare_single_hop_path_and_storage_options`: - Do not pass HF authentication headers and HF user-agent to non-HF HTTP URLs - Do not overwrite passed `storage_options` nested values: - Before, when passed ```DownloadConfig(storage_options={"https": {"client_kwargs": {"raise_for_status": True}}})```, ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7068/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7068/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/7067
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7067/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7067/comments
https://api.github.com/repos/huggingface/datasets/issues/7067/events
https://github.com/huggingface/datasets/issues/7067
2,425,460,168
I_kwDODunzps6QkZXI
7,067
Convert_to_parquet fails for datasets with multiple configs
{ "login": "HuangZhen02", "id": 97585031, "node_id": "U_kgDOBdEHhw", "avatar_url": "https://avatars.githubusercontent.com/u/97585031?v=4", "gravatar_id": "", "url": "https://api.github.com/users/HuangZhen02", "html_url": "https://github.com/HuangZhen02", "followers_url": "https://api.github.com/users/Hu...
[]
open
false
null
[]
null
[ "Many users have encountered the same issue, which has caused inconvenience.\r\n\r\nhttps://discuss.huggingface.co/t/convert-to-parquet-fails-for-datasets-with-multiple-configs/86733\r\n" ]
2024-07-23T15:09:33
2024-07-23T15:10:44
null
NONE
null
null
null
If the dataset has multiple configs, when using the `datasets-cli convert_to_parquet` command to avoid issues with the data viewer caused by loading scripts, the conversion process only successfully converts the data corresponding to the first config. When it starts converting the second config, it throws an error: ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7067/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7067/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7066
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7066/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7066/comments
https://api.github.com/repos/huggingface/datasets/issues/7066/events
https://github.com/huggingface/datasets/issues/7066
2,425,125,160
I_kwDODunzps6QjHko
7,066
One subset per file in repo ?
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
open
false
null
[]
null
[]
2024-07-23T12:43:59
2024-07-23T12:43:59
null
MEMBER
null
null
null
Right now we consider all the files of a dataset to be the same data, e.g. ``` single_subset_dataset/ ├── train0.jsonl ├── train1.jsonl └── train2.jsonl ``` but in cases like this, each file is actually a different subset of the dataset and should be loaded separately ``` many_subsets_dataset/ ├── animals.jso...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7066/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7066/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7065
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7065/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7065/comments
https://api.github.com/repos/huggingface/datasets/issues/7065/events
https://github.com/huggingface/datasets/issues/7065
2,424,734,953
I_kwDODunzps6QhoTp
7,065
Cannot get item after loading from disk and then converting to iterable.
{ "login": "happyTonakai", "id": 21305646, "node_id": "MDQ6VXNlcjIxMzA1NjQ2", "avatar_url": "https://avatars.githubusercontent.com/u/21305646?v=4", "gravatar_id": "", "url": "https://api.github.com/users/happyTonakai", "html_url": "https://github.com/happyTonakai", "followers_url": "https://api.github.c...
[]
open
false
null
[]
null
[]
2024-07-23T09:37:56
2024-07-23T09:37:56
null
NONE
null
null
null
### Describe the bug The dataset generated from local file works fine. ```py root = "/home/data/train" file_list1 = glob(os.path.join(root, "*part1.flac")) file_list2 = glob(os.path.join(root, "*part2.flac")) ds = ( Dataset.from_dict({"part1": file_list1, "part2": file_list2}) .cast_column("part1", Au...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7065/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7065/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7064
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7064/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7064/comments
https://api.github.com/repos/huggingface/datasets/issues/7064/events
https://github.com/huggingface/datasets/pull/7064
2,424,613,104
PR_kwDODunzps52Lz2-
7,064
Add `batch` method to `Dataset` class
{ "login": "lappemic", "id": 61876623, "node_id": "MDQ6VXNlcjYxODc2NjIz", "avatar_url": "https://avatars.githubusercontent.com/u/61876623?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lappemic", "html_url": "https://github.com/lappemic", "followers_url": "https://api.github.com/users/lap...
[]
open
false
null
[]
null
[ "Looks good to me ! :)\r\n\r\nyou might want to add the `map` num_proc argument as well, for people who want to make it run faster", "Thanks for the feedback @lhoestq! The last commits include:\r\n- Adding the `num_proc` parameter to `batch`\r\n- Adding tests similar to the one done for `IterableDataset.batch()`\...
2024-07-23T08:40:43
2024-07-24T06:17:45
null
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7064", "html_url": "https://github.com/huggingface/datasets/pull/7064", "diff_url": "https://github.com/huggingface/datasets/pull/7064.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7064.patch", "merged_at": null }
This PR introduces a new `batch` method to the `Dataset` class, aligning its functionality with the `IterableDataset.batch()` method (implemented in #7054). The implementation uses as well the existing `map` method for efficient batching of examples. Key changes: - Add `batch` method to `Dataset` class in `arrow_da...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7064/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7064/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/7063
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7063/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7063/comments
https://api.github.com/repos/huggingface/datasets/issues/7063/events
https://github.com/huggingface/datasets/issues/7063
2,424,488,648
I_kwDODunzps6QgsLI
7,063
Add `batch` method to `Dataset`
{ "login": "lappemic", "id": 61876623, "node_id": "MDQ6VXNlcjYxODc2NjIz", "avatar_url": "https://avatars.githubusercontent.com/u/61876623?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lappemic", "html_url": "https://github.com/lappemic", "followers_url": "https://api.github.com/users/lap...
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
open
false
null
[]
null
[]
2024-07-23T07:36:59
2024-07-23T07:36:59
null
CONTRIBUTOR
null
null
null
### Feature request Add a `batch` method to the Dataset class, similar to the one recently implemented for `IterableDataset` in PR #7054. ### Motivation A batched iteration speeds up data loading significantly (see e.g. #6279) ### Your contribution I plan to open a PR to implement this.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7063/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7063/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7062
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7062/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7062/comments
https://api.github.com/repos/huggingface/datasets/issues/7062/events
https://github.com/huggingface/datasets/pull/7062
2,424,467,484
PR_kwDODunzps52LUPR
7,062
Avoid calling http_head for non-HTTP URLs
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7062). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>...
2024-07-23T07:25:09
2024-07-23T14:28:27
2024-07-23T14:21:08
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7062", "html_url": "https://github.com/huggingface/datasets/pull/7062", "diff_url": "https://github.com/huggingface/datasets/pull/7062.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7062.patch", "merged_at": "2024-07-23T14:21...
Avoid calling `http_head` for non-HTTP URLs, by adding and `else` statement. Currently, it makes an unnecessary HTTP call (which adds latency) for non-HTTP protocols, like FTP, S3,... I discovered this while working in an unrelated issue.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7062/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7062/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/7061
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7061/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7061/comments
https://api.github.com/repos/huggingface/datasets/issues/7061/events
https://github.com/huggingface/datasets/issues/7061
2,423,786,881
I_kwDODunzps6QeA2B
7,061
Custom Dataset | Still Raise Error while handling errors in _generate_examples
{ "login": "hahmad2008", "id": 68266028, "node_id": "MDQ6VXNlcjY4MjY2MDI4", "avatar_url": "https://avatars.githubusercontent.com/u/68266028?v=4", "gravatar_id": "", "url": "https://api.github.com/users/hahmad2008", "html_url": "https://github.com/hahmad2008", "followers_url": "https://api.github.com/use...
[]
open
false
null
[]
null
[]
2024-07-22T21:18:12
2024-07-22T21:18:12
null
NONE
null
null
null
### Describe the bug I follow this [example](https://discuss.huggingface.co/t/error-handling-in-iterabledataset/72827/3) to handle errors in custom dataset. I am writing a dataset script which read jsonl files and i need to handle errors and continue reading files without raising exception and exit the execution. `...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7061/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7061/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7060
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7060/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7060/comments
https://api.github.com/repos/huggingface/datasets/issues/7060/events
https://github.com/huggingface/datasets/pull/7060
2,423,188,419
PR_kwDODunzps52G71g
7,060
WebDataset BuilderConfig
{ "login": "hlky", "id": 106811348, "node_id": "U_kgDOBl3P1A", "avatar_url": "https://avatars.githubusercontent.com/u/106811348?v=4", "gravatar_id": "", "url": "https://api.github.com/users/hlky", "html_url": "https://github.com/hlky", "followers_url": "https://api.github.com/users/hlky/followers", "f...
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7060). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update." ]
2024-07-22T15:41:07
2024-07-23T13:28:44
2024-07-23T13:28:44
NONE
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7060", "html_url": "https://github.com/huggingface/datasets/pull/7060", "diff_url": "https://github.com/huggingface/datasets/pull/7060.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7060.patch", "merged_at": null }
This PR adds `WebDatasetConfig`. Closes #7055
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7060/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7060/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/7059
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7059/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7059/comments
https://api.github.com/repos/huggingface/datasets/issues/7059/events
https://github.com/huggingface/datasets/issues/7059
2,422,827,892
I_kwDODunzps6QaWt0
7,059
None values are skipped when reading jsonl in subobjects
{ "login": "PonteIneptique", "id": 1929830, "node_id": "MDQ6VXNlcjE5Mjk4MzA=", "avatar_url": "https://avatars.githubusercontent.com/u/1929830?v=4", "gravatar_id": "", "url": "https://api.github.com/users/PonteIneptique", "html_url": "https://github.com/PonteIneptique", "followers_url": "https://api.gith...
[]
open
false
null
[]
null
[]
2024-07-22T13:02:42
2024-07-22T13:02:53
null
NONE
null
null
null
### Describe the bug I have been fighting against my machine since this morning only to find out this is some kind of a bug. When loading a dataset composed of `metadata.jsonl`, if you have nullable values (Optional[str]), they can be ignored by the parser, shifting things around. E.g., let's take this example ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7059/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7059/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7058
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7058/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7058/comments
https://api.github.com/repos/huggingface/datasets/issues/7058/events
https://github.com/huggingface/datasets/issues/7058
2,422,560,355
I_kwDODunzps6QZVZj
7,058
New feature type: Document
{ "login": "severo", "id": 1676121, "node_id": "MDQ6VXNlcjE2NzYxMjE=", "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "gravatar_id": "", "url": "https://api.github.com/users/severo", "html_url": "https://github.com/severo", "followers_url": "https://api.github.com/users/severo/foll...
[]
open
false
null
[]
null
[]
2024-07-22T10:49:20
2024-07-22T10:49:20
null
CONTRIBUTOR
null
null
null
It would be useful for PDF. https://github.com/huggingface/dataset-viewer/issues/2991#issuecomment-2242656069
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7058/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7058/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7057
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7057/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7057/comments
https://api.github.com/repos/huggingface/datasets/issues/7057/events
https://github.com/huggingface/datasets/pull/7057
2,422,498,520
PR_kwDODunzps52EjGC
7,057
Update load_hub.mdx
{ "login": "severo", "id": 1676121, "node_id": "MDQ6VXNlcjE2NzYxMjE=", "avatar_url": "https://avatars.githubusercontent.com/u/1676121?v=4", "gravatar_id": "", "url": "https://api.github.com/users/severo", "html_url": "https://github.com/severo", "followers_url": "https://api.github.com/users/severo/foll...
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7057). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>...
2024-07-22T10:17:46
2024-07-22T10:34:14
2024-07-22T10:28:10
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7057", "html_url": "https://github.com/huggingface/datasets/pull/7057", "diff_url": "https://github.com/huggingface/datasets/pull/7057.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7057.patch", "merged_at": "2024-07-22T10:28...
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7057/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7057/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/7056
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7056/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7056/comments
https://api.github.com/repos/huggingface/datasets/issues/7056/events
https://github.com/huggingface/datasets/pull/7056
2,422,192,257
PR_kwDODunzps52DgOu
7,056
Make `BufferShuffledExamplesIterable` resumable
{ "login": "yzhangcs", "id": 18402347, "node_id": "MDQ6VXNlcjE4NDAyMzQ3", "avatar_url": "https://avatars.githubusercontent.com/u/18402347?v=4", "gravatar_id": "", "url": "https://api.github.com/users/yzhangcs", "html_url": "https://github.com/yzhangcs", "followers_url": "https://api.github.com/users/yzh...
[]
open
false
null
[]
null
[ "Oh cool !\r\n\r\nThe time it takes to resume depends on the expected maximum distance in this case right ? Do you know its relationship with $B$ ?\r\n\r\nIn your test it already as high as 15k for $B=1024$, which is ok for text datasets but is maybe not ideal for datasets with heavy samples like audio/image/video ...
2024-07-22T07:50:02
2024-07-22T15:37:01
null
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7056", "html_url": "https://github.com/huggingface/datasets/pull/7056", "diff_url": "https://github.com/huggingface/datasets/pull/7056.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7056.patch", "merged_at": null }
This PR aims to implement a resumable `BufferShuffledExamplesIterable`. Instead of saving the entire buffer content, which is very memory-intensive, the newly implemented `BufferShuffledExamplesIterable` saves only the minimal state necessary for recovery, e.g., the random generator states and the state of the first e...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7056/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7056/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/7055
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7055/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7055/comments
https://api.github.com/repos/huggingface/datasets/issues/7055/events
https://github.com/huggingface/datasets/issues/7055
2,421,708,891
I_kwDODunzps6QWFhb
7,055
WebDataset with different prefixes are unsupported
{ "login": "hlky", "id": 106811348, "node_id": "U_kgDOBl3P1A", "avatar_url": "https://avatars.githubusercontent.com/u/106811348?v=4", "gravatar_id": "", "url": "https://api.github.com/users/hlky", "html_url": "https://github.com/hlky", "followers_url": "https://api.github.com/users/hlky/followers", "f...
[]
closed
false
null
[]
null
[ "Since `datasets` uses is built on Arrow to store the data, it requires each sample to have the same columns.\r\n\r\nThis can be fixed by specifyign in advance the name of all the possible columns in the `dataset_info` in YAML, and missing values will be `None`", "Thanks. This currently doesn't work for WebDatase...
2024-07-22T01:14:19
2024-07-23T13:58:01
2024-07-23T13:28:46
NONE
null
null
null
### Describe the bug Consider a WebDataset with multiple images for each item where the number of images may vary: [example](https://huggingface.co/datasets/bigdata-pw/fashion-150k) Due to this [code](https://github.com/huggingface/datasets/blob/87f4c2088854ff33e817e724e75179e9975c1b02/src/datasets/packaged_modules...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7055/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7055/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/7054
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7054/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7054/comments
https://api.github.com/repos/huggingface/datasets/issues/7054/events
https://github.com/huggingface/datasets/pull/7054
2,418,548,995
PR_kwDODunzps514T1f
7,054
Add batching to `IterableDataset`
{ "login": "lappemic", "id": 61876623, "node_id": "MDQ6VXNlcjYxODc2NjIz", "avatar_url": "https://avatars.githubusercontent.com/u/61876623?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lappemic", "html_url": "https://github.com/lappemic", "followers_url": "https://api.github.com/users/lap...
[]
closed
false
null
[]
null
[ "Cool ! Thanks for diving into it :)\r\n\r\nYour implementation is great and indeed supports shuffling and batching, you just need to additionally account for state_dict (for dataset [checkpointing+resuming](https://huggingface.co/docs/datasets/main/en/use_with_pytorch#checkpoint-and-resume))\r\n\r\nThat being said...
2024-07-19T10:11:47
2024-07-23T13:25:13
2024-07-23T10:34:28
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7054", "html_url": "https://github.com/huggingface/datasets/pull/7054", "diff_url": "https://github.com/huggingface/datasets/pull/7054.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7054.patch", "merged_at": "2024-07-23T10:34...
I've taken a try at implementing a batched `IterableDataset` as requested in issue #6279. This PR adds a new `BatchedExamplesIterable` class and a `.batch()` method to the `IterableDataset` class. The main changes are: 1. A new `BatchedExamplesIterable` that groups examples into batches. 2. A `.batch()` method for...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7054/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7054/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/7053
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7053/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7053/comments
https://api.github.com/repos/huggingface/datasets/issues/7053/events
https://github.com/huggingface/datasets/issues/7053
2,416,423,791
I_kwDODunzps6QB7Nv
7,053
Datasets.datafiles resolve_pattern `TypeError: can only concatenate tuple (not "str") to tuple`
{ "login": "MatthewYZhang", "id": 48289218, "node_id": "MDQ6VXNlcjQ4Mjg5MjE4", "avatar_url": "https://avatars.githubusercontent.com/u/48289218?v=4", "gravatar_id": "", "url": "https://api.github.com/users/MatthewYZhang", "html_url": "https://github.com/MatthewYZhang", "followers_url": "https://api.githu...
[]
closed
false
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[ { "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_...
null
[ "Hi,\r\n\r\nThis issue was fixed in `datasets` 2.15.0:\r\n- #6105\r\n\r\nYou will need to update your `datasets`:\r\n```\r\npip install -U datasets\r\n```", "Duplicate of:\r\n- #6100" ]
2024-07-18T13:42:35
2024-07-18T15:17:42
2024-07-18T15:16:18
NONE
null
null
null
### Describe the bug in data_files.py, line 332, `fs, _, _ = get_fs_token_paths(pattern, storage_options=storage_options)` If we run the code on AWS, as fs.protocol will be a tuple like: `('file', 'local')` So, `isinstance(fs.protocol, str) == False` and `protocol_prefix = fs.protocol + "://" if fs.protocol != ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7053/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7053/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/7052
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7052/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7052/comments
https://api.github.com/repos/huggingface/datasets/issues/7052/events
https://github.com/huggingface/datasets/pull/7052
2,411,682,730
PR_kwDODunzps51iuop
7,052
Adding `Music` feature for symbolic music modality (MIDI, abc)
{ "login": "Natooz", "id": 56734983, "node_id": "MDQ6VXNlcjU2NzM0OTgz", "avatar_url": "https://avatars.githubusercontent.com/u/56734983?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Natooz", "html_url": "https://github.com/Natooz", "followers_url": "https://api.github.com/users/Natooz/fo...
[]
open
false
null
[]
null
[]
2024-07-16T17:26:04
2024-07-16T17:26:04
null
NONE
null
true
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7052", "html_url": "https://github.com/huggingface/datasets/pull/7052", "diff_url": "https://github.com/huggingface/datasets/pull/7052.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7052.patch", "merged_at": null }
⚠️ (WIP) ⚠️ ### What this PR does This PR adds a `Music` feature for the symbolic music modality, in particular [MIDI](https://en.wikipedia.org/wiki/Musical_Instrument_Digital_Interface) and [abc](https://en.wikipedia.org/wiki/ABC_notation) files. ### Motivations These two file formats are widely used in th...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7052/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7052/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/7051
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7051/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7051/comments
https://api.github.com/repos/huggingface/datasets/issues/7051/events
https://github.com/huggingface/datasets/issues/7051
2,409,353,929
I_kwDODunzps6Pm9LJ
7,051
How to set_epoch with interleave_datasets?
{ "login": "jonathanasdf", "id": 511073, "node_id": "MDQ6VXNlcjUxMTA3Mw==", "avatar_url": "https://avatars.githubusercontent.com/u/511073?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jonathanasdf", "html_url": "https://github.com/jonathanasdf", "followers_url": "https://api.github.com/u...
[]
open
false
null
[]
null
[ "This is not possible right now afaik :/\r\n\r\nMaybe we could have something like this ? wdyt ?\r\n\r\n```python\r\nds = interleave_datasets(\r\n [shuffled_dataset_a, dataset_b],\r\n probabilities=probabilities,\r\n stopping_strategy='all_exhausted',\r\n reshuffle_each_iteration=True,\r\n)", "That wo...
2024-07-15T18:24:52
2024-07-22T16:52:07
null
NONE
null
null
null
Let's say I have dataset A which has 100k examples, and dataset B which has 100m examples. I want to train on an interleaved dataset of A+B, with stopping_strategy='all_exhausted' so dataset B doesn't repeat any examples. But every time A is exhausted I want it to be reshuffled (eg. calling set_epoch) Of course I...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7051/reactions", "total_count": 1, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 1 }
https://api.github.com/repos/huggingface/datasets/issues/7051/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7050
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7050/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7050/comments
https://api.github.com/repos/huggingface/datasets/issues/7050/events
https://github.com/huggingface/datasets/pull/7050
2,409,048,733
PR_kwDODunzps51Z1Yp
7,050
add checkpoint and resume title in docs
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7050). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>...
2024-07-15T15:38:04
2024-07-15T16:06:15
2024-07-15T15:59:56
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7050", "html_url": "https://github.com/huggingface/datasets/pull/7050", "diff_url": "https://github.com/huggingface/datasets/pull/7050.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7050.patch", "merged_at": "2024-07-15T15:59...
(minor) just to make it more prominent in the docs page for the soon-to-be-released new torchdata
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7050/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7050/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/7049
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7049/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7049/comments
https://api.github.com/repos/huggingface/datasets/issues/7049/events
https://github.com/huggingface/datasets/issues/7049
2,408,514,366
I_kwDODunzps6PjwM-
7,049
Save nparray as list
{ "login": "Sakurakdx", "id": 48399040, "node_id": "MDQ6VXNlcjQ4Mzk5MDQw", "avatar_url": "https://avatars.githubusercontent.com/u/48399040?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Sakurakdx", "html_url": "https://github.com/Sakurakdx", "followers_url": "https://api.github.com/users/...
[]
closed
false
null
[]
null
[ "In addition, when I use `set_format ` and index the ds, the following error occurs:\r\nthe code\r\n```python\r\nds.set_format(type=\"np\", colums=\"pixel_values\")\r\n```\r\nerror\r\n<img width=\"918\" alt=\"image\" src=\"https://github.com/user-attachments/assets/b28bbff2-20ea-4d28-ab62-b4ed2d944996\">\r\n", ">...
2024-07-15T11:36:11
2024-07-18T11:33:34
2024-07-18T11:33:34
NONE
null
null
null
### Describe the bug When I use the `map` function to convert images into features, datasets saves nparray as a list. Some people use the `set_format` function to convert the column back, but doesn't this lose precision? ### Steps to reproduce the bug the map function ```python def convert_image_to_features(inst, ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7049/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7049/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/7048
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7048/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7048/comments
https://api.github.com/repos/huggingface/datasets/issues/7048/events
https://github.com/huggingface/datasets/issues/7048
2,408,487,547
I_kwDODunzps6Pjpp7
7,048
ImportError: numpy.core.multiarray when using `filter`
{ "login": "kamilakesbi", "id": 45195979, "node_id": "MDQ6VXNlcjQ1MTk1OTc5", "avatar_url": "https://avatars.githubusercontent.com/u/45195979?v=4", "gravatar_id": "", "url": "https://api.github.com/users/kamilakesbi", "html_url": "https://github.com/kamilakesbi", "followers_url": "https://api.github.com/...
[]
closed
false
null
[]
null
[ "Could you please check your `numpy` version?", "I got this issue while using numpy version 2.0. \r\n\r\nI solved it by switching back to numpy 1.26.0 :) ", "We recently added support for numpy 2.0, but it is not released yet.", "Ok I see, thanks! I think we can close this issue for now as switching back to v...
2024-07-15T11:21:04
2024-07-16T10:11:25
2024-07-16T10:11:25
NONE
null
null
null
### Describe the bug I can't apply the filter method on my dataset. ### Steps to reproduce the bug The following snippet generates a bug: ```python from datasets import load_dataset ami = load_dataset('kamilakesbi/ami', 'ihm') ami['train'].filter( lambda example: example["file_name"] == 'EN2001a' ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7048/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7048/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/7047
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7047/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7047/comments
https://api.github.com/repos/huggingface/datasets/issues/7047/events
https://github.com/huggingface/datasets/issues/7047
2,406,495,084
I_kwDODunzps6PcDNs
7,047
Save Dataset as Sharded Parquet
{ "login": "tom-p-reichel", "id": 43631024, "node_id": "MDQ6VXNlcjQzNjMxMDI0", "avatar_url": "https://avatars.githubusercontent.com/u/43631024?v=4", "gravatar_id": "", "url": "https://api.github.com/users/tom-p-reichel", "html_url": "https://github.com/tom-p-reichel", "followers_url": "https://api.githu...
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
open
false
null
[]
null
[ "To anyone else who finds themselves in this predicament, it's possible to read the parquet file in the same way that datasets writes it, and then manually break it into pieces. Although, you need a couple of magic options (`thrift_*`) to deal with the huge metadata, otherwise pyarrow immediately crashes.\r\n```pyt...
2024-07-12T23:47:51
2024-07-17T12:07:08
null
NONE
null
null
null
### Feature request `to_parquet` currently saves the dataset as one massive, monolithic parquet file, rather than as several small parquet files. It should shard large datasets automatically. ### Motivation This default behavior makes me very sad because a program I ran for 6 hours saved its results using `to_...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7047/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7047/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7046
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7046/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7046/comments
https://api.github.com/repos/huggingface/datasets/issues/7046/events
https://github.com/huggingface/datasets/pull/7046
2,405,485,582
PR_kwDODunzps51N05n
7,046
Support librosa and numpy 2.0 for Python 3.10
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7046). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>...
2024-07-12T12:42:47
2024-07-12T13:04:40
2024-07-12T12:58:17
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7046", "html_url": "https://github.com/huggingface/datasets/pull/7046", "diff_url": "https://github.com/huggingface/datasets/pull/7046.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7046.patch", "merged_at": "2024-07-12T12:58...
Support librosa and numpy 2.0 for Python 3.10 by installing soxr 0.4.0b1 pre-release: - https://github.com/dofuuz/python-soxr/releases/tag/v0.4.0b1 - https://github.com/dofuuz/python-soxr/issues/28
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7046/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7046/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/7045
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7045/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7045/comments
https://api.github.com/repos/huggingface/datasets/issues/7045/events
https://github.com/huggingface/datasets/pull/7045
2,405,447,858
PR_kwDODunzps51Nsie
7,045
Fix tensorflow min version depending on Python version
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7045). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>...
2024-07-12T12:20:23
2024-07-12T12:38:53
2024-07-12T12:33:00
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7045", "html_url": "https://github.com/huggingface/datasets/pull/7045", "diff_url": "https://github.com/huggingface/datasets/pull/7045.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7045.patch", "merged_at": "2024-07-12T12:33...
Fix tensorflow min version depending on Python version. Related to: - #6991
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7045/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7045/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/7044
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7044/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7044/comments
https://api.github.com/repos/huggingface/datasets/issues/7044/events
https://github.com/huggingface/datasets/pull/7044
2,405,002,987
PR_kwDODunzps51MLbh
7,044
Mark tests that require librosa
{ "login": "albertvillanova", "id": 8515462, "node_id": "MDQ6VXNlcjg1MTU0NjI=", "avatar_url": "https://avatars.githubusercontent.com/u/8515462?v=4", "gravatar_id": "", "url": "https://api.github.com/users/albertvillanova", "html_url": "https://github.com/albertvillanova", "followers_url": "https://api.g...
[]
closed
false
null
[]
null
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_7044). All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.", "<details>\n<summary>Show benchmarks</summary>\n\nPyArrow==8.0.0\n\n<details>\n<summary>...
2024-07-12T08:06:59
2024-07-12T09:06:32
2024-07-12T09:00:09
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/7044", "html_url": "https://github.com/huggingface/datasets/pull/7044", "diff_url": "https://github.com/huggingface/datasets/pull/7044.diff", "patch_url": "https://github.com/huggingface/datasets/pull/7044.patch", "merged_at": "2024-07-12T09:00...
Mark tests that require `librosa`. Note that `librosa` is an optional dependency (installed with `audio` option) and we should be able to test environments without that library installed. This is the case if we want to test Numpy 2.0, which is currently incompatible with `librosa` due to its dependency on `soxr`: -...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/7044/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/7044/timeline
null
null
true
End of preview. Expand in Data Studio
README.md exists but content is empty.
Downloads last month
13