url
stringlengths
58
61
repository_url
stringclasses
1 value
labels_url
stringlengths
72
75
comments_url
stringlengths
67
70
events_url
stringlengths
65
68
html_url
stringlengths
46
51
id
int64
599M
2.8B
node_id
stringlengths
18
32
number
int64
1
7.38k
title
stringlengths
1
290
user
dict
labels
listlengths
0
4
state
stringclasses
2 values
locked
bool
1 class
assignee
dict
assignees
listlengths
0
4
milestone
dict
comments
sequencelengths
0
0
created_at
unknown
updated_at
unknown
closed_at
timestamp[us]
author_association
stringclasses
4 values
sub_issues_summary
dict
active_lock_reason
float64
body
stringlengths
0
228k
closed_by
dict
reactions
dict
timeline_url
stringlengths
67
70
performed_via_github_app
float64
state_reason
stringclasses
3 values
draft
float64
0
1
pull_request
dict
is_pull_request
bool
2 classes
https://api.github.com/repos/huggingface/datasets/issues/7378
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7378/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7378/comments
https://api.github.com/repos/huggingface/datasets/issues/7378/events
https://github.com/huggingface/datasets/issues/7378
2,802,957,388
I_kwDODunzps6nEbxM
7,378
Allow pushing config version to hub
{ "avatar_url": "https://avatars.githubusercontent.com/u/129072?v=4", "events_url": "https://api.github.com/users/momeara/events{/privacy}", "followers_url": "https://api.github.com/users/momeara/followers", "following_url": "https://api.github.com/users/momeara/following{/other_user}", "gists_url": "https://api.github.com/users/momeara/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/momeara", "id": 129072, "login": "momeara", "node_id": "MDQ6VXNlcjEyOTA3Mg==", "organizations_url": "https://api.github.com/users/momeara/orgs", "received_events_url": "https://api.github.com/users/momeara/received_events", "repos_url": "https://api.github.com/users/momeara/repos", "site_admin": false, "starred_url": "https://api.github.com/users/momeara/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/momeara/subscriptions", "type": "User", "url": "https://api.github.com/users/momeara", "user_view_type": "public" }
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
open
false
null
[]
null
[]
"2025-01-21T22:35:07"
"2025-01-21T22:35:07"
1970-01-01T00:00:00
NONE
{ "completed": 0, "percent_completed": 0, "total": 0 }
null
### Feature request Currently, when datasets are created, they can be versioned by passing the `version` argument to `load_dataset(...)`. For example creating `outcomes.csv` on the command line ``` echo "id,value\n1,0\n2,0\n3,1\n4,1\n" > outcomes.csv ``` and creating it ``` import datasets dataset = datasets.load_dataset( "csv", data_files ="outcomes.csv", keep_in_memory = True, version = '1.0.0') ``` The version info is stored in the `info` and can be accessed e.g. by `next(iter(dataset.values())).info.version` This dataset can be uploaded to the hub with `dataset.push_to_hub(repo_id = "maomlab/example_dataset")`. This will create a dataset on the hub with the following in the `README.md`, but it doesn't upload the version information: ``` --- dataset_info: features: - name: id dtype: int64 - name: value dtype: int64 splits: - name: train num_bytes: 64 num_examples: 4 download_size: 1332 dataset_size: 64 configs: - config_name: default data_files: - split: train path: data/train-* --- ``` However, when I download from the hub, the version information is missing: ``` dataset_from_hub_no_version = datasets.load_dataset("maomlab/example_dataset") next(iter(dataset.values())).info.version ``` I can add the version information manually to the hub, by appending it to the end of config section: ``` ... configs: - config_name: default data_files: - split: train path: data/train-* version: 1.0.0 --- ``` And then when I download it, the version information is correct. ### Motivation ### Why adding version information for each config makes sense 1. The version information is already recorded in the dataset config info data structure and is able to parse it correctly, so it makes sense to sync it with `push_to_hub`. 2. Keeping the version info in at the config level is different from version info at the branch level. As the former relates to the version of the specific dataset the config refers to rather than the version of the dataset curation itself. ## A explanation for the current behavior: In [datasets/src/datasets/info.py:159](https://github.com/huggingface/datasets/blob/fb91fd3c9ea91a818681a777faf8d0c46f14c680/src/datasets/info.py#L159C1-L160C1 ), the `_INCLUDED_INFO_IN_YAML` variable doesn't include `"version"`. If my reading of the code is right, adding `"version"` to `_INCLUDED_INFO_IN_YAML`, would allow the version information to be uploaded to the hub. ### Your contribution Request: add `"version"` to `_INCLUDE_INFO_IN_YAML` in [datasets/src/datasets/info.py:159](https://github.com/huggingface/datasets/blob/fb91fd3c9ea91a818681a777faf8d0c46f14c680/src/datasets/info.py#L159C1-L160C1 )
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7378/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7378/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7377
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7377/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7377/comments
https://api.github.com/repos/huggingface/datasets/issues/7377/events
https://github.com/huggingface/datasets/issues/7377
2,802,723,285
I_kwDODunzps6nDinV
7,377
Support for sparse arrays with the Arrow Sparse Tensor format?
{ "avatar_url": "https://avatars.githubusercontent.com/u/3231217?v=4", "events_url": "https://api.github.com/users/JulesGM/events{/privacy}", "followers_url": "https://api.github.com/users/JulesGM/followers", "following_url": "https://api.github.com/users/JulesGM/following{/other_user}", "gists_url": "https://api.github.com/users/JulesGM/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/JulesGM", "id": 3231217, "login": "JulesGM", "node_id": "MDQ6VXNlcjMyMzEyMTc=", "organizations_url": "https://api.github.com/users/JulesGM/orgs", "received_events_url": "https://api.github.com/users/JulesGM/received_events", "repos_url": "https://api.github.com/users/JulesGM/repos", "site_admin": false, "starred_url": "https://api.github.com/users/JulesGM/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/JulesGM/subscriptions", "type": "User", "url": "https://api.github.com/users/JulesGM", "user_view_type": "public" }
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
open
false
null
[]
null
[]
"2025-01-21T20:14:35"
"2025-01-21T20:17:17"
1970-01-01T00:00:00
NONE
{ "completed": 0, "percent_completed": 0, "total": 0 }
null
### Feature request AI in biology is becoming a big thing. One thing that would be a huge benefit to the field that Huggingface Datasets doesn't currently have is native support for **sparse arrays**. Arrow has support for sparse tensors. https://arrow.apache.org/docs/format/Other.html#sparse-tensor It would be a big deal if Hugging Face Datasets supported sparse tensors as a feature type, natively. ### Motivation This is important for example in the field of transcriptomics (modeling and understanding gene expression), because a large fraction of the genes are not expressed (zero). More generally, in science, sparse arrays are very common, so adding support for them would be very benefitial, it would make just using Hugging Face Dataset objects a lot more straightforward and clean. ### Your contribution We can discuss this further once the team comments of what they think about the feature, and if there were previous attempts at making it work, and understanding their evaluation of how hard it would be. My intuition is that it should be fairly straightforward, as the Arrow backend already supports it.
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7377/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7377/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7376
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7376/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7376/comments
https://api.github.com/repos/huggingface/datasets/issues/7376/events
https://github.com/huggingface/datasets/pull/7376
2,802,621,104
PR_kwDODunzps6IiO9j
7,376
[docs] uv install
{ "avatar_url": "https://avatars.githubusercontent.com/u/59462357?v=4", "events_url": "https://api.github.com/users/stevhliu/events{/privacy}", "followers_url": "https://api.github.com/users/stevhliu/followers", "following_url": "https://api.github.com/users/stevhliu/following{/other_user}", "gists_url": "https://api.github.com/users/stevhliu/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/stevhliu", "id": 59462357, "login": "stevhliu", "node_id": "MDQ6VXNlcjU5NDYyMzU3", "organizations_url": "https://api.github.com/users/stevhliu/orgs", "received_events_url": "https://api.github.com/users/stevhliu/received_events", "repos_url": "https://api.github.com/users/stevhliu/repos", "site_admin": false, "starred_url": "https://api.github.com/users/stevhliu/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/stevhliu/subscriptions", "type": "User", "url": "https://api.github.com/users/stevhliu", "user_view_type": "public" }
[]
open
false
null
[]
null
[]
"2025-01-21T19:15:48"
"2025-01-21T19:39:29"
1970-01-01T00:00:00
MEMBER
{ "completed": 0, "percent_completed": 0, "total": 0 }
null
Proposes adding uv to installation docs (see Slack thread [here](https://huggingface.slack.com/archives/C01N44FJDHT/p1737377177709279) for more context) if you're interested!
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7376/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7376/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7376.diff", "html_url": "https://github.com/huggingface/datasets/pull/7376", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7376.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7376" }
true
https://api.github.com/repos/huggingface/datasets/issues/7375
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7375/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7375/comments
https://api.github.com/repos/huggingface/datasets/issues/7375/events
https://github.com/huggingface/datasets/issues/7375
2,800,609,218
I_kwDODunzps6m7efC
7,375
vllm批量推理报错
{ "avatar_url": "https://avatars.githubusercontent.com/u/51228154?v=4", "events_url": "https://api.github.com/users/YuShengzuishuai/events{/privacy}", "followers_url": "https://api.github.com/users/YuShengzuishuai/followers", "following_url": "https://api.github.com/users/YuShengzuishuai/following{/other_user}", "gists_url": "https://api.github.com/users/YuShengzuishuai/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/YuShengzuishuai", "id": 51228154, "login": "YuShengzuishuai", "node_id": "MDQ6VXNlcjUxMjI4MTU0", "organizations_url": "https://api.github.com/users/YuShengzuishuai/orgs", "received_events_url": "https://api.github.com/users/YuShengzuishuai/received_events", "repos_url": "https://api.github.com/users/YuShengzuishuai/repos", "site_admin": false, "starred_url": "https://api.github.com/users/YuShengzuishuai/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/YuShengzuishuai/subscriptions", "type": "User", "url": "https://api.github.com/users/YuShengzuishuai", "user_view_type": "public" }
[]
open
false
null
[]
null
[]
"2025-01-21T03:22:23"
"2025-01-21T03:22:23"
1970-01-01T00:00:00
NONE
{ "completed": 0, "percent_completed": 0, "total": 0 }
null
### Describe the bug ![Image](https://github.com/user-attachments/assets/3d958e43-28dc-4467-9333-5990c7af3b3f) ### Steps to reproduce the bug ![Image](https://github.com/user-attachments/assets/3067eeca-a54d-4956-b0fd-3fc5ea93dabb) ### Expected behavior ![Image](https://github.com/user-attachments/assets/77d32936-488f-4572-9365-bfb4170e555b) ### Environment info ![Image](https://github.com/user-attachments/assets/230335c4-825f-4db1-b07d-4776ef63ead8)
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7375/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7375/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7374
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7374/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7374/comments
https://api.github.com/repos/huggingface/datasets/issues/7374/events
https://github.com/huggingface/datasets/pull/7374
2,793,442,320
PR_kwDODunzps6IC66n
7,374
Remove .h5 from imagefolder extensions
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
closed
false
null
[]
null
[]
"2025-01-16T18:17:24"
"2025-01-16T18:26:40"
2025-01-16T18:26:38
MEMBER
{ "completed": 0, "percent_completed": 0, "total": 0 }
null
the format is not relevant for imagefolder, and makes the viewer fail to process datasets on HF (so many that the viewer takes more time to process new datasets)
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7374/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7374/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7374.diff", "html_url": "https://github.com/huggingface/datasets/pull/7374", "merged_at": "2025-01-16T18:26:38Z", "patch_url": "https://github.com/huggingface/datasets/pull/7374.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7374" }
true
https://api.github.com/repos/huggingface/datasets/issues/7373
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7373/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7373/comments
https://api.github.com/repos/huggingface/datasets/issues/7373/events
https://github.com/huggingface/datasets/issues/7373
2,793,237,139
I_kwDODunzps6mfWqT
7,373
Excessive RAM Usage After Dataset Concatenation concatenate_datasets
{ "avatar_url": "https://avatars.githubusercontent.com/u/40773225?v=4", "events_url": "https://api.github.com/users/sam-hey/events{/privacy}", "followers_url": "https://api.github.com/users/sam-hey/followers", "following_url": "https://api.github.com/users/sam-hey/following{/other_user}", "gists_url": "https://api.github.com/users/sam-hey/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/sam-hey", "id": 40773225, "login": "sam-hey", "node_id": "MDQ6VXNlcjQwNzczMjI1", "organizations_url": "https://api.github.com/users/sam-hey/orgs", "received_events_url": "https://api.github.com/users/sam-hey/received_events", "repos_url": "https://api.github.com/users/sam-hey/repos", "site_admin": false, "starred_url": "https://api.github.com/users/sam-hey/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sam-hey/subscriptions", "type": "User", "url": "https://api.github.com/users/sam-hey", "user_view_type": "public" }
[]
open
false
null
[]
null
[]
"2025-01-16T16:33:10"
"2025-01-17T08:05:22"
1970-01-01T00:00:00
NONE
{ "completed": 0, "percent_completed": 0, "total": 0 }
null
### Describe the bug When loading a dataset from disk, concatenating it, and starting the training process, the RAM usage progressively increases until the kernel terminates the process due to excessive memory consumption. https://github.com/huggingface/datasets/issues/2276 ### Steps to reproduce the bug ``` rom datasets import DatasetDict, concatenate_datasets dataset = DatasetDict.load_from_disk("data") ... ... combined_dataset = concatenate_datasets( [dataset[split] for split in dataset] ) #start SentenceTransformer training ``` ### Expected behavior I would not expect RAM utilization to increase after concatenation. Removing the concatenation step resolves the issue ### Environment info sentence-transformers==3.1.1 datasets==3.2.0 python3.10
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7373/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7373/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7372
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7372/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7372/comments
https://api.github.com/repos/huggingface/datasets/issues/7372/events
https://github.com/huggingface/datasets/issues/7372
2,791,760,968
I_kwDODunzps6mZuRI
7,372
Inconsistent Behavior Between `load_dataset` and `load_from_disk` When Loading Sharded Datasets
{ "avatar_url": "https://avatars.githubusercontent.com/u/38203359?v=4", "events_url": "https://api.github.com/users/gaohongkui/events{/privacy}", "followers_url": "https://api.github.com/users/gaohongkui/followers", "following_url": "https://api.github.com/users/gaohongkui/following{/other_user}", "gists_url": "https://api.github.com/users/gaohongkui/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/gaohongkui", "id": 38203359, "login": "gaohongkui", "node_id": "MDQ6VXNlcjM4MjAzMzU5", "organizations_url": "https://api.github.com/users/gaohongkui/orgs", "received_events_url": "https://api.github.com/users/gaohongkui/received_events", "repos_url": "https://api.github.com/users/gaohongkui/repos", "site_admin": false, "starred_url": "https://api.github.com/users/gaohongkui/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/gaohongkui/subscriptions", "type": "User", "url": "https://api.github.com/users/gaohongkui", "user_view_type": "public" }
[]
open
false
null
[]
null
[]
"2025-01-16T05:47:20"
"2025-01-16T05:47:20"
1970-01-01T00:00:00
NONE
{ "completed": 0, "percent_completed": 0, "total": 0 }
null
### Description I encountered an inconsistency in behavior between `load_dataset` and `load_from_disk` when loading sharded datasets. Here is a minimal example to reproduce the issue: #### Code 1: Using `load_dataset` ```python from datasets import Dataset, load_dataset # First save with max_shard_size=10 Dataset.from_dict({"id": range(1000)}).train_test_split(test_size=0.1).save_to_disk("my_sharded_datasetdict", max_shard_size=10) # Second save with max_shard_size=10 Dataset.from_dict({"id": range(500)}).train_test_split(test_size=0.1).save_to_disk("my_sharded_datasetdict", max_shard_size=10) # Load the DatasetDict loaded_datasetdict = load_dataset("my_sharded_datasetdict") print(loaded_datasetdict) ``` **Output**: - `train` has 1350 samples. - `test` has 150 samples. #### Code 2: Using `load_from_disk` ```python from datasets import Dataset, load_from_disk # First save with max_shard_size=10 Dataset.from_dict({"id": range(1000)}).train_test_split(test_size=0.1).save_to_disk("my_sharded_datasetdict", max_shard_size=10) # Second save with max_shard_size=10 Dataset.from_dict({"id": range(500)}).train_test_split(test_size=0.1).save_to_disk("my_sharded_datasetdict", max_shard_size=10) # Load the DatasetDict loaded_datasetdict = load_from_disk("my_sharded_datasetdict") print(loaded_datasetdict) ``` **Output**: - `train` has 450 samples. - `test` has 50 samples. ### Expected Behavior I expected both `load_dataset` and `load_from_disk` to load the same dataset, as they are pointing to the same directory. However, the results differ significantly: - `load_dataset` seems to merge all shards, resulting in a combined dataset. - `load_from_disk` only loads the last saved dataset, ignoring previous shards. ### Questions 1. Is this behavior intentional? If so, could you clarify the difference between `load_dataset` and `load_from_disk` in the documentation? 2. If this is not intentional, could this be considered a bug? 3. What is the recommended way to handle cases where multiple datasets are saved to the same directory? Thank you for your time and effort in maintaining this great library! I look forward to your feedback.
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7372/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7372/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7371
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7371/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7371/comments
https://api.github.com/repos/huggingface/datasets/issues/7371/events
https://github.com/huggingface/datasets/issues/7371
2,790,549,889
I_kwDODunzps6mVGmB
7,371
500 Server error with pushing a dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/7677814?v=4", "events_url": "https://api.github.com/users/martinmatak/events{/privacy}", "followers_url": "https://api.github.com/users/martinmatak/followers", "following_url": "https://api.github.com/users/martinmatak/following{/other_user}", "gists_url": "https://api.github.com/users/martinmatak/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/martinmatak", "id": 7677814, "login": "martinmatak", "node_id": "MDQ6VXNlcjc2Nzc4MTQ=", "organizations_url": "https://api.github.com/users/martinmatak/orgs", "received_events_url": "https://api.github.com/users/martinmatak/received_events", "repos_url": "https://api.github.com/users/martinmatak/repos", "site_admin": false, "starred_url": "https://api.github.com/users/martinmatak/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/martinmatak/subscriptions", "type": "User", "url": "https://api.github.com/users/martinmatak", "user_view_type": "public" }
[]
open
false
null
[]
null
[]
"2025-01-15T18:23:02"
"2025-01-15T20:06:05"
1970-01-01T00:00:00
NONE
{ "completed": 0, "percent_completed": 0, "total": 0 }
null
### Describe the bug Suddenly, I started getting this error message saying it was an internal error. `Error creating/pushing dataset: 500 Server Error: Internal Server Error for url: https://huggingface.co./api/datasets/ll4ma-lab/grasp-dataset/commit/main (Request ID: Root=1-6787f0b7-66d5bd45413e481c4c2fb22d;670d04ff-65f5-4741-a353-2eacc47a3928) Internal Error - We're working hard to fix this as soon as possible! Traceback (most recent call last): File "/uufs/chpc.utah.edu/common/home/hermans-group1/martin/software/pkg/miniforge3/envs/myenv2/lib/python3.10/site-packages/huggingface_hub/utils/_http.py", line 406, in hf_raise_for_status response.raise_for_status() File "/uufs/chpc.utah.edu/common/home/hermans-group1/martin/software/pkg/miniforge3/envs/myenv2/lib/python3.10/site-packages/requests/models.py", line 1024, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 500 Server Error: Internal Server Error for url: https://huggingface.co./api/datasets/ll4ma-lab/grasp-dataset/commit/main The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/uufs/chpc.utah.edu/common/home/u1295595/grasp_dataset_converter/src/grasp_dataset_converter/main.py", line 142, in main subset_train.push_to_hub(dataset_name, split='train') File "/uufs/chpc.utah.edu/common/home/hermans-group1/martin/software/pkg/miniforge3/envs/myenv2/lib/python3.10/site-packages/datasets/arrow_dataset.py", line 5624, in push_to_hub commit_info = api.create_commit( File "/uufs/chpc.utah.edu/common/home/hermans-group1/martin/software/pkg/miniforge3/envs/myenv2/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 114, in _inner_fn return fn(*args, **kwargs) File "/uufs/chpc.utah.edu/common/home/hermans-group1/martin/software/pkg/miniforge3/envs/myenv2/lib/python3.10/site-packages/huggingface_hub/hf_api.py", line 1518, in _inner return fn(self, *args, **kwargs) File "/uufs/chpc.utah.edu/common/home/hermans-group1/martin/software/pkg/miniforge3/envs/myenv2/lib/python3.10/site-packages/huggingface_hub/hf_api.py", line 4087, in create_commit hf_raise_for_status(commit_resp, endpoint_name="commit") File "/uufs/chpc.utah.edu/common/home/hermans-group1/martin/software/pkg/miniforge3/envs/myenv2/lib/python3.10/site-packages/huggingface_hub/utils/_http.py", line 477, in hf_raise_for_status raise _format(HfHubHTTPError, str(e), response) from e huggingface_hub.errors.HfHubHTTPError: 500 Server Error: Internal Server Error for url: https://huggingface.co./api/datasets/ll4ma-lab/grasp-dataset/commit/main (Request ID: Root=1-6787f0b7-66d5bd45413e481c4c2fb22d;670d04ff-65f5-4741-a353-2eacc47a3928) Internal Error - We're working hard to fix this as soon as possible!` ### Steps to reproduce the bug I am pushing a Dataset in a loop via push_to_hub API ### Expected behavior It worked fine until it stopped working suddenly. Expected behavior: It should start working again ### Environment info - `datasets` version: 3.2.0 - Platform: Linux-4.18.0-477.15.1.el8_8.x86_64-x86_64-with-glibc2.28 - Python version: 3.10.0 - `huggingface_hub` version: 0.27.1 - PyArrow version: 18.1.0 - Pandas version: 2.2.3 - `fsspec` version: 2024.9.0
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7371/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7371/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7370
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7370/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7370/comments
https://api.github.com/repos/huggingface/datasets/issues/7370/events
https://github.com/huggingface/datasets/pull/7370
2,787,972,786
PR_kwDODunzps6HwAu7
7,370
Support faster processing using pandas or polars functions in `IterableDataset.map()`
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
open
false
null
[]
null
[]
"2025-01-14T18:14:13"
"2025-01-14T18:30:13"
1970-01-01T00:00:00
MEMBER
{ "completed": 0, "percent_completed": 0, "total": 0 }
null
Allow super fast processing using pandas or polars functions in `IterableDataset.map()` by adding support to pandas and polars formatting in `IterableDataset` ```python import polars as pl from datasets import Dataset ds = Dataset.from_dict({"i": range(10)}).to_iterable_dataset() ds = ds.with_format("polars") ds = ds.map(lambda df: df.with_columns(pl.col("i"), pl.col("i").add(1).alias("i+1")), batched=True) ds = ds.with_format(None) print(next(iter(ds))) # {'i': 0, 'i+1': 1} ``` It leverages arrow's zero-copy features from/to pandas and polars. related to https://github.com/huggingface/datasets/issues/3444 #6762
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7370/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7370/timeline
null
null
1
{ "diff_url": "https://github.com/huggingface/datasets/pull/7370.diff", "html_url": "https://github.com/huggingface/datasets/pull/7370", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7370.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7370" }
true
https://api.github.com/repos/huggingface/datasets/issues/7369
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7369/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7369/comments
https://api.github.com/repos/huggingface/datasets/issues/7369/events
https://github.com/huggingface/datasets/issues/7369
2,787,193,238
I_kwDODunzps6mITGW
7,369
Importing dataset gives unhelpful error message when filenames in metadata.csv are not found in the directory
{ "avatar_url": "https://avatars.githubusercontent.com/u/38278139?v=4", "events_url": "https://api.github.com/users/svencornetsdegroot/events{/privacy}", "followers_url": "https://api.github.com/users/svencornetsdegroot/followers", "following_url": "https://api.github.com/users/svencornetsdegroot/following{/other_user}", "gists_url": "https://api.github.com/users/svencornetsdegroot/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/svencornetsdegroot", "id": 38278139, "login": "svencornetsdegroot", "node_id": "MDQ6VXNlcjM4Mjc4MTM5", "organizations_url": "https://api.github.com/users/svencornetsdegroot/orgs", "received_events_url": "https://api.github.com/users/svencornetsdegroot/received_events", "repos_url": "https://api.github.com/users/svencornetsdegroot/repos", "site_admin": false, "starred_url": "https://api.github.com/users/svencornetsdegroot/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/svencornetsdegroot/subscriptions", "type": "User", "url": "https://api.github.com/users/svencornetsdegroot", "user_view_type": "public" }
[]
open
false
null
[]
null
[]
"2025-01-14T13:53:21"
"2025-01-14T15:05:51"
1970-01-01T00:00:00
NONE
{ "completed": 0, "percent_completed": 0, "total": 0 }
null
### Describe the bug While importing an audiofolder dataset, where the names of the audiofiles don't correspond to the filenames in the metadata.csv, we get an unclear error message that is not helpful for the debugging, i.e. ``` ValueError: Instruction "train" corresponds to no data! ``` ### Steps to reproduce the bug Assume an audiofolder with audiofiles, filename1.mp3, filename2.mp3 etc and a file metadata.csv which contains the columns file_name and sentence. The file_names are formatted like filename1.mp3, filename2.mp3 etc. Load the audio ``` from datasets import load_dataset load_dataset("audiofolder", data_dir='/path/to/audiofolder') ``` When the file_names in the csv are not in sync with the filenames in the audiofolder, then we get an Error message: ``` File /opt/conda/lib/python3.12/site-packages/datasets/arrow_reader.py:251, in BaseReader.read(self, name, instructions, split_infos, in_memory) 249 if not files: 250 msg = f'Instruction "{instructions}" corresponds to no data!' --> 251 raise ValueError(msg) 252 return self.read_files(files=files, original_instructions=instructions, in_memory=in_memory) ValueError: Instruction "train" corresponds to no data! ``` load_dataset has a default value for the argument split = 'train'. ### Expected behavior It would be better to get an error report something like: ``` The metadata.csv file has different filenames than the files in the datadirectory. ``` It would have saved me 4 hours of debugging. ### Environment info - `datasets` version: 3.2.0 - Platform: Linux-5.14.0-427.40.1.el9_4.x86_64-x86_64-with-glibc2.39 - Python version: 3.12.8 - `huggingface_hub` version: 0.27.0 - PyArrow version: 18.1.0 - Pandas version: 2.2.3 - `fsspec` version: 2024.9.0
null
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/7369/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7369/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7368
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7368/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7368/comments
https://api.github.com/repos/huggingface/datasets/issues/7368/events
https://github.com/huggingface/datasets/pull/7368
2,784,272,477
PR_kwDODunzps6HjE97
7,368
Add with_split to DatasetDict.map
{ "avatar_url": "https://avatars.githubusercontent.com/u/93233241?v=4", "events_url": "https://api.github.com/users/jp1924/events{/privacy}", "followers_url": "https://api.github.com/users/jp1924/followers", "following_url": "https://api.github.com/users/jp1924/following{/other_user}", "gists_url": "https://api.github.com/users/jp1924/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/jp1924", "id": 93233241, "login": "jp1924", "node_id": "U_kgDOBY6gWQ", "organizations_url": "https://api.github.com/users/jp1924/orgs", "received_events_url": "https://api.github.com/users/jp1924/received_events", "repos_url": "https://api.github.com/users/jp1924/repos", "site_admin": false, "starred_url": "https://api.github.com/users/jp1924/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jp1924/subscriptions", "type": "User", "url": "https://api.github.com/users/jp1924", "user_view_type": "public" }
[]
open
false
null
[]
null
[]
"2025-01-13T15:09:56"
"2025-01-22T02:32:11"
1970-01-01T00:00:00
NONE
{ "completed": 0, "percent_completed": 0, "total": 0 }
null
#7356
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7368/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7368/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7368.diff", "html_url": "https://github.com/huggingface/datasets/pull/7368", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7368.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7368" }
true
https://api.github.com/repos/huggingface/datasets/issues/7366
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7366/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7366/comments
https://api.github.com/repos/huggingface/datasets/issues/7366/events
https://github.com/huggingface/datasets/issues/7366
2,781,522,894
I_kwDODunzps6lyqvO
7,366
Dataset.from_dict() can't handle large dict
{ "avatar_url": "https://avatars.githubusercontent.com/u/164967134?v=4", "events_url": "https://api.github.com/users/CSU-OSS/events{/privacy}", "followers_url": "https://api.github.com/users/CSU-OSS/followers", "following_url": "https://api.github.com/users/CSU-OSS/following{/other_user}", "gists_url": "https://api.github.com/users/CSU-OSS/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/CSU-OSS", "id": 164967134, "login": "CSU-OSS", "node_id": "U_kgDOCdUy3g", "organizations_url": "https://api.github.com/users/CSU-OSS/orgs", "received_events_url": "https://api.github.com/users/CSU-OSS/received_events", "repos_url": "https://api.github.com/users/CSU-OSS/repos", "site_admin": false, "starred_url": "https://api.github.com/users/CSU-OSS/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/CSU-OSS/subscriptions", "type": "User", "url": "https://api.github.com/users/CSU-OSS", "user_view_type": "public" }
[]
open
false
null
[]
null
[]
"2025-01-11T02:05:21"
"2025-01-11T02:05:21"
1970-01-01T00:00:00
NONE
{ "completed": 0, "percent_completed": 0, "total": 0 }
null
### Describe the bug I have 26,000,000 3-tuples. When I use Dataset.from_dict() to load, neither. py nor Jupiter notebook can run successfully. This is my code: ``` # len(example_data) is 26,000,000, 'diff' is a text diff1_list = [example_data[i].texts[0] for i in range(len(example_data))] diff2_list = [example_data[i].texts[1] for i in range(len(example_data))] label_list = [example_data[i].label for i in range(len(example_data))] embedding_dataset = Dataset.from_dict({ "diff1": diff1_list, "diff2": diff2_list, "label": label_list }) ``` ### Steps to reproduce the bug 1. Initialize a large 3-tuple, e.g. 26,000,000 2. Use Dataset.from_dict() to load ### Expected behavior Dataset.from_dict() run successfully ### Environment info sentence-transformers 3.3.1
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7366/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7366/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7365
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7365/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7365/comments
https://api.github.com/repos/huggingface/datasets/issues/7365/events
https://github.com/huggingface/datasets/issues/7365
2,780,216,199
I_kwDODunzps6ltruH
7,365
A parameter is specified but not used in datasets.arrow_dataset.Dataset.from_pandas()
{ "avatar_url": "https://avatars.githubusercontent.com/u/69003192?v=4", "events_url": "https://api.github.com/users/NourOM02/events{/privacy}", "followers_url": "https://api.github.com/users/NourOM02/followers", "following_url": "https://api.github.com/users/NourOM02/following{/other_user}", "gists_url": "https://api.github.com/users/NourOM02/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/NourOM02", "id": 69003192, "login": "NourOM02", "node_id": "MDQ6VXNlcjY5MDAzMTky", "organizations_url": "https://api.github.com/users/NourOM02/orgs", "received_events_url": "https://api.github.com/users/NourOM02/received_events", "repos_url": "https://api.github.com/users/NourOM02/repos", "site_admin": false, "starred_url": "https://api.github.com/users/NourOM02/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/NourOM02/subscriptions", "type": "User", "url": "https://api.github.com/users/NourOM02", "user_view_type": "public" }
[]
open
false
null
[]
null
[]
"2025-01-10T13:39:33"
"2025-01-10T13:39:33"
1970-01-01T00:00:00
NONE
{ "completed": 0, "percent_completed": 0, "total": 0 }
null
### Describe the bug I am interested in creating train, test and eval splits from a pandas Dataframe, therefore I was looking at the possibilities I can follow. I noticed the split parameter and was hopeful to use it in order to generate the 3 at once, however, while trying to understand the code, i noticed that it has no added value (correct me if I am wrong or misunderstood the code). from_pandas function code : ```python if info is not None and features is not None and info.features != features: raise ValueError( f"Features specified in `features` and `info.features` can't be different:\n{features}\n{info.features}" ) features = features if features is not None else info.features if info is not None else None if info is None: info = DatasetInfo() info.features = features table = InMemoryTable.from_pandas( df=df, preserve_index=preserve_index, ) if features is not None: # more expensive cast than InMemoryTable.from_pandas(..., schema=features.arrow_schema) # needed to support the str to Audio conversion for instance table = table.cast(features.arrow_schema) return cls(table, info=info, split=split) ``` ### Steps to reproduce the bug ```python from datasets import Dataset # Filling the split parameter with whatever causes no harm at all data = Dataset.from_pandas(self.raw_data, split='egiojegoierjgoiejgrefiergiuorenvuirgurthgi') ``` ### Expected behavior Would be great if there is no split parameter (if it isn't working), or to add a concrete example of how it can be used. ### Environment info - `datasets` version: 3.2.0 - Platform: Linux-5.15.0-127-generic-x86_64-with-glibc2.35 - Python version: 3.10.12 - `huggingface_hub` version: 0.27.1 - PyArrow version: 18.1.0 - Pandas version: 2.2.3 - `fsspec` version: 2024.9.0
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7365/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7365/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7364
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7364/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7364/comments
https://api.github.com/repos/huggingface/datasets/issues/7364/events
https://github.com/huggingface/datasets/issues/7364
2,776,929,268
I_kwDODunzps6lhJP0
7,364
API endpoints for gated dataset access requests
{ "avatar_url": "https://avatars.githubusercontent.com/u/6140840?v=4", "events_url": "https://api.github.com/users/jerome-white/events{/privacy}", "followers_url": "https://api.github.com/users/jerome-white/followers", "following_url": "https://api.github.com/users/jerome-white/following{/other_user}", "gists_url": "https://api.github.com/users/jerome-white/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/jerome-white", "id": 6140840, "login": "jerome-white", "node_id": "MDQ6VXNlcjYxNDA4NDA=", "organizations_url": "https://api.github.com/users/jerome-white/orgs", "received_events_url": "https://api.github.com/users/jerome-white/received_events", "repos_url": "https://api.github.com/users/jerome-white/repos", "site_admin": false, "starred_url": "https://api.github.com/users/jerome-white/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jerome-white/subscriptions", "type": "User", "url": "https://api.github.com/users/jerome-white", "user_view_type": "public" }
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
closed
false
null
[]
null
[]
"2025-01-09T06:21:20"
"2025-01-09T11:17:40"
2025-01-09T11:17:20
NONE
{ "completed": 0, "percent_completed": 0, "total": 0 }
null
### Feature request I would like a programatic way of requesting access to gated datasets. The current solution to gain access forces me to visit a website and physically click an "agreement" button (as per the [documentation](https://huggingface.co./docs/hub/en/datasets-gated#access-gated-datasets-as-a-user)). An ideal approach would be HF API download methods that negotiate access on my behalf based on information from my CLI login and/or token. I realise that may be naive given the various types of access semantics available to dataset authors (automatic versus manual approval, for example) and complexities it might add to existing methods, but something along those lines would be nice. Perhaps using the `*_access_request` methods available to dataset authors can be a precedent; see [`reject_access_request`](https://huggingface.co./docs/huggingface_hub/main/en/package_reference/hf_api#huggingface_hub.HfApi.reject_access_request) for example. ### Motivation When trying to download files from a gated dataset, I'm met with a `GatedRepoError` and instructed to visit the repository's website to gain access: ``` Cannot access gated repo for url https://huggingface.co./datasets/open-llm-leaderboard/meta-llama__Meta-Llama-3.1-70B-Instruct-details/resolve/main/meta-llama__Meta-Llama-3.1-70B-Instruct/samples_leaderboard_math_precalculus_hard_2024-07-19T18-47-29.522341.jsonl. Access to dataset open-llm-leaderboard/meta-llama__Meta-Llama-3.1-70B-Instruct-details is restricted and you are not in the authorized list. Visit https://huggingface.co./datasets/open-llm-leaderboard/meta-llama__Meta-Llama-3.1-70B-Instruct-details to ask for access. ``` This makes task automation extremely difficult. For example, I'm interested in studying sample-level responses of models on the LLM leaderboard -- how they answered particular questions on a given evaluation framework. As I come across more and more participants that gate their data, it's becoming unwieldy to continue my work (there over 2,000 participants, so in the worst case that's the number of website visits I'd need to manually undertake). One approach is use Selenium to react to the `GatedRepoError`, but that seems like overkill; and a potential violation HF terms of service (?). As mentioned in the previous section, there seems to be an [API for gated dataset owners](https://huggingface.co./docs/hub/en/datasets-gated#via-the-api) to managed access requests, and thus some appetite for allowing automated management of gating. This feature request is to extend that to dataset users. ### Your contribution Whether I can help depends on a few things; one being the complexity of the underlying gated access design. If this feature request is accepted I am open to being involved in discussions and testing, and even development under the right time-outcome tradeoff.
{ "avatar_url": "https://avatars.githubusercontent.com/u/6140840?v=4", "events_url": "https://api.github.com/users/jerome-white/events{/privacy}", "followers_url": "https://api.github.com/users/jerome-white/followers", "following_url": "https://api.github.com/users/jerome-white/following{/other_user}", "gists_url": "https://api.github.com/users/jerome-white/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/jerome-white", "id": 6140840, "login": "jerome-white", "node_id": "MDQ6VXNlcjYxNDA4NDA=", "organizations_url": "https://api.github.com/users/jerome-white/orgs", "received_events_url": "https://api.github.com/users/jerome-white/received_events", "repos_url": "https://api.github.com/users/jerome-white/repos", "site_admin": false, "starred_url": "https://api.github.com/users/jerome-white/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jerome-white/subscriptions", "type": "User", "url": "https://api.github.com/users/jerome-white", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7364/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7364/timeline
null
not_planned
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7363
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7363/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7363/comments
https://api.github.com/repos/huggingface/datasets/issues/7363/events
https://github.com/huggingface/datasets/issues/7363
2,774,090,012
I_kwDODunzps6lWUEc
7,363
ImportError: To support decoding images, please install 'Pillow'.
{ "avatar_url": "https://avatars.githubusercontent.com/u/1394644?v=4", "events_url": "https://api.github.com/users/jamessdixon/events{/privacy}", "followers_url": "https://api.github.com/users/jamessdixon/followers", "following_url": "https://api.github.com/users/jamessdixon/following{/other_user}", "gists_url": "https://api.github.com/users/jamessdixon/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/jamessdixon", "id": 1394644, "login": "jamessdixon", "node_id": "MDQ6VXNlcjEzOTQ2NDQ=", "organizations_url": "https://api.github.com/users/jamessdixon/orgs", "received_events_url": "https://api.github.com/users/jamessdixon/received_events", "repos_url": "https://api.github.com/users/jamessdixon/repos", "site_admin": false, "starred_url": "https://api.github.com/users/jamessdixon/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jamessdixon/subscriptions", "type": "User", "url": "https://api.github.com/users/jamessdixon", "user_view_type": "public" }
[]
open
false
null
[]
null
[]
"2025-01-08T02:22:57"
"2025-01-16T08:54:47"
1970-01-01T00:00:00
NONE
{ "completed": 0, "percent_completed": 0, "total": 0 }
null
### Describe the bug Following this tutorial locally using a macboko and VSCode: https://huggingface.co./docs/diffusers/en/tutorials/basic_training This line of code: for i, image in enumerate(dataset[:4]["image"]): throws: ImportError: To support decoding images, please install 'Pillow'. Pillow is installed. ### Steps to reproduce the bug Run the tutorial ### Expected behavior Images should be rendered ### Environment info MacBook, VSCode
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7363/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7363/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7362
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7362/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7362/comments
https://api.github.com/repos/huggingface/datasets/issues/7362/events
https://github.com/huggingface/datasets/issues/7362
2,773,731,829
I_kwDODunzps6lU8n1
7,362
HuggingFace CLI dataset download raises error
{ "avatar_url": "https://avatars.githubusercontent.com/u/3870355?v=4", "events_url": "https://api.github.com/users/ajayvohra2005/events{/privacy}", "followers_url": "https://api.github.com/users/ajayvohra2005/followers", "following_url": "https://api.github.com/users/ajayvohra2005/following{/other_user}", "gists_url": "https://api.github.com/users/ajayvohra2005/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/ajayvohra2005", "id": 3870355, "login": "ajayvohra2005", "node_id": "MDQ6VXNlcjM4NzAzNTU=", "organizations_url": "https://api.github.com/users/ajayvohra2005/orgs", "received_events_url": "https://api.github.com/users/ajayvohra2005/received_events", "repos_url": "https://api.github.com/users/ajayvohra2005/repos", "site_admin": false, "starred_url": "https://api.github.com/users/ajayvohra2005/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/ajayvohra2005/subscriptions", "type": "User", "url": "https://api.github.com/users/ajayvohra2005", "user_view_type": "public" }
[]
closed
false
null
[]
null
[]
"2025-01-07T21:03:30"
"2025-01-08T15:00:37"
2025-01-08T14:35:52
NONE
{ "completed": 0, "percent_completed": 0, "total": 0 }
null
### Describe the bug Trying to download Hugging Face datasets using Hugging Face CLI raises error. This error only started after December 27th, 2024. For example: ``` huggingface-cli download --repo-type dataset gboleda/wikicorpus Traceback (most recent call last): File "/home/ubuntu/test_venv/bin/huggingface-cli", line 8, in <module> sys.exit(main()) File "/home/ubuntu/test_venv/lib/python3.10/site-packages/huggingface_hub/commands/huggingface_cli.py", line 51, in main service.run() File "/home/ubuntu/test_venv/lib/python3.10/site-packages/huggingface_hub/commands/download.py", line 146, in run print(self._download()) # Print path to downloaded files File "/home/ubuntu/test_venv/lib/python3.10/site-packages/huggingface_hub/commands/download.py", line 180, in _download return snapshot_download( File "/home/ubuntu/test_venv/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 114, in _inner_fn return fn(*args, **kwargs) File "/home/ubuntu/test_venv/lib/python3.10/site-packages/huggingface_hub/_snapshot_download.py", line 164, in snapshot_download repo_info = api.repo_info(repo_id=repo_id, repo_type=repo_type, revision=revision, token=token) File "/home/ubuntu/test_venv/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 114, in _inner_fn return fn(*args, **kwargs) File "/home/ubuntu/test_venv/lib/python3.10/site-packages/huggingface_hub/hf_api.py", line 2491, in repo_info return method( File "/home/ubuntu/test_venv/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 114, in _inner_fn return fn(*args, **kwargs) File "/home/ubuntu/test_venv/lib/python3.10/site-packages/huggingface_hub/hf_api.py", line 2366, in dataset_info return DatasetInfo(**data) File "/home/ubuntu/test_venv/lib/python3.10/site-packages/huggingface_hub/hf_api.py", line 799, in __init__ self.tags = kwargs.pop("tags") KeyError: 'tags' ``` ### Steps to reproduce the bug ``` 1. huggingface-cli download --repo-type dataset gboleda/wikicorpus ``` ### Expected behavior There should be no error. ### Environment info - `datasets` version: 2.19.1 - Platform: Linux-6.8.0-1015-aws-x86_64-with-glibc2.35 - Python version: 3.10.12 - `huggingface_hub` version: 0.23.5 - PyArrow version: 18.1.0 - Pandas version: 2.2.3 - `fsspec` version: 2024.3.1
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/7362/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7362/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7361
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7361/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7361/comments
https://api.github.com/repos/huggingface/datasets/issues/7361/events
https://github.com/huggingface/datasets/pull/7361
2,771,859,244
PR_kwDODunzps6G4t2p
7,361
Fix lock permission
{ "avatar_url": "https://avatars.githubusercontent.com/u/11530592?v=4", "events_url": "https://api.github.com/users/cih9088/events{/privacy}", "followers_url": "https://api.github.com/users/cih9088/followers", "following_url": "https://api.github.com/users/cih9088/following{/other_user}", "gists_url": "https://api.github.com/users/cih9088/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/cih9088", "id": 11530592, "login": "cih9088", "node_id": "MDQ6VXNlcjExNTMwNTky", "organizations_url": "https://api.github.com/users/cih9088/orgs", "received_events_url": "https://api.github.com/users/cih9088/received_events", "repos_url": "https://api.github.com/users/cih9088/repos", "site_admin": false, "starred_url": "https://api.github.com/users/cih9088/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/cih9088/subscriptions", "type": "User", "url": "https://api.github.com/users/cih9088", "user_view_type": "public" }
[]
open
false
null
[]
null
[]
"2025-01-07T04:15:53"
"2025-01-07T04:49:46"
1970-01-01T00:00:00
NONE
{ "completed": 0, "percent_completed": 0, "total": 0 }
null
All files except lock file have proper permission obeying `ACL` property if it is set. If the cache directory has `ACL` property, it should be respected instead of just using `umask` for permission. To fix it, just create a lock file and pass the created `mode`. By creating a lock file with `touch()` before `FileLock` create it with `mode`, - if `ACL` is not set, same as before - if `ACL` is set, `ACL` is respected If it is acceptable, it should be also applied to [`huggingface_hub`](https://github.com/huggingface/huggingface_hub/blob/2702ec2a2bd0124cc1fddfd72ccb1297b2478148/src/huggingface_hub/utils/_fixes.py#L95) I guess.
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7361/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7361/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7361.diff", "html_url": "https://github.com/huggingface/datasets/pull/7361", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7361.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7361" }
true
https://api.github.com/repos/huggingface/datasets/issues/7360
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7360/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7360/comments
https://api.github.com/repos/huggingface/datasets/issues/7360/events
https://github.com/huggingface/datasets/issues/7360
2,771,751,406
I_kwDODunzps6lNZHu
7,360
error when loading dataset in Hugging Face: NoneType error is not callable
{ "avatar_url": "https://avatars.githubusercontent.com/u/189343338?v=4", "events_url": "https://api.github.com/users/nanu23333/events{/privacy}", "followers_url": "https://api.github.com/users/nanu23333/followers", "following_url": "https://api.github.com/users/nanu23333/following{/other_user}", "gists_url": "https://api.github.com/users/nanu23333/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/nanu23333", "id": 189343338, "login": "nanu23333", "node_id": "U_kgDOC0kmag", "organizations_url": "https://api.github.com/users/nanu23333/orgs", "received_events_url": "https://api.github.com/users/nanu23333/received_events", "repos_url": "https://api.github.com/users/nanu23333/repos", "site_admin": false, "starred_url": "https://api.github.com/users/nanu23333/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/nanu23333/subscriptions", "type": "User", "url": "https://api.github.com/users/nanu23333", "user_view_type": "public" }
[]
open
false
null
[]
null
[]
"2025-01-07T02:11:36"
"2025-01-10T10:44:38"
1970-01-01T00:00:00
NONE
{ "completed": 0, "percent_completed": 0, "total": 0 }
null
### Describe the bug I met an error when running a notebook provide by Hugging Face, and met the error. ``` --------------------------------------------------------------------------- TypeError Traceback (most recent call last) Cell In[2], line 5 3 # Load the enhancers dataset from the InstaDeep Hugging Face ressources 4 dataset_name = "enhancers_types" ----> 5 train_dataset_enhancers = load_dataset( 6 "InstaDeepAI/nucleotide_transformer_downstream_tasks_revised", 7 dataset_name, 8 split="train", 9 streaming= False, 10 ) 11 test_dataset_enhancers = load_dataset( 12 "InstaDeepAI/nucleotide_transformer_downstream_tasks_revised", 13 dataset_name, 14 split="test", 15 streaming= False, 16 ) File /public/home/hhl/miniconda3/envs/transformer/lib/python3.9/site-packages/datasets/load.py:2129, in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, verification_mode, keep_in_memory, save_infos, revision, token, streaming, num_proc, storage_options, trust_remote_code, **config_kwargs) 2124 verification_mode = VerificationMode( 2125 (verification_mode or VerificationMode.BASIC_CHECKS) if not save_infos else VerificationMode.ALL_CHECKS 2126 ) 2128 # Create a dataset builder -> 2129 builder_instance = load_dataset_builder( 2130 path=path, 2131 name=name, 2132 data_dir=data_dir, 2133 data_files=data_files, 2134 cache_dir=cache_dir, 2135 features=features, 2136 download_config=download_config, 2137 download_mode=download_mode, 2138 revision=revision, 2139 token=token, 2140 storage_options=storage_options, 2141 trust_remote_code=trust_remote_code, 2142 _require_default_config_name=name is None, 2143 **config_kwargs, 2144 ) 2146 # Return iterable dataset in case of streaming 2147 if streaming: File /public/home/hhl/miniconda3/envs/transformer/lib/python3.9/site-packages/datasets/load.py:1886, in load_dataset_builder(path, name, data_dir, data_files, cache_dir, features, download_config, download_mode, revision, token, storage_options, trust_remote_code, _require_default_config_name, **config_kwargs) 1884 builder_cls = get_dataset_builder_class(dataset_module, dataset_name=dataset_name) 1885 # Instantiate the dataset builder -> 1886 builder_instance: DatasetBuilder = builder_cls( 1887 cache_dir=cache_dir, 1888 dataset_name=dataset_name, 1889 config_name=config_name, 1890 data_dir=data_dir, 1891 data_files=data_files, 1892 hash=dataset_module.hash, 1893 info=info, 1894 features=features, 1895 token=token, 1896 storage_options=storage_options, 1897 **builder_kwargs, 1898 **config_kwargs, 1899 ) 1900 builder_instance._use_legacy_cache_dir_if_possible(dataset_module) 1902 return builder_instance TypeError: 'NoneType' object is not callable ``` I have checked my internet, it worked well. And the dataset name was just copied from the Hugging Face. Totally no idea what is wrong! ### Steps to reproduce the bug To reproduce the bug you may run ``` from datasets import load_dataset, Dataset # Load the enhancers dataset from the InstaDeep Hugging Face ressources dataset_name = "enhancers_types" train_dataset_enhancers = load_dataset( "InstaDeepAI/nucleotide_transformer_downstream_tasks_revised", dataset_name, split="train", streaming= False, ) test_dataset_enhancers = load_dataset( "InstaDeepAI/nucleotide_transformer_downstream_tasks_revised", dataset_name, split="test", streaming= False, ) ``` ### Expected behavior 1. what may be the reasons of the error 2. how can I fine which reason lead to the error 3. how can I save the problem ### Environment info ``` - `datasets` version: 3.2.0 - Platform: Linux-5.15.0-117-generic-x86_64-with-glibc2.31 - Python version: 3.9.21 - `huggingface_hub` version: 0.27.0 - PyArrow version: 18.1.0 - Pandas version: 2.2.3 - `fsspec` version: 2024.9.0 ```
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7360/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7360/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7359
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7359/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7359/comments
https://api.github.com/repos/huggingface/datasets/issues/7359/events
https://github.com/huggingface/datasets/issues/7359
2,771,137,842
I_kwDODunzps6lLDUy
7,359
There are multiple 'mteb/arguana' configurations in the cache: default, corpus, queries with HF_HUB_OFFLINE=1
{ "avatar_url": "https://avatars.githubusercontent.com/u/723146?v=4", "events_url": "https://api.github.com/users/Bhavya6187/events{/privacy}", "followers_url": "https://api.github.com/users/Bhavya6187/followers", "following_url": "https://api.github.com/users/Bhavya6187/following{/other_user}", "gists_url": "https://api.github.com/users/Bhavya6187/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/Bhavya6187", "id": 723146, "login": "Bhavya6187", "node_id": "MDQ6VXNlcjcyMzE0Ng==", "organizations_url": "https://api.github.com/users/Bhavya6187/orgs", "received_events_url": "https://api.github.com/users/Bhavya6187/received_events", "repos_url": "https://api.github.com/users/Bhavya6187/repos", "site_admin": false, "starred_url": "https://api.github.com/users/Bhavya6187/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/Bhavya6187/subscriptions", "type": "User", "url": "https://api.github.com/users/Bhavya6187", "user_view_type": "public" }
[]
open
false
null
[]
null
[]
"2025-01-06T17:42:49"
"2025-01-06T17:43:31"
1970-01-01T00:00:00
NONE
{ "completed": 0, "percent_completed": 0, "total": 0 }
null
### Describe the bug Hey folks, I am trying to run this code - ```python from datasets import load_dataset, get_dataset_config_names ds = load_dataset("mteb/arguana") ``` with HF_HUB_OFFLINE=1 But I get the following error - ```python Using the latest cached version of the dataset since mteb/arguana couldn't be found on the Hugging Face Hub (offline mode is enabled). --------------------------------------------------------------------------- ValueError Traceback (most recent call last) Cell In[2], line 1 ----> 1 ds = load_dataset("mteb/arguana") File ~/env/lib/python3.10/site-packages/datasets/load.py:2129, in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, verification_mode, keep_in_memory, save_infos, revision, token, streaming, num_proc, storage_options, trust_remote_code, **config_kwargs) 2124 verification_mode = VerificationMode( 2125 (verification_mode or VerificationMode.BASIC_CHECKS) if not save_infos else VerificationMode.ALL_CHECKS 2126 ) 2128 # Create a dataset builder -> 2129 builder_instance = load_dataset_builder( 2130 path=path, 2131 name=name, 2132 data_dir=data_dir, 2133 data_files=data_files, 2134 cache_dir=cache_dir, 2135 features=features, 2136 download_config=download_config, 2137 download_mode=download_mode, 2138 revision=revision, 2139 token=token, 2140 storage_options=storage_options, 2141 trust_remote_code=trust_remote_code, 2142 _require_default_config_name=name is None, 2143 **config_kwargs, 2144 ) 2146 # Return iterable dataset in case of streaming 2147 if streaming: File ~/env/lib/python3.10/site-packages/datasets/load.py:1886, in load_dataset_builder(path, name, data_dir, data_files, cache_dir, features, download_config, download_mode, revision, token, storage_options, trust_remote_code, _require_default_config_name, **config_kwargs) 1884 builder_cls = get_dataset_builder_class(dataset_module, dataset_name=dataset_name) 1885 # Instantiate the dataset builder -> 1886 builder_instance: DatasetBuilder = builder_cls( 1887 cache_dir=cache_dir, 1888 dataset_name=dataset_name, 1889 config_name=config_name, 1890 data_dir=data_dir, 1891 data_files=data_files, 1892 hash=dataset_module.hash, 1893 info=info, 1894 features=features, 1895 token=token, 1896 storage_options=storage_options, 1897 **builder_kwargs, 1898 **config_kwargs, 1899 ) 1900 builder_instance._use_legacy_cache_dir_if_possible(dataset_module) 1902 return builder_instance File ~/env/lib/python3.10/site-packages/datasets/packaged_modules/cache/cache.py:124, in Cache.__init__(self, cache_dir, dataset_name, config_name, version, hash, base_path, info, features, token, repo_id, data_files, data_dir, storage_options, writer_batch_size, **config_kwargs) 122 config_kwargs["data_dir"] = data_dir 123 if hash == "auto" and version == "auto": --> 124 config_name, version, hash = _find_hash_in_cache( 125 dataset_name=repo_id or dataset_name, 126 config_name=config_name, 127 cache_dir=cache_dir, 128 config_kwargs=config_kwargs, 129 custom_features=features, 130 ) 131 elif hash == "auto" or version == "auto": 132 raise NotImplementedError("Pass both hash='auto' and version='auto' instead") File ~/env/lib/python3.10/site-packages/datasets/packaged_modules/cache/cache.py:84, in _find_hash_in_cache(dataset_name, config_name, cache_dir, config_kwargs, custom_features) 72 other_configs = [ 73 Path(_cached_directory_path).parts[-3] 74 for _cached_directory_path in glob.glob(os.path.join(cached_datasets_directory_path_root, "*", version, hash)) (...) 81 ) 82 ] 83 if not config_id and len(other_configs) > 1: ---> 84 raise ValueError( 85 f"There are multiple '{dataset_name}' configurations in the cache: {', '.join(other_configs)}" 86 f"\nPlease specify which configuration to reload from the cache, e.g." 87 f"\n\tload_dataset('{dataset_name}', '{other_configs[0]}')" 88 ) 89 config_name = cached_directory_path.parts[-3] 90 warning_msg = ( 91 f"Found the latest cached dataset configuration '{config_name}' at {cached_directory_path} " 92 f"(last modified on {time.ctime(_get_modification_time(cached_directory_path))})." 93 ) ValueError: There are multiple 'mteb/arguana' configurations in the cache: queries, corpus, default Please specify which configuration to reload from the cache, e.g. load_dataset('mteb/arguana', 'queries') ``` It works when I run the same code with HF_HUB_OFFLINE=0, but after the data is downloaded, I turn off the HF hub cache with HF_HUB_OFFLINE=1, and then this error appears. Are there some files I am missing with hub disabled? ### Steps to reproduce the bug from datasets import load_dataset, get_dataset_config_names ds = load_dataset("mteb/arguana") with HF_HUB_OFFLINE=1 (after already running it with HF_HUB_OFFLINE=0 and populating the datasets cache) ### Expected behavior Dataset loaded successfully as it does with HF_HUB_OFFLINE=1 ### Environment info - `datasets` version: 3.2.0 - Platform: Linux-5.15.148.2-2.cm2-x86_64-with-glibc2.35 - Python version: 3.10.14 - `huggingface_hub` version: 0.27.0 - PyArrow version: 17.0.0 - Pandas version: 2.2.3 - `fsspec` version: 2024.6.1
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7359/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7359/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7358
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7358/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7358/comments
https://api.github.com/repos/huggingface/datasets/issues/7358/events
https://github.com/huggingface/datasets/pull/7358
2,770,927,769
PR_kwDODunzps6G1kka
7,358
Fix remove_columns in the formatted case
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
open
false
null
[]
null
[]
"2025-01-06T15:44:23"
"2025-01-06T15:46:46"
1970-01-01T00:00:00
MEMBER
{ "completed": 0, "percent_completed": 0, "total": 0 }
null
`remove_columns` had no effect when running a function in `.map()` on dataset that is formatted This aligns the logic of `map()` with the non formatted case and also with with https://github.com/huggingface/datasets/pull/7353
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7358/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7358/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7358.diff", "html_url": "https://github.com/huggingface/datasets/pull/7358", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7358.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7358" }
true
https://api.github.com/repos/huggingface/datasets/issues/7357
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7357/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7357/comments
https://api.github.com/repos/huggingface/datasets/issues/7357/events
https://github.com/huggingface/datasets/issues/7357
2,770,456,127
I_kwDODunzps6lIc4_
7,357
Python process aborded with GIL issue when using image dataset
{ "avatar_url": "https://avatars.githubusercontent.com/u/25342812?v=4", "events_url": "https://api.github.com/users/AlexKoff88/events{/privacy}", "followers_url": "https://api.github.com/users/AlexKoff88/followers", "following_url": "https://api.github.com/users/AlexKoff88/following{/other_user}", "gists_url": "https://api.github.com/users/AlexKoff88/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/AlexKoff88", "id": 25342812, "login": "AlexKoff88", "node_id": "MDQ6VXNlcjI1MzQyODEy", "organizations_url": "https://api.github.com/users/AlexKoff88/orgs", "received_events_url": "https://api.github.com/users/AlexKoff88/received_events", "repos_url": "https://api.github.com/users/AlexKoff88/repos", "site_admin": false, "starred_url": "https://api.github.com/users/AlexKoff88/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/AlexKoff88/subscriptions", "type": "User", "url": "https://api.github.com/users/AlexKoff88", "user_view_type": "public" }
[]
open
false
null
[]
null
[]
"2025-01-06T11:29:30"
"2025-01-09T14:00:46"
1970-01-01T00:00:00
NONE
{ "completed": 0, "percent_completed": 0, "total": 0 }
null
### Describe the bug The issue is visible only with the latest `datasets==3.2.0`. When using image dataset the Python process gets aborted right before the exit with the following error: ``` Fatal Python error: PyGILState_Release: thread state 0x7fa1f409ade0 must be current when releasing Python runtime state: finalizing (tstate=0x0000000000ad2958) Thread 0x00007fa33d157740 (most recent call first): <no Python frame> Extension modules: numpy.core._multiarray_umath, numpy.core._multiarray_tests, numpy.linalg._umath_linalg, numpy.fft._pocketfft_internal, numpy.random._common, numpy.random.bit_generator, numpy.random._boun ded_integers, numpy.random._mt19937, numpy.random.mtrand, numpy.random._philox, numpy.random._pcg64, numpy.random._sfc64, numpy.random._generator, pyarrow.lib, pandas._libs.tslibs.ccalendar, pandas._libs.ts libs.np_datetime, pandas._libs.tslibs.dtypes, pandas._libs.tslibs.base, pandas._libs.tslibs.nattype, pandas._libs.tslibs.timezones, pandas._libs.tslibs.fields, pandas._libs.tslibs.timedeltas, pandas._libs.t slibs.tzconversion, pandas._libs.tslibs.timestamps, pandas._libs.properties, pandas._libs.tslibs.offsets, pandas._libs.tslibs.strptime, pandas._libs.tslibs.parsing, pandas._libs.tslibs.conversion, pandas._l ibs.tslibs.period, pandas._libs.tslibs.vectorized, pandas._libs.ops_dispatch, pandas._libs.missing, pandas._libs.hashtable, pandas._libs.algos, pandas._libs.interval, pandas._libs.lib, pyarrow._compute, pan das._libs.ops, pandas._libs.hashing, pandas._libs.arrays, pandas._libs.tslib, pandas._libs.sparse, pandas._libs.internals, pandas._libs.indexing, pandas._libs.index, pandas._libs.writers, pandas._libs.join, pandas._libs.window.aggregations, pandas._libs.window.indexers, pandas._libs.reshape, pandas._libs.groupby, pandas._libs.json, pandas._libs.parsers, pandas._libs.testing, charset_normalizer.md, requests.pa ckages.charset_normalizer.md, requests.packages.chardet.md, yaml._yaml, markupsafe._speedups, PIL._imaging, torch._C, torch._C._dynamo.autograd_compiler, torch._C._dynamo.eval_frame, torch._C._dynamo.guards , torch._C._dynamo.utils, torch._C._fft, torch._C._linalg, torch._C._nested, torch._C._nn, torch._C._sparse, torch._C._special, sentencepiece._sentencepiece, sklearn.__check_build._check_build, psutil._psut il_linux, psutil._psutil_posix, scipy._lib._ccallback_c, scipy.sparse._sparsetools, _csparsetools, scipy.sparse._csparsetools, scipy.linalg._fblas, scipy.linalg._flapack, scipy.linalg.cython_lapack, scipy.l inalg._cythonized_array_utils, scipy.linalg._solve_toeplitz, scipy.linalg._decomp_lu_cython, scipy.linalg._matfuncs_sqrtm_triu, scipy.linalg.cython_blas, scipy.linalg._matfuncs_expm, scipy.linalg._decomp_up date, scipy.sparse.linalg._dsolve._superlu, scipy.sparse.linalg._eigen.arpack._arpack, scipy.sparse.linalg._propack._spropack, scipy.sparse.linalg._propack._dpropack, scipy.sparse.linalg._propack._cpropack, scipy.sparse.linalg._propack._zpropack, scipy.sparse.csgraph._tools, scipy.sparse.csgraph._shortest_path, scipy.sparse.csgraph._traversal, scipy.sparse.csgraph._min_spanning_tree, scipy.sparse.csgraph._flo w, scipy.sparse.csgraph._matching, scipy.sparse.csgraph._reordering, scipy.special._ufuncs_cxx, scipy.special._ufuncs, scipy.special._specfun, scipy.special._comb, scipy.special._ellip_harm_2, scipy.spatial ._ckdtree, scipy._lib.messagestream, scipy.spatial._qhull, scipy.spatial._voronoi, scipy.spatial._distance_wrap, scipy.spatial._hausdorff, scipy.spatial.transform._rotation, scipy.optimize._group_columns, s cipy.optimize._trlib._trlib, scipy.optimize._lbfgsb, _moduleTNC, scipy.optimize._moduleTNC, scipy.optimize._cobyla, scipy.optimize._slsqp, scipy.optimize._minpack, scipy.optimize._lsq.givens_elimination, sc ipy.optimize._zeros, scipy.optimize._highs.cython.src._highs_wrapper, scipy.optimize._highs._highs_wrapper, scipy.optimize._highs.cython.src._highs_constants, scipy.optimize._highs._highs_constants, scipy.l inalg._interpolative, scipy.optimize._bglu_dense, scipy.optimize._lsap, scipy.optimize._direct, scipy.integrate._odepack, scipy.integrate._quadpack, scipy.integrate._vode, scipy.integrate._dop, scipy.integr ate._lsoda, scipy.interpolate._fitpack, scipy.interpolate._dfitpack, scipy.interpolate._bspl, scipy.interpolate._ppoly, scipy.interpolate.interpnd, scipy.interpolate._rbfinterp_pythran, scipy.interpolate._r gi_cython, scipy.special.cython_special, scipy.stats._stats, scipy.stats._biasedurn, scipy.stats._levy_stable.levyst, scipy.stats._stats_pythran, scipy._lib._uarray._uarray, scipy.stats._ansari_swilk_statis tics, scipy.stats._sobol, scipy.stats._qmc_cy, scipy.stats._mvn, scipy.stats._rcont.rcont, scipy.stats._unuran.unuran_wrapper, scipy.ndimage._nd_image, _ni_label, scipy.ndimage._ni_label, sklearn.utils._isf inite, sklearn.utils.sparsefuncs_fast, sklearn.utils.murmurhash, sklearn.utils._openmp_helpers, sklearn.metrics.cluster._expected_mutual_info_fast, sklearn.preprocessing._csr_polynomial_expansion, sklearn.p reprocessing._target_encoder_fast, sklearn.metrics._dist_metrics, sklearn.metrics._pairwise_distances_reduction._datasets_pair, sklearn.utils._cython_blas, sklearn.metrics._pairwise_distances_reduction._bas e, sklearn.metrics._pairwise_distances_reduction._middle_term_computer, sklearn.utils._heap, sklearn.utils._sorting, sklearn.metrics._pairwise_distances_reduction._argkmin, sklearn.metrics._pairwise_distanc es_reduction._argkmin_classmode, sklearn.utils._vector_sentinel, sklearn.metrics._pairwise_distances_reduction._radius_neighbors, sklearn.metrics._pairwise_distances_reduction._radius_neighbors_classmode, s klearn.metrics._pairwise_fast, PIL._imagingft, google._upb._message, h5py._errors, h5py.defs, h5py._objects, h5py.h5, h5py.utils, h5py.h5t, h5py.h5s, h5py.h5ac, h5py.h5p, h5py.h5r, h5py._proxy, h5py._conv, h5py.h5z, h5py.h5a, h5py.h5d, h5py.h5ds, h5py.h5g, h5py.h5i, h5py.h5o, h5py.h5f, h5py.h5fd, h5py.h5pl, h5py.h5l, h5py._selector, _cffi_backend, pyarrow._parquet, pyarrow._fs, pyarrow._azurefs, pyarrow._hdfs , pyarrow._gcsfs, pyarrow._s3fs, multidict._multidict, propcache._helpers_c, yarl._quoting_c, aiohttp._helpers, aiohttp._http_writer, aiohttp._http_parser, aiohttp._websocket, frozenlist._frozenlist, xxhash ._xxhash, pyarrow._json, pyarrow._acero, pyarrow._csv, pyarrow._dataset, pyarrow._dataset_orc, pyarrow._parquet_encryption, pyarrow._dataset_parquet_encryption, pyarrow._dataset_parquet, regex._regex, scipy .io.matlab._mio_utils, scipy.io.matlab._streams, scipy.io.matlab._mio5_utils, PIL._imagingmath, PIL._webp (total: 236) Aborted (core dumped) ```an ### Steps to reproduce the bug Install `datasets==3.2.0` Run the following script: ```python import datasets DATASET_NAME = "phiyodr/InpaintCOCO" NUM_SAMPLES = 10 def preprocess_fn(example): return { "prompts": example["inpaint_caption"], "images": example["coco_image"], "masks": example["mask"], } default_dataset = datasets.load_dataset( DATASET_NAME, split="test", streaming=True ).filter(lambda example: example["inpaint_caption"] != "").take(NUM_SAMPLES) test_data = default_dataset.map( lambda x: preprocess_fn(x), remove_columns=default_dataset.column_names ) for data in test_data: print(data["prompts"]) `` ### Expected behavior The script should not hang or crash. ### Environment info - `datasets` version: 3.2.0 - Platform: Linux-5.15.0-50-generic-x86_64-with-glibc2.31 - Python version: 3.11.0 - `huggingface_hub` version: 0.25.1 - PyArrow version: 17.0.0 - Pandas version: 2.2.3 - `fsspec` version: 2024.2.0
null
{ "+1": 1, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 1, "url": "https://api.github.com/repos/huggingface/datasets/issues/7357/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7357/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7356
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7356/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7356/comments
https://api.github.com/repos/huggingface/datasets/issues/7356/events
https://github.com/huggingface/datasets/issues/7356
2,770,095,103
I_kwDODunzps6lHEv_
7,356
How about adding a feature to pass the key when performing map on DatasetDict?
{ "avatar_url": "https://avatars.githubusercontent.com/u/93233241?v=4", "events_url": "https://api.github.com/users/jp1924/events{/privacy}", "followers_url": "https://api.github.com/users/jp1924/followers", "following_url": "https://api.github.com/users/jp1924/following{/other_user}", "gists_url": "https://api.github.com/users/jp1924/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/jp1924", "id": 93233241, "login": "jp1924", "node_id": "U_kgDOBY6gWQ", "organizations_url": "https://api.github.com/users/jp1924/orgs", "received_events_url": "https://api.github.com/users/jp1924/received_events", "repos_url": "https://api.github.com/users/jp1924/repos", "site_admin": false, "starred_url": "https://api.github.com/users/jp1924/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jp1924/subscriptions", "type": "User", "url": "https://api.github.com/users/jp1924", "user_view_type": "public" }
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
open
false
null
[]
null
[]
"2025-01-06T08:13:52"
"2025-01-13T14:30:48"
1970-01-01T00:00:00
NONE
{ "completed": 0, "percent_completed": 0, "total": 0 }
null
### Feature request Add a feature to pass the key of the DatasetDict when performing map ### Motivation I often preprocess using map on DatasetDict. Sometimes, I need to preprocess train and valid data differently depending on the task. So, I thought it would be nice to pass the key (like train, valid) when performing map on DatasetDict. What do you think? ### Your contribution I can submit a pull request to add the feature to pass the key of the DatasetDict when performing map.
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7356/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7356/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7355
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7355/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7355/comments
https://api.github.com/repos/huggingface/datasets/issues/7355/events
https://github.com/huggingface/datasets/issues/7355
2,768,958,211
I_kwDODunzps6lCvMD
7,355
Not available datasets[audio] on python 3.13
{ "avatar_url": "https://avatars.githubusercontent.com/u/70306948?v=4", "events_url": "https://api.github.com/users/sergiosinlimites/events{/privacy}", "followers_url": "https://api.github.com/users/sergiosinlimites/followers", "following_url": "https://api.github.com/users/sergiosinlimites/following{/other_user}", "gists_url": "https://api.github.com/users/sergiosinlimites/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/sergiosinlimites", "id": 70306948, "login": "sergiosinlimites", "node_id": "MDQ6VXNlcjcwMzA2OTQ4", "organizations_url": "https://api.github.com/users/sergiosinlimites/orgs", "received_events_url": "https://api.github.com/users/sergiosinlimites/received_events", "repos_url": "https://api.github.com/users/sergiosinlimites/repos", "site_admin": false, "starred_url": "https://api.github.com/users/sergiosinlimites/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/sergiosinlimites/subscriptions", "type": "User", "url": "https://api.github.com/users/sergiosinlimites", "user_view_type": "public" }
[]
open
false
null
[]
null
[]
"2025-01-04T18:37:08"
"2025-01-10T10:46:00"
1970-01-01T00:00:00
NONE
{ "completed": 0, "percent_completed": 0, "total": 0 }
null
### Describe the bug This is the error I got, it seems numba package does not support python 3.13 PS C:\Users\sergi\Documents> pip install datasets[audio] Defaulting to user installation because normal site-packages is not writeable Collecting datasets[audio] Using cached datasets-3.2.0-py3-none-any.whl.metadata (20 kB) ... (OTHER PACKAGES) Collecting numba>=0.51.0 (from librosa->datasets[audio]) Downloading numba-0.60.0.tar.gz (2.7 MB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 2.7/2.7 MB 44.1 MB/s eta 0:00:00 Installing build dependencies ... done Getting requirements to build wheel ... error error: subprocess-exited-with-error × Getting requirements to build wheel did not run successfully. │ exit code: 1 ╰─> [24 lines of output] Traceback (most recent call last): File "C:\Program Files\WindowsApps\PythonSoftwareFoundation.Python.3.13_3.13.496.0_x64__qbz5n2kfra8p0\Lib\site-packages\pip\_vendor\pyproject_hooks\_in_process\_in_process.py", line 353, in <module> main() ~~~~^^ File "C:\Program Files\WindowsApps\PythonSoftwareFoundation.Python.3.13_3.13.496.0_x64__qbz5n2kfra8p0\Lib\site-packages\pip\_vendor\pyproject_hooks\_in_process\_in_process.py", line 335, in main json_out['return_val'] = hook(**hook_input['kwargs']) ~~~~^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Program Files\WindowsApps\PythonSoftwareFoundation.Python.3.13_3.13.496.0_x64__qbz5n2kfra8p0\Lib\site-packages\pip\_vendor\pyproject_hooks\_in_process\_in_process.py", line 118, in get_requires_for_build_wheel return hook(config_settings) File "C:\Users\sergi\AppData\Local\Temp\pip-build-env-yauns_qh\overlay\Lib\site-packages\setuptools\build_meta.py", line 334, in get_requires_for_build_wheel return self._get_build_requires(config_settings, requirements=[]) ~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Users\sergi\AppData\Local\Temp\pip-build-env-yauns_qh\overlay\Lib\site-packages\setuptools\build_meta.py", line 304, in _get_build_requires self.run_setup() ~~~~~~~~~~~~~~^^ RuntimeError: Cannot install on Python version 3.13.1; only versions >=3.9,<3.13 are supported. [end of output] note: This error originates from a subprocess, and is likely not a problem with pip. error: subprocess-exited-with-error × Getting requirements to build wheel did not run successfully. │ exit code: 1 ╰─> See above for output. ### Steps to reproduce the bug 1. install python >=3.13 2. !pip install datasets[audio] ### Expected behavior I needed datasets[audio] in the python 3.13 ### Environment info python 3.13.1
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7355/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7355/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7354
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7354/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7354/comments
https://api.github.com/repos/huggingface/datasets/issues/7354/events
https://github.com/huggingface/datasets/issues/7354
2,768,955,917
I_kwDODunzps6lCuoN
7,354
A module that was compiled using NumPy 1.x cannot be run in NumPy 2.0.2 as it may crash. To support both 1.x and 2.x versions of NumPy, modules must be compiled with NumPy 2.0. Some module may need to rebuild instead e.g. with 'pybind11>=2.12'.
{ "avatar_url": "https://avatars.githubusercontent.com/u/1394644?v=4", "events_url": "https://api.github.com/users/jamessdixon/events{/privacy}", "followers_url": "https://api.github.com/users/jamessdixon/followers", "following_url": "https://api.github.com/users/jamessdixon/following{/other_user}", "gists_url": "https://api.github.com/users/jamessdixon/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/jamessdixon", "id": 1394644, "login": "jamessdixon", "node_id": "MDQ6VXNlcjEzOTQ2NDQ=", "organizations_url": "https://api.github.com/users/jamessdixon/orgs", "received_events_url": "https://api.github.com/users/jamessdixon/received_events", "repos_url": "https://api.github.com/users/jamessdixon/repos", "site_admin": false, "starred_url": "https://api.github.com/users/jamessdixon/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jamessdixon/subscriptions", "type": "User", "url": "https://api.github.com/users/jamessdixon", "user_view_type": "public" }
[]
closed
false
null
[]
null
[]
"2025-01-04T18:30:17"
"2025-01-08T02:20:58"
2025-01-08T02:20:58
NONE
{ "completed": 0, "percent_completed": 0, "total": 0 }
null
### Describe the bug Following this tutorial: https://huggingface.co./docs/diffusers/en/tutorials/basic_training and running it locally using VSCode on my MacBook. The first line in the tutorial fails: from datasets import load_dataset dataset = load_dataset('huggan/smithsonian_butterflies_subset', split="train"). with this error: A module that was compiled using NumPy 1.x cannot be run in NumPy 2.0.2 as it may crash. To support both 1.x and 2.x versions of NumPy, modules must be compiled with NumPy 2.0. Some module may need to rebuild instead e.g. with 'pybind11>=2.12'. If you are a user of the module, the easiest solution will be to downgrade to 'numpy<2' or try to upgrade the affected module. We expect that some modules will need time to support NumPy 2. and ImportError: numpy.core.multiarray failed to import. Does from datasets import load_dataset really use NumPy 1.x? ### Steps to reproduce the bug Open VSCode. create a new venv. Create a new ipynb file. Import pip install diffusers[training] try to run this line of code: from datasets import load_dataset ### Expected behavior data is loaded ### Environment info ran this: datasets-cli env and got A module that was compiled using NumPy 1.x cannot be run in NumPy 2.0.2 as it may crash. To support both 1.x and 2.x versions of NumPy, modules must be compiled with NumPy 2.0. Some module may need to rebuild instead e.g. with 'pybind11>=2.12'. If you are a user of the module, the easiest solution will be to downgrade to 'numpy<2' or try to upgrade the affected module. We expect that some modules will need time to support NumPy 2.
{ "avatar_url": "https://avatars.githubusercontent.com/u/1394644?v=4", "events_url": "https://api.github.com/users/jamessdixon/events{/privacy}", "followers_url": "https://api.github.com/users/jamessdixon/followers", "following_url": "https://api.github.com/users/jamessdixon/following{/other_user}", "gists_url": "https://api.github.com/users/jamessdixon/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/jamessdixon", "id": 1394644, "login": "jamessdixon", "node_id": "MDQ6VXNlcjEzOTQ2NDQ=", "organizations_url": "https://api.github.com/users/jamessdixon/orgs", "received_events_url": "https://api.github.com/users/jamessdixon/received_events", "repos_url": "https://api.github.com/users/jamessdixon/repos", "site_admin": false, "starred_url": "https://api.github.com/users/jamessdixon/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/jamessdixon/subscriptions", "type": "User", "url": "https://api.github.com/users/jamessdixon", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7354/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7354/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7353
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7353/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7353/comments
https://api.github.com/repos/huggingface/datasets/issues/7353/events
https://github.com/huggingface/datasets/pull/7353
2,768,484,726
PR_kwDODunzps6Gtd6K
7,353
changes to MappedExamplesIterable to resolve #7345
{ "avatar_url": "https://avatars.githubusercontent.com/u/12157034?v=4", "events_url": "https://api.github.com/users/vttrifonov/events{/privacy}", "followers_url": "https://api.github.com/users/vttrifonov/followers", "following_url": "https://api.github.com/users/vttrifonov/following{/other_user}", "gists_url": "https://api.github.com/users/vttrifonov/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/vttrifonov", "id": 12157034, "login": "vttrifonov", "node_id": "MDQ6VXNlcjEyMTU3MDM0", "organizations_url": "https://api.github.com/users/vttrifonov/orgs", "received_events_url": "https://api.github.com/users/vttrifonov/received_events", "repos_url": "https://api.github.com/users/vttrifonov/repos", "site_admin": false, "starred_url": "https://api.github.com/users/vttrifonov/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/vttrifonov/subscriptions", "type": "User", "url": "https://api.github.com/users/vttrifonov", "user_view_type": "public" }
[]
closed
false
null
[]
null
[]
"2025-01-04T06:01:15"
"2025-01-07T11:56:41"
2025-01-07T11:56:41
CONTRIBUTOR
{ "completed": 0, "percent_completed": 0, "total": 0 }
null
modified `MappedExamplesIterable` and `test_iterable_dataset.py::test_mapped_examples_iterable_with_indices` fix #7345 @lhoestq
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7353/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7353/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7353.diff", "html_url": "https://github.com/huggingface/datasets/pull/7353", "merged_at": "2025-01-07T11:56:41Z", "patch_url": "https://github.com/huggingface/datasets/pull/7353.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7353" }
true
https://api.github.com/repos/huggingface/datasets/issues/7352
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7352/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7352/comments
https://api.github.com/repos/huggingface/datasets/issues/7352/events
https://github.com/huggingface/datasets/pull/7352
2,767,763,850
PR_kwDODunzps6GrBB5
7,352
fsspec 2024.12.0
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
closed
false
null
[]
null
[]
"2025-01-03T15:32:25"
"2025-01-03T15:34:54"
2025-01-03T15:34:11
MEMBER
{ "completed": 0, "percent_completed": 0, "total": 0 }
null
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7352/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7352/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7352.diff", "html_url": "https://github.com/huggingface/datasets/pull/7352", "merged_at": "2025-01-03T15:34:11Z", "patch_url": "https://github.com/huggingface/datasets/pull/7352.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7352" }
true
https://api.github.com/repos/huggingface/datasets/issues/7350
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7350/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7350/comments
https://api.github.com/repos/huggingface/datasets/issues/7350/events
https://github.com/huggingface/datasets/pull/7350
2,767,731,707
PR_kwDODunzps6Gq6Bf
7,350
Bump hfh to 0.24 to fix ci
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
closed
false
null
[]
null
[]
"2025-01-03T15:09:40"
"2025-01-03T15:12:17"
2025-01-03T15:10:27
MEMBER
{ "completed": 0, "percent_completed": 0, "total": 0 }
null
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7350/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7350/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7350.diff", "html_url": "https://github.com/huggingface/datasets/pull/7350", "merged_at": "2025-01-03T15:10:27Z", "patch_url": "https://github.com/huggingface/datasets/pull/7350.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7350" }
true
https://api.github.com/repos/huggingface/datasets/issues/7349
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7349/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7349/comments
https://api.github.com/repos/huggingface/datasets/issues/7349/events
https://github.com/huggingface/datasets/pull/7349
2,767,670,454
PR_kwDODunzps6GqseO
7,349
Webdataset special columns in last position
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
closed
false
null
[]
null
[]
"2025-01-03T14:32:15"
"2025-01-03T14:34:39"
2025-01-03T14:32:30
MEMBER
{ "completed": 0, "percent_completed": 0, "total": 0 }
null
Place columns "__key__" and "__url__" in last position in the Dataset Viewer since they are not the main content before: <img width="1012" alt="image" src="https://github.com/user-attachments/assets/b556c1fe-2674-4ba0-9643-c074aa9716fd" />
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7349/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7349/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7349.diff", "html_url": "https://github.com/huggingface/datasets/pull/7349", "merged_at": "2025-01-03T14:32:30Z", "patch_url": "https://github.com/huggingface/datasets/pull/7349.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7349" }
true
https://api.github.com/repos/huggingface/datasets/issues/7348
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7348/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7348/comments
https://api.github.com/repos/huggingface/datasets/issues/7348/events
https://github.com/huggingface/datasets/pull/7348
2,766,128,230
PR_kwDODunzps6Gldcy
7,348
Catch OSError for arrow
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
closed
false
null
[]
null
[]
"2025-01-02T14:30:00"
"2025-01-09T14:25:06"
2025-01-09T14:25:04
MEMBER
{ "completed": 0, "percent_completed": 0, "total": 0 }
null
fixes https://github.com/huggingface/datasets/issues/7346 (also updated `ruff` and appleid style changes)
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7348/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7348/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7348.diff", "html_url": "https://github.com/huggingface/datasets/pull/7348", "merged_at": "2025-01-09T14:25:04Z", "patch_url": "https://github.com/huggingface/datasets/pull/7348.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7348" }
true
https://api.github.com/repos/huggingface/datasets/issues/7347
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7347/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7347/comments
https://api.github.com/repos/huggingface/datasets/issues/7347/events
https://github.com/huggingface/datasets/issues/7347
2,760,282,339
I_kwDODunzps6khpDj
7,347
Converting Arrow to WebDataset TAR Format for Offline Use
{ "avatar_url": "https://avatars.githubusercontent.com/u/91370128?v=4", "events_url": "https://api.github.com/users/katie312/events{/privacy}", "followers_url": "https://api.github.com/users/katie312/followers", "following_url": "https://api.github.com/users/katie312/following{/other_user}", "gists_url": "https://api.github.com/users/katie312/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/katie312", "id": 91370128, "login": "katie312", "node_id": "MDQ6VXNlcjkxMzcwMTI4", "organizations_url": "https://api.github.com/users/katie312/orgs", "received_events_url": "https://api.github.com/users/katie312/received_events", "repos_url": "https://api.github.com/users/katie312/repos", "site_admin": false, "starred_url": "https://api.github.com/users/katie312/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/katie312/subscriptions", "type": "User", "url": "https://api.github.com/users/katie312", "user_view_type": "public" }
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
closed
false
null
[]
null
[]
"2024-12-27T01:40:44"
"2024-12-31T17:38:00"
2024-12-28T15:38:03
NONE
{ "completed": 0, "percent_completed": 0, "total": 0 }
null
### Feature request Hi, I've downloaded an Arrow-formatted dataset offline using the hugggingface's datasets library by: ``` import json from datasets import load_dataset dataset = load_dataset("pixparse/cc3m-wds") dataset.save_to_disk("./cc3m_1") ``` now I need to convert it to WebDataset's TAR format for offline data ingestion. Is there a straightforward method to achieve this conversion without an internet connection? Can I simply convert it by ``` tar -cvf ``` btw, when I tried: ``` import webdataset as wds from huggingface_hub import get_token from torch.utils.data import DataLoader hf_token = get_token() url = "https://huggingface.co./datasets/timm/imagenet-12k-wds/resolve/main/imagenet12k-train-{{0000..1023}}.tar" url = f"pipe:curl -s -L {url} -H 'Authorization:Bearer {hf_token}'" dataset = wds.WebDataset(url).decode() dataset.save_to_disk("./cc3m_webdataset") ``` error occured: ``` AttributeError: 'WebDataset' object has no attribute 'save_to_disk' ``` Thanks a lot! ### Motivation Converting Arrow to WebDataset TAR Format ### Your contribution No clue yet
{ "avatar_url": "https://avatars.githubusercontent.com/u/91370128?v=4", "events_url": "https://api.github.com/users/katie312/events{/privacy}", "followers_url": "https://api.github.com/users/katie312/followers", "following_url": "https://api.github.com/users/katie312/following{/other_user}", "gists_url": "https://api.github.com/users/katie312/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/katie312", "id": 91370128, "login": "katie312", "node_id": "MDQ6VXNlcjkxMzcwMTI4", "organizations_url": "https://api.github.com/users/katie312/orgs", "received_events_url": "https://api.github.com/users/katie312/received_events", "repos_url": "https://api.github.com/users/katie312/repos", "site_admin": false, "starred_url": "https://api.github.com/users/katie312/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/katie312/subscriptions", "type": "User", "url": "https://api.github.com/users/katie312", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7347/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7347/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7346
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7346/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7346/comments
https://api.github.com/repos/huggingface/datasets/issues/7346/events
https://github.com/huggingface/datasets/issues/7346
2,758,752,118
I_kwDODunzps6kbzd2
7,346
OSError: Invalid flatbuffers message.
{ "avatar_url": "https://avatars.githubusercontent.com/u/46232487?v=4", "events_url": "https://api.github.com/users/antecede/events{/privacy}", "followers_url": "https://api.github.com/users/antecede/followers", "following_url": "https://api.github.com/users/antecede/following{/other_user}", "gists_url": "https://api.github.com/users/antecede/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/antecede", "id": 46232487, "login": "antecede", "node_id": "MDQ6VXNlcjQ2MjMyNDg3", "organizations_url": "https://api.github.com/users/antecede/orgs", "received_events_url": "https://api.github.com/users/antecede/received_events", "repos_url": "https://api.github.com/users/antecede/repos", "site_admin": false, "starred_url": "https://api.github.com/users/antecede/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/antecede/subscriptions", "type": "User", "url": "https://api.github.com/users/antecede", "user_view_type": "public" }
[]
closed
false
null
[]
null
[]
"2024-12-25T11:38:52"
"2025-01-09T14:25:29"
2025-01-09T14:25:05
NONE
{ "completed": 0, "percent_completed": 0, "total": 0 }
null
### Describe the bug When loading a large 2D data (1000 × 1152) with a large number of (2,000 data in this case) in `load_dataset`, the error message `OSError: Invalid flatbuffers message` is reported. When only 300 pieces of data of this size (1000 × 1152) are stored, they can be loaded correctly. When 2,000 2D arrays are stored in each file, about 100 files are generated, each with a file size of about 5-6GB. But when 300 2D arrays are stored in each file, **about 600 files are generated, which is too many files**. ### Steps to reproduce the bug error: ```python --------------------------------------------------------------------------- OSError Traceback (most recent call last) Cell In[2], line 4 1 from datasets import Dataset 2 from datasets import load_dataset ----> 4 real_dataset = load_dataset("arrow", data_files='tensorData/real_ResidueTensor/*', split="train")#.with_format("torch") # , split="train" 5 # sim_dataset = load_dataset("arrow", data_files='tensorData/sim_ResidueTensor/*', split="train").with_format("torch") 6 real_dataset File [~/miniforge3/envs/esmIne3/lib/python3.12/site-packages/datasets/load.py:2151](http://localhost:8899/lab/tree/RTC%3Anew_world/esm3/~/miniforge3/envs/esmIne3/lib/python3.12/site-packages/datasets/load.py#line=2150), in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, verification_mode, keep_in_memory, save_infos, revision, token, streaming, num_proc, storage_options, trust_remote_code, **config_kwargs) 2148 return builder_instance.as_streaming_dataset(split=split) 2150 # Download and prepare data -> 2151 builder_instance.download_and_prepare( 2152 download_config=download_config, 2153 download_mode=download_mode, 2154 verification_mode=verification_mode, 2155 num_proc=num_proc, 2156 storage_options=storage_options, 2157 ) 2159 # Build dataset for splits 2160 keep_in_memory = ( 2161 keep_in_memory if keep_in_memory is not None else is_small_dataset(builder_instance.info.dataset_size) 2162 ) File [~/miniforge3/envs/esmIne3/lib/python3.12/site-packages/datasets/builder.py:924](http://localhost:8899/lab/tree/RTC%3Anew_world/esm3/~/miniforge3/envs/esmIne3/lib/python3.12/site-packages/datasets/builder.py#line=923), in DatasetBuilder.download_and_prepare(self, output_dir, download_config, download_mode, verification_mode, dl_manager, base_path, file_format, max_shard_size, num_proc, storage_options, **download_and_prepare_kwargs) 922 if num_proc is not None: 923 prepare_split_kwargs["num_proc"] = num_proc --> 924 self._download_and_prepare( 925 dl_manager=dl_manager, 926 verification_mode=verification_mode, 927 **prepare_split_kwargs, 928 **download_and_prepare_kwargs, 929 ) 930 # Sync info 931 self.info.dataset_size = sum(split.num_bytes for split in self.info.splits.values()) File [~/miniforge3/envs/esmIne3/lib/python3.12/site-packages/datasets/builder.py:978](http://localhost:8899/lab/tree/RTC%3Anew_world/esm3/~/miniforge3/envs/esmIne3/lib/python3.12/site-packages/datasets/builder.py#line=977), in DatasetBuilder._download_and_prepare(self, dl_manager, verification_mode, **prepare_split_kwargs) 976 split_dict = SplitDict(dataset_name=self.dataset_name) 977 split_generators_kwargs = self._make_split_generators_kwargs(prepare_split_kwargs) --> 978 split_generators = self._split_generators(dl_manager, **split_generators_kwargs) 980 # Checksums verification 981 if verification_mode == VerificationMode.ALL_CHECKS and dl_manager.record_checksums: File [~/miniforge3/envs/esmIne3/lib/python3.12/site-packages/datasets/packaged_modules/arrow/arrow.py:47](http://localhost:8899/lab/tree/RTC%3Anew_world/esm3/~/miniforge3/envs/esmIne3/lib/python3.12/site-packages/datasets/packaged_modules/arrow/arrow.py#line=46), in Arrow._split_generators(self, dl_manager) 45 with open(file, "rb") as f: 46 try: ---> 47 reader = pa.ipc.open_stream(f) 48 except pa.lib.ArrowInvalid: 49 reader = pa.ipc.open_file(f) File [~/miniforge3/envs/esmIne3/lib/python3.12/site-packages/pyarrow/ipc.py:190](http://localhost:8899/lab/tree/RTC%3Anew_world/esm3/~/miniforge3/envs/esmIne3/lib/python3.12/site-packages/pyarrow/ipc.py#line=189), in open_stream(source, options, memory_pool) 171 def open_stream(source, *, options=None, memory_pool=None): 172 """ 173 Create reader for Arrow streaming format. 174 (...) 188 A reader for the given source 189 """ --> 190 return RecordBatchStreamReader(source, options=options, 191 memory_pool=memory_pool) File [~/miniforge3/envs/esmIne3/lib/python3.12/site-packages/pyarrow/ipc.py:52](http://localhost:8899/lab/tree/RTC%3Anew_world/esm3/~/miniforge3/envs/esmIne3/lib/python3.12/site-packages/pyarrow/ipc.py#line=51), in RecordBatchStreamReader.__init__(self, source, options, memory_pool) 50 def __init__(self, source, *, options=None, memory_pool=None): 51 options = _ensure_default_ipc_read_options(options) ---> 52 self._open(source, options=options, memory_pool=memory_pool) File [~/miniforge3/envs/esmIne3/lib/python3.12/site-packages/pyarrow/ipc.pxi:1006](http://localhost:8899/lab/tree/RTC%3Anew_world/esm3/~/miniforge3/envs/esmIne3/lib/python3.12/site-packages/pyarrow/ipc.pxi#line=1005), in pyarrow.lib._RecordBatchStreamReader._open() File [~/miniforge3/envs/esmIne3/lib/python3.12/site-packages/pyarrow/error.pxi:155](http://localhost:8899/lab/tree/RTC%3Anew_world/esm3/~/miniforge3/envs/esmIne3/lib/python3.12/site-packages/pyarrow/error.pxi#line=154), in pyarrow.lib.pyarrow_internal_check_status() File [~/miniforge3/envs/esmIne3/lib/python3.12/site-packages/pyarrow/error.pxi:92](http://localhost:8899/lab/tree/RTC%3Anew_world/esm3/~/miniforge3/envs/esmIne3/lib/python3.12/site-packages/pyarrow/error.pxi#line=91), in pyarrow.lib.check_status() OSError: Invalid flatbuffers message. ``` reproduce:Here is just an example result, the real 2D matrix is the output of the ESM large model, and the matrix size is approximate ```python import numpy as np import pyarrow as pa random_arrays_list = [np.random.rand(1000, 1152) for _ in range(2000)] table = pa.Table.from_pydict({ 'tensor': [tensor.tolist() for tensor in random_arrays_list] }) import pyarrow.feather as feather feather.write_feather(table, 'test.arrow') from datasets import load_dataset dataset = load_dataset("arrow", data_files='test.arrow', split="train") ``` ### Expected behavior `load_dataset` load the dataset as normal as `feather.read_feather` ```python import pyarrow.feather as feather feather.read_feather('tensorData/real_ResidueTensor/real_tensor_1.arrow') ``` Plus `load_dataset("parquet", data_files='test.arrow', split="train")` works fine ### Environment info - `datasets` version: 3.2.0 - Platform: Linux-6.8.0-49-generic-x86_64-with-glibc2.39 - Python version: 3.12.3 - `huggingface_hub` version: 0.26.5 - PyArrow version: 18.1.0 - Pandas version: 2.2.3 - `fsspec` version: 2024.9.0
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7346/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7346/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7345
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7345/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7345/comments
https://api.github.com/repos/huggingface/datasets/issues/7345/events
https://github.com/huggingface/datasets/issues/7345
2,758,585,709
I_kwDODunzps6kbK1t
7,345
Different behaviour of IterableDataset.map vs Dataset.map with remove_columns
{ "avatar_url": "https://avatars.githubusercontent.com/u/12157034?v=4", "events_url": "https://api.github.com/users/vttrifonov/events{/privacy}", "followers_url": "https://api.github.com/users/vttrifonov/followers", "following_url": "https://api.github.com/users/vttrifonov/following{/other_user}", "gists_url": "https://api.github.com/users/vttrifonov/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/vttrifonov", "id": 12157034, "login": "vttrifonov", "node_id": "MDQ6VXNlcjEyMTU3MDM0", "organizations_url": "https://api.github.com/users/vttrifonov/orgs", "received_events_url": "https://api.github.com/users/vttrifonov/received_events", "repos_url": "https://api.github.com/users/vttrifonov/repos", "site_admin": false, "starred_url": "https://api.github.com/users/vttrifonov/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/vttrifonov/subscriptions", "type": "User", "url": "https://api.github.com/users/vttrifonov", "user_view_type": "public" }
[]
closed
false
null
[]
null
[]
"2024-12-25T07:36:48"
"2025-01-07T11:56:42"
2025-01-07T11:56:42
CONTRIBUTOR
{ "completed": 0, "percent_completed": 0, "total": 0 }
null
### Describe the bug The following code ```python import datasets as hf ds1 = hf.Dataset.from_list([{'i': i} for i in [0,1]]) #ds1 = ds1.to_iterable_dataset() ds2 = ds1.map( lambda i: {'i': i+1}, input_columns = ['i'], remove_columns = ['i'] ) list(ds2) ``` produces ```python [{'i': 1}, {'i': 2}] ``` as expected. If the line that converts `ds1` to iterable is uncommented so that the `ds2` is a map of an `IterableDataset`, the result is ```python [{},{}] ``` I expected the output to be the same as before. It seems that in the second case the removed column is not added back into the output. The issue seems to be [here](https://github.com/huggingface/datasets/blob/6c6a82a573f946c4a81069f56446caed15cee9c2/src/datasets/iterable_dataset.py#L1093): the columns are removed after the mapping which is not what we want (or what the [documentation says](https://github.com/huggingface/datasets/blob/6c6a82a573f946c4a81069f56446caed15cee9c2/src/datasets/iterable_dataset.py#L2370)) because we want the columns removed from the transformed example but then added if the map produced them. This is `datasets==3.2.0` and `python==3.10` ### Steps to reproduce the bug see above ### Expected behavior see above ### Environment info see above
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7345/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7345/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7344
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7344/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7344/comments
https://api.github.com/repos/huggingface/datasets/issues/7344/events
https://github.com/huggingface/datasets/issues/7344
2,754,735,951
I_kwDODunzps6kMe9P
7,344
HfHubHTTPError: 429 Client Error: Too Many Requests for URL when trying to access SlimPajama-627B or c4 on TPUs
{ "avatar_url": "https://avatars.githubusercontent.com/u/9397233?v=4", "events_url": "https://api.github.com/users/clankur/events{/privacy}", "followers_url": "https://api.github.com/users/clankur/followers", "following_url": "https://api.github.com/users/clankur/following{/other_user}", "gists_url": "https://api.github.com/users/clankur/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/clankur", "id": 9397233, "login": "clankur", "node_id": "MDQ6VXNlcjkzOTcyMzM=", "organizations_url": "https://api.github.com/users/clankur/orgs", "received_events_url": "https://api.github.com/users/clankur/received_events", "repos_url": "https://api.github.com/users/clankur/repos", "site_admin": false, "starred_url": "https://api.github.com/users/clankur/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/clankur/subscriptions", "type": "User", "url": "https://api.github.com/users/clankur", "user_view_type": "public" }
[]
closed
false
null
[]
null
[]
"2024-12-22T16:30:07"
"2025-01-15T05:32:00"
2025-01-15T05:31:58
NONE
{ "completed": 0, "percent_completed": 0, "total": 0 }
null
### Describe the bug I am trying to run some trainings on Google's TPUs using Huggingface's DataLoader on [SlimPajama-627B](https://huggingface.co./datasets/cerebras/SlimPajama-627B) and [c4](https://huggingface.co./datasets/allenai/c4), but I end up running into `429 Client Error: Too Many Requests for URL` error when I call `load_dataset`. The even odder part is that I am able to sucessfully run trainings with the [wikitext dataset](https://huggingface.co./datasets/Salesforce/wikitext). Is there something I need to setup to specifically train with SlimPajama or C4 with TPUs because I am not clear why I am getting these errors. ### Steps to reproduce the bug These are the commands you could run to produce the error below but you will require a ClearML account (you can create one [here](https://app.clear.ml/login?redirect=%2Fdashboard)) with a queue setup to run on Google TPUs ```bash git clone https://github.com/clankur/muGPT.git cd muGPT python -m train --config-name=slim_v4-32_84m.yaml +training.queue={NAME_OF_CLEARML_QUEUE} ``` The error I see: ``` Traceback (most recent call last): File "/home/clankur/conda/envs/jax/lib/python3.10/site-packages/clearml/binding/hydra_bind.py", line 230, in _patched_task_function return task_function(a_config, *a_args, **a_kwargs) File "/home/clankur/.clearml/venvs-builds/3.10/task_repository/muGPT.git/train.py", line 1037, in main main_contained(config, logger) File "/home/clankur/.clearml/venvs-builds/3.10/task_repository/muGPT.git/train.py", line 840, in main_contained loader = get_loader("train", config.training_data, config.training.tokens) File "/home/clankur/.clearml/venvs-builds/3.10/task_repository/muGPT.git/input_loader.py", line 549, in get_loader return HuggingFaceDataLoader(split, config, token_batch_params) File "/home/clankur/.clearml/venvs-builds/3.10/task_repository/muGPT.git/input_loader.py", line 395, in __init__ self.dataset = load_dataset( File "/home/clankur/conda/envs/jax/lib/python3.10/site-packages/datasets/load.py", line 2112, in load_dataset builder_instance = load_dataset_builder( File "/home/clankur/conda/envs/jax/lib/python3.10/site-packages/datasets/load.py", line 1798, in load_dataset_builder dataset_module = dataset_module_factory( File "/home/clankur/conda/envs/jax/lib/python3.10/site-packages/datasets/load.py", line 1495, in dataset_module_factory raise e1 from None File "/home/clankur/conda/envs/jax/lib/python3.10/site-packages/datasets/load.py", line 1479, in dataset_module_factory ).get_module() File "/home/clankur/conda/envs/jax/lib/python3.10/site-packages/datasets/load.py", line 1034, in get_module else get_data_patterns(base_path, download_config=self.download_config) File "/home/clankur/conda/envs/jax/lib/python3.10/site-packages/datasets/data_files.py", line 457, in get_data_patterns return _get_data_files_patterns(resolver) File "/home/clankur/conda/envs/jax/lib/python3.10/site-packages/datasets/data_files.py", line 248, in _get_data_files_patterns data_files = pattern_resolver(pattern) File "/home/clankur/conda/envs/jax/lib/python3.10/site-packages/datasets/data_files.py", line 340, in resolve_pattern for filepath, info in fs.glob(pattern, detail=True).items() File "/home/clankur/conda/envs/jax/lib/python3.10/site-packages/huggingface_hub/hf_file_system.py", line 409, in glob return super().glob(path, **kwargs) File "/home/clankur/.clearml/venvs-builds/3.10/lib/python3.10/site-packages/fsspec/spec.py", line 602, in glob allpaths = self.find(root, maxdepth=depth, withdirs=True, detail=True, **kwargs) File "/home/clankur/conda/envs/jax/lib/python3.10/site-packages/huggingface_hub/hf_file_system.py", line 429, in find out = self._ls_tree(path, recursive=True, refresh=refresh, revision=resolved_path.revision, **kwargs) File "/home/clankur/conda/envs/jax/lib/python3.10/site-packages/huggingface_hub/hf_file_system.py", line 358, in _ls_tree self._ls_tree( File "/home/clankur/conda/envs/jax/lib/python3.10/site-packages/huggingface_hub/hf_file_system.py", line 375, in _ls_tree for path_info in tree: File "/home/clankur/conda/envs/jax/lib/python3.10/site-packages/huggingface_hub/hf_api.py", line 3080, in list_repo_tree for path_info in paginate(path=tree_url, headers=headers, params={"recursive": recursive, "expand": expand}): File "/home/clankur/conda/envs/jax/lib/python3.10/site-packages/huggingface_hub/utils/_pagination.py", line 46, in paginate hf_raise_for_status(r) File "/home/clankur/conda/envs/jax/lib/python3.10/site-packages/huggingface_hub/utils/_http.py", line 477, in hf_raise_for_status raise _format(HfHubHTTPError, str(e), response) from e huggingface_hub.errors.HfHubHTTPError: 429 Client Error: Too Many Requests for url: https://huggingface.co./api/datasets/cerebras/SlimPajama-627B/tree/2d0accdd58c5d5511943ca1f5ff0e3eb5e293543?recursive=True&expand=True&cursor=ZXlKbWFXeGxYMjVoYldVaU9pSjBaWE4wTDJOb2RXNXJNUzlsZUdGdGNHeGxYMmh2YkdSdmRYUmZPVFEzTG1wemIyNXNMbnB6ZENKOTo2MjUw (Request ID: Root=1-67673de9-1413900606ede7712b08ef2c;1304c09c-3e69-4222-be14-f10ee709d49c) maximum queue size reached Set the environment variable HYDRA_FULL_ERROR=1 for a complete stack trace. ``` ### Expected behavior I'd expect the DataLoader to load from the SlimPajama-627B and c4 dataset without issue. ### Environment info - `datasets` version: 2.14.4 - Platform: Linux-5.8.0-1035-gcp-x86_64-with-glibc2.31 - Python version: 3.10.16 - Huggingface_hub version: 0.26.5 - PyArrow version: 18.1.0 - Pandas version: 2.2.3
{ "avatar_url": "https://avatars.githubusercontent.com/u/9397233?v=4", "events_url": "https://api.github.com/users/clankur/events{/privacy}", "followers_url": "https://api.github.com/users/clankur/followers", "following_url": "https://api.github.com/users/clankur/following{/other_user}", "gists_url": "https://api.github.com/users/clankur/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/clankur", "id": 9397233, "login": "clankur", "node_id": "MDQ6VXNlcjkzOTcyMzM=", "organizations_url": "https://api.github.com/users/clankur/orgs", "received_events_url": "https://api.github.com/users/clankur/received_events", "repos_url": "https://api.github.com/users/clankur/repos", "site_admin": false, "starred_url": "https://api.github.com/users/clankur/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/clankur/subscriptions", "type": "User", "url": "https://api.github.com/users/clankur", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7344/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7344/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7343
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7343/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7343/comments
https://api.github.com/repos/huggingface/datasets/issues/7343/events
https://github.com/huggingface/datasets/issues/7343
2,750,525,823
I_kwDODunzps6j8bF_
7,343
[Bug] Inconsistent behavior of data_files and data_dir in load_dataset method.
{ "avatar_url": "https://avatars.githubusercontent.com/u/74161960?v=4", "events_url": "https://api.github.com/users/JasonCZH4/events{/privacy}", "followers_url": "https://api.github.com/users/JasonCZH4/followers", "following_url": "https://api.github.com/users/JasonCZH4/following{/other_user}", "gists_url": "https://api.github.com/users/JasonCZH4/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/JasonCZH4", "id": 74161960, "login": "JasonCZH4", "node_id": "MDQ6VXNlcjc0MTYxOTYw", "organizations_url": "https://api.github.com/users/JasonCZH4/orgs", "received_events_url": "https://api.github.com/users/JasonCZH4/received_events", "repos_url": "https://api.github.com/users/JasonCZH4/repos", "site_admin": false, "starred_url": "https://api.github.com/users/JasonCZH4/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/JasonCZH4/subscriptions", "type": "User", "url": "https://api.github.com/users/JasonCZH4", "user_view_type": "public" }
[]
closed
false
null
[]
null
[]
"2024-12-19T14:31:27"
"2025-01-03T15:54:09"
2025-01-03T15:54:09
NONE
{ "completed": 0, "percent_completed": 0, "total": 0 }
null
### Describe the bug Inconsistent operation of data_files and data_dir in load_dataset method. ### Steps to reproduce the bug # First I have three files, named 'train.json', 'val.json', 'test.json'. Each one has a simple dict `{text:'aaa'}`. Their path are `/data/train.json`, `/data/val.json`, `/data/test.json` I load dataset with `data_files` argument: ```py files = [os.path.join('./data',file) for file in os.listdir('./data')] ds = load_dataset( path='json', data_files=files,) ``` And I get: ```py DatasetDict({ train: Dataset({ features: ['text'], num_rows: 3 }) }) ``` However, If I load dataset with `data_dir` argument: ```py ds = load_dataset( path='json', data_dir='./data',) ``` And I get: ```py DatasetDict({ train: Dataset({ features: ['text'], num_rows: 1 }) validation: Dataset({ features: ['text'], num_rows: 1 }) test: Dataset({ features: ['text'], num_rows: 1 }) }) ``` Two results are not the same. Their behaviors are not equal, even if the statement [here](https://github.com/huggingface/datasets/blob/d0c152a979d91cc34b605c0298aebc650ab7dd27/src/datasets/load.py#L1790) said that their behaviors are equal. # Second If some filename include 'test' while others do not, `load_dataset` only return `test` dataset and others files are **abandoned**. Given two files named `test.json` and `1.json` Each one has a simple dict `{text:'aaa'}`. I load the dataset using: ```py ds = load_dataset( path='json', data_dir='./data',) ``` Only `test` is returned, `1.json` is missing: ```py DatasetDict({ test: Dataset({ features: ['text'], num_rows: 1 }) }) ``` Things do not change even I manually set `split='train'` ### Expected behavior 1. Fix the above bugs. 2. Although the document says that load_dataset method will `Find which file goes into which split (e.g. train/test) based on file and directory names or on the YAML configuration`, I hope I can manually decide whether to do so. Sometimes users may accidentally put a `test` string in the filename but they just want a single `train` dataset. If the number of files in `data_dir` is huge, it's not easy to find out what cause the second situation metioned above. ### Environment info datasets==3.2.0 Ubuntu18.84
{ "avatar_url": "https://avatars.githubusercontent.com/u/74161960?v=4", "events_url": "https://api.github.com/users/JasonCZH4/events{/privacy}", "followers_url": "https://api.github.com/users/JasonCZH4/followers", "following_url": "https://api.github.com/users/JasonCZH4/following{/other_user}", "gists_url": "https://api.github.com/users/JasonCZH4/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/JasonCZH4", "id": 74161960, "login": "JasonCZH4", "node_id": "MDQ6VXNlcjc0MTYxOTYw", "organizations_url": "https://api.github.com/users/JasonCZH4/orgs", "received_events_url": "https://api.github.com/users/JasonCZH4/received_events", "repos_url": "https://api.github.com/users/JasonCZH4/repos", "site_admin": false, "starred_url": "https://api.github.com/users/JasonCZH4/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/JasonCZH4/subscriptions", "type": "User", "url": "https://api.github.com/users/JasonCZH4", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7343/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7343/timeline
null
completed
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7342
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7342/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7342/comments
https://api.github.com/repos/huggingface/datasets/issues/7342/events
https://github.com/huggingface/datasets/pull/7342
2,749,572,310
PR_kwDODunzps6FvgcK
7,342
Update LICENSE
{ "avatar_url": "https://avatars.githubusercontent.com/u/97572401?v=4", "events_url": "https://api.github.com/users/eliebak/events{/privacy}", "followers_url": "https://api.github.com/users/eliebak/followers", "following_url": "https://api.github.com/users/eliebak/following{/other_user}", "gists_url": "https://api.github.com/users/eliebak/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/eliebak", "id": 97572401, "login": "eliebak", "node_id": "U_kgDOBdDWMQ", "organizations_url": "https://api.github.com/users/eliebak/orgs", "received_events_url": "https://api.github.com/users/eliebak/received_events", "repos_url": "https://api.github.com/users/eliebak/repos", "site_admin": false, "starred_url": "https://api.github.com/users/eliebak/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/eliebak/subscriptions", "type": "User", "url": "https://api.github.com/users/eliebak", "user_view_type": "public" }
[]
closed
false
null
[]
null
[]
"2024-12-19T08:17:50"
"2024-12-19T08:44:08"
2024-12-19T08:44:08
NONE
{ "completed": 0, "percent_completed": 0, "total": 0 }
null
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/97572401?v=4", "events_url": "https://api.github.com/users/eliebak/events{/privacy}", "followers_url": "https://api.github.com/users/eliebak/followers", "following_url": "https://api.github.com/users/eliebak/following{/other_user}", "gists_url": "https://api.github.com/users/eliebak/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/eliebak", "id": 97572401, "login": "eliebak", "node_id": "U_kgDOBdDWMQ", "organizations_url": "https://api.github.com/users/eliebak/orgs", "received_events_url": "https://api.github.com/users/eliebak/received_events", "repos_url": "https://api.github.com/users/eliebak/repos", "site_admin": false, "starred_url": "https://api.github.com/users/eliebak/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/eliebak/subscriptions", "type": "User", "url": "https://api.github.com/users/eliebak", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7342/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7342/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7342.diff", "html_url": "https://github.com/huggingface/datasets/pull/7342", "merged_at": null, "patch_url": "https://github.com/huggingface/datasets/pull/7342.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7342" }
true
https://api.github.com/repos/huggingface/datasets/issues/7341
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7341/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7341/comments
https://api.github.com/repos/huggingface/datasets/issues/7341/events
https://github.com/huggingface/datasets/pull/7341
2,745,658,561
PR_kwDODunzps6FiGlt
7,341
minor video docs on how to install
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
closed
false
null
[]
null
[]
"2024-12-17T18:06:17"
"2024-12-17T18:11:17"
2024-12-17T18:11:15
MEMBER
{ "completed": 0, "percent_completed": 0, "total": 0 }
null
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7341/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7341/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7341.diff", "html_url": "https://github.com/huggingface/datasets/pull/7341", "merged_at": "2024-12-17T18:11:14Z", "patch_url": "https://github.com/huggingface/datasets/pull/7341.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7341" }
true
https://api.github.com/repos/huggingface/datasets/issues/7340
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7340/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7340/comments
https://api.github.com/repos/huggingface/datasets/issues/7340/events
https://github.com/huggingface/datasets/pull/7340
2,745,473,274
PR_kwDODunzps6FhdR2
7,340
don't import soundfile in tests
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
closed
false
null
[]
null
[]
"2024-12-17T16:49:55"
"2024-12-17T16:54:04"
2024-12-17T16:50:24
MEMBER
{ "completed": 0, "percent_completed": 0, "total": 0 }
null
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7340/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7340/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7340.diff", "html_url": "https://github.com/huggingface/datasets/pull/7340", "merged_at": "2024-12-17T16:50:24Z", "patch_url": "https://github.com/huggingface/datasets/pull/7340.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7340" }
true
https://api.github.com/repos/huggingface/datasets/issues/7339
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7339/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7339/comments
https://api.github.com/repos/huggingface/datasets/issues/7339/events
https://github.com/huggingface/datasets/pull/7339
2,745,460,060
PR_kwDODunzps6FhaTl
7,339
Update CONTRIBUTING.md
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
[]
closed
false
null
[]
null
[]
"2024-12-17T16:45:25"
"2024-12-17T16:51:36"
2024-12-17T16:46:30
MEMBER
{ "completed": 0, "percent_completed": 0, "total": 0 }
null
null
{ "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "events_url": "https://api.github.com/users/lhoestq/events{/privacy}", "followers_url": "https://api.github.com/users/lhoestq/followers", "following_url": "https://api.github.com/users/lhoestq/following{/other_user}", "gists_url": "https://api.github.com/users/lhoestq/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/lhoestq", "id": 42851186, "login": "lhoestq", "node_id": "MDQ6VXNlcjQyODUxMTg2", "organizations_url": "https://api.github.com/users/lhoestq/orgs", "received_events_url": "https://api.github.com/users/lhoestq/received_events", "repos_url": "https://api.github.com/users/lhoestq/repos", "site_admin": false, "starred_url": "https://api.github.com/users/lhoestq/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/lhoestq/subscriptions", "type": "User", "url": "https://api.github.com/users/lhoestq", "user_view_type": "public" }
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7339/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7339/timeline
null
null
0
{ "diff_url": "https://github.com/huggingface/datasets/pull/7339.diff", "html_url": "https://github.com/huggingface/datasets/pull/7339", "merged_at": "2024-12-17T16:46:30Z", "patch_url": "https://github.com/huggingface/datasets/pull/7339.patch", "url": "https://api.github.com/repos/huggingface/datasets/pulls/7339" }
true
https://api.github.com/repos/huggingface/datasets/issues/7337
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7337/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7337/comments
https://api.github.com/repos/huggingface/datasets/issues/7337/events
https://github.com/huggingface/datasets/issues/7337
2,744,877,569
I_kwDODunzps6jm4IB
7,337
One or several metadata.jsonl were found, but not in the same directory or in a parent directory of
{ "avatar_url": "https://avatars.githubusercontent.com/u/67250532?v=4", "events_url": "https://api.github.com/users/mst272/events{/privacy}", "followers_url": "https://api.github.com/users/mst272/followers", "following_url": "https://api.github.com/users/mst272/following{/other_user}", "gists_url": "https://api.github.com/users/mst272/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/mst272", "id": 67250532, "login": "mst272", "node_id": "MDQ6VXNlcjY3MjUwNTMy", "organizations_url": "https://api.github.com/users/mst272/orgs", "received_events_url": "https://api.github.com/users/mst272/received_events", "repos_url": "https://api.github.com/users/mst272/repos", "site_admin": false, "starred_url": "https://api.github.com/users/mst272/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/mst272/subscriptions", "type": "User", "url": "https://api.github.com/users/mst272", "user_view_type": "public" }
[]
open
false
null
[]
null
[]
"2024-12-17T12:58:43"
"2025-01-03T15:28:13"
1970-01-01T00:00:00
NONE
{ "completed": 0, "percent_completed": 0, "total": 0 }
null
### Describe the bug ImageFolder with metadata.jsonl error. I downloaded liuhaotian/LLaVA-CC3M-Pretrain-595K locally from Hugging Face. According to the tutorial in https://huggingface.co./docs/datasets/image_dataset#image-captioning, only put images.zip and metadata.jsonl containing information in the same folder. However, after loading, an error was reported: One or several metadata.jsonl were found, but not in the same directory or in a parent directory of. The data in my jsonl file is as follows: > {"id": "GCC_train_002448550", "file_name": "GCC_train_002448550.jpg", "conversations": [{"from": "human", "value": "<image>\nProvide a brief description of the given image."}, {"from": "gpt", "value": "a view of a city , where the flyover was proposed to reduce the increasing traffic on thursday ."}]} ### Steps to reproduce the bug from datasets import load_dataset image = load_dataset("imagefolder",data_dir='data/opensource_data') ### Expected behavior success ### Environment info datasets==3.2.0
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7337/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7337/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7336
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7336/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7336/comments
https://api.github.com/repos/huggingface/datasets/issues/7336/events
https://github.com/huggingface/datasets/issues/7336
2,744,746,456
I_kwDODunzps6jmYHY
7,336
Clarify documentation or Create DatasetCard
{ "avatar_url": "https://avatars.githubusercontent.com/u/145011209?v=4", "events_url": "https://api.github.com/users/August-murr/events{/privacy}", "followers_url": "https://api.github.com/users/August-murr/followers", "following_url": "https://api.github.com/users/August-murr/following{/other_user}", "gists_url": "https://api.github.com/users/August-murr/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/August-murr", "id": 145011209, "login": "August-murr", "node_id": "U_kgDOCKSyCQ", "organizations_url": "https://api.github.com/users/August-murr/orgs", "received_events_url": "https://api.github.com/users/August-murr/received_events", "repos_url": "https://api.github.com/users/August-murr/repos", "site_admin": false, "starred_url": "https://api.github.com/users/August-murr/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/August-murr/subscriptions", "type": "User", "url": "https://api.github.com/users/August-murr", "user_view_type": "public" }
[ { "color": "a2eeef", "default": true, "description": "New feature or request", "id": 1935892871, "name": "enhancement", "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement" } ]
open
false
null
[]
null
[]
"2024-12-17T12:01:00"
"2024-12-17T12:01:00"
1970-01-01T00:00:00
NONE
{ "completed": 0, "percent_completed": 0, "total": 0 }
null
### Feature request I noticed that you can use a Model Card instead of a Dataset Card when pushing a dataset to the Hub, but this isn’t clearly mentioned in [the docs.](https://huggingface.co./docs/datasets/dataset_card) - Update the docs to clarify that a Model Card can work for datasets too. - It might be worth creating a dedicated DatasetCard module, similar to the ModelCard module, for consistency and better support. Not sure if this belongs here or on the [Hub repo](https://github.com/huggingface/huggingface_hub), but thought I’d bring it up! ### Motivation I just spent an hour like on [this issue](https://github.com/huggingface/trl/pull/2491) trying to create a `DatasetCard` for a script. ### Your contribution might later
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7336/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7336/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7335
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7335/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7335/comments
https://api.github.com/repos/huggingface/datasets/issues/7335/events
https://github.com/huggingface/datasets/issues/7335
2,743,437,260
I_kwDODunzps6jhYfM
7,335
Too many open files: '/root/.cache/huggingface/token'
{ "avatar_url": "https://avatars.githubusercontent.com/u/17604849?v=4", "events_url": "https://api.github.com/users/kopyl/events{/privacy}", "followers_url": "https://api.github.com/users/kopyl/followers", "following_url": "https://api.github.com/users/kopyl/following{/other_user}", "gists_url": "https://api.github.com/users/kopyl/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/kopyl", "id": 17604849, "login": "kopyl", "node_id": "MDQ6VXNlcjE3NjA0ODQ5", "organizations_url": "https://api.github.com/users/kopyl/orgs", "received_events_url": "https://api.github.com/users/kopyl/received_events", "repos_url": "https://api.github.com/users/kopyl/repos", "site_admin": false, "starred_url": "https://api.github.com/users/kopyl/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/kopyl/subscriptions", "type": "User", "url": "https://api.github.com/users/kopyl", "user_view_type": "public" }
[]
open
false
null
[]
null
[]
"2024-12-16T21:30:24"
"2024-12-16T21:30:24"
1970-01-01T00:00:00
NONE
{ "completed": 0, "percent_completed": 0, "total": 0 }
null
### Describe the bug I ran this code: ``` from datasets import load_dataset dataset = load_dataset("common-canvas/commoncatalog-cc-by", cache_dir="/datadrive/datasets/cc", num_proc=1000) ``` And got this error. Before it was some other file though (lie something...incomplete) runnting ``` ulimit -n 8192 ``` did not help at all. ### Steps to reproduce the bug Run the code i sent ### Expected behavior Should be no errors ### Environment info linux, jupyter lab.
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7335/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7335/timeline
null
null
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/7334
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/7334/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/7334/comments
https://api.github.com/repos/huggingface/datasets/issues/7334/events
https://github.com/huggingface/datasets/issues/7334
2,740,266,503
I_kwDODunzps6jVSYH
7,334
TypeError: Value.__init__() missing 1 required positional argument: 'dtype'
{ "avatar_url": "https://avatars.githubusercontent.com/u/185799756?v=4", "events_url": "https://api.github.com/users/kakamond/events{/privacy}", "followers_url": "https://api.github.com/users/kakamond/followers", "following_url": "https://api.github.com/users/kakamond/following{/other_user}", "gists_url": "https://api.github.com/users/kakamond/gists{/gist_id}", "gravatar_id": "", "html_url": "https://github.com/kakamond", "id": 185799756, "login": "kakamond", "node_id": "U_kgDOCxMUTA", "organizations_url": "https://api.github.com/users/kakamond/orgs", "received_events_url": "https://api.github.com/users/kakamond/received_events", "repos_url": "https://api.github.com/users/kakamond/repos", "site_admin": false, "starred_url": "https://api.github.com/users/kakamond/starred{/owner}{/repo}", "subscriptions_url": "https://api.github.com/users/kakamond/subscriptions", "type": "User", "url": "https://api.github.com/users/kakamond", "user_view_type": "public" }
[]
open
false
null
[]
null
[]
"2024-12-15T04:08:46"
"2024-12-15T04:08:46"
1970-01-01T00:00:00
NONE
{ "completed": 0, "percent_completed": 0, "total": 0 }
null
### Describe the bug ds = load_dataset( "./xxx.py", name="default", split="train", ) The datasets does not support debugging locally anymore... ### Steps to reproduce the bug ``` from datasets import load_dataset ds = load_dataset( "./repo.py", name="default", split="train", ) for item in ds: print(item) ``` It works fine for "username/repo", but it does not work for "./repo.py" when debugging locally... Running above code template will report TypeError: Value.__init__() missing 1 required positional argument: 'dtype' ### Expected behavior fix this bug ### Environment info python 3.10 datasets==2.21
null
{ "+1": 0, "-1": 0, "confused": 0, "eyes": 0, "heart": 0, "hooray": 0, "laugh": 0, "rocket": 0, "total_count": 0, "url": "https://api.github.com/repos/huggingface/datasets/issues/7334/reactions" }
https://api.github.com/repos/huggingface/datasets/issues/7334/timeline
null
null
null
null
false

Dataset Card for GitHub Issues

Dataset Summary

GitHub Issues is a dataset consisting of GitHub issues and pull requests associated with the 🤗 Datasets repository. It is intended for educational purposes and can be used for semantic search or multilabel text classification. The contents of each GitHub issue are in English and concern the domain of datasets for NLP, computer vision, and beyond.

Supported Tasks and Leaderboards

For each of the tasks tagged for this dataset, give a brief description of the tag, metrics, and suggested models (with a link to their HuggingFace implementation if available). Give a similar description of tasks that were not covered by the structured tag set (repace the task-category-tag with an appropriate other:other-task-name).

  • task-category-tag: The dataset can be used to train a model for [TASK NAME], which consists in [TASK DESCRIPTION]. Success on this task is typically measured by achieving a high/low metric name. The (model name or model class) model currently achieves the following score. [IF A LEADERBOARD IS AVAILABLE]: This task has an active leaderboard which can be found at leaderboard url and ranks models based on metric name while also reporting other metric name.

Languages

Provide a brief overview of the languages represented in the dataset. Describe relevant details about specifics of the language such as whether it is social media text, African American English,...

When relevant, please provide BCP-47 codes, which consist of a primary language subtag, with a script subtag and/or region subtag if available.

Dataset Structure

Data Instances

Provide an JSON-formatted example and brief description of a typical instance in the dataset. If available, provide a link to further examples.

{
  'example_field': ...,
  ...
}

Provide any additional information that is not covered in the other sections about the data here. In particular describe any relationships between data points and if these relationships are made explicit.

Data Fields

List and describe the fields present in the dataset. Mention their data type, and whether they are used as input or output in any of the tasks the dataset currently supports. If the data has span indices, describe their attributes, such as whether they are at the character level or word level, whether they are contiguous or not, etc. If the datasets contains example IDs, state whether they have an inherent meaning, such as a mapping to other datasets or pointing to relationships between data points.

  • example_field: description of example_field

Note that the descriptions can be initialized with the Show Markdown Data Fields output of the tagging app, you will then only need to refine the generated descriptions.

Data Splits

Describe and name the splits in the dataset if there are more than one.

Describe any criteria for splitting the data, if used. If their are differences between the splits (e.g. if the training annotations are machine-generated and the dev and test ones are created by humans, or if different numbers of annotators contributed to each example), describe them here.

Provide the sizes of each split. As appropriate, provide any descriptive statistics for the features, such as average length. For example:

Tain Valid Test
Input Sentences
Average Sentence Length

Dataset Creation

Curation Rationale

What need motivated the creation of this dataset? What are some of the reasons underlying the major choices involved in putting it together?

Source Data

This section describes the source data (e.g. news text and headlines, social media posts, translated sentences,...)

Initial Data Collection and Normalization

Describe the data collection process. Describe any criteria for data selection or filtering. List any key words or search terms used. If possible, include runtime information for the collection process.

If data was collected from other pre-existing datasets, link to source here and to their Hugging Face version.

If the data was modified or normalized after being collected (e.g. if the data is word-tokenized), describe the process and the tools used.

Who are the source language producers?

State whether the data was produced by humans or machine generated. Describe the people or systems who originally created the data.

If available, include self-reported demographic or identity information for the source data creators, but avoid inferring this information. Instead state that this information is unknown. See Larson 2017 for using identity categories as a variables, particularly gender.

Describe the conditions under which the data was created (for example, if the producers were crowdworkers, state what platform was used, or if the data was found, what website the data was found on). If compensation was provided, include that information here.

Describe other people represented or mentioned in the data. Where possible, link to references for the information.

Annotations

If the dataset contains annotations which are not part of the initial data collection, describe them in the following paragraphs.

Annotation process

If applicable, describe the annotation process and any tools used, or state otherwise. Describe the amount of data annotated, if not all. Describe or reference annotation guidelines provided to the annotators. If available, provide interannotator statistics. Describe any annotation validation processes.

Who are the annotators?

If annotations were collected for the source data (such as class labels or syntactic parses), state whether the annotations were produced by humans or machine generated.

Describe the people or systems who originally created the annotations and their selection criteria if applicable.

If available, include self-reported demographic or identity information for the annotators, but avoid inferring this information. Instead state that this information is unknown. See Larson 2017 for using identity categories as a variables, particularly gender.

Describe the conditions under which the data was annotated (for example, if the annotators were crowdworkers, state what platform was used, or if the data was found, what website the data was found on). If compensation was provided, include that information here.

Personal and Sensitive Information

State whether the dataset uses identity categories and, if so, how the information is used. Describe where this information comes from (i.e. self-reporting, collecting from profiles, inferring, etc.). See Larson 2017 for using identity categories as a variables, particularly gender. State whether the data is linked to individuals and whether those individuals can be identified in the dataset, either directly or indirectly (i.e., in combination with other data).

State whether the dataset contains other data that might be considered sensitive (e.g., data that reveals racial or ethnic origins, sexual orientations, religious beliefs, political opinions or union memberships, or locations; financial or health data; biometric or genetic data; forms of government identification, such as social security numbers; criminal history).

If efforts were made to anonymize the data, describe the anonymization process.

Considerations for Using the Data

Social Impact of Dataset

Please discuss some of the ways you believe the use of this dataset will impact society.

The statement should include both positive outlooks, such as outlining how technologies developed through its use may improve people's lives, and discuss the accompanying risks. These risks may range from making important decisions more opaque to people who are affected by the technology, to reinforcing existing harmful biases (whose specifics should be discussed in the next section), among other considerations.

Also describe in this section if the proposed dataset contains a low-resource or under-represented language. If this is the case or if this task has any impact on underserved communities, please elaborate here.

Discussion of Biases

Provide descriptions of specific biases that are likely to be reflected in the data, and state whether any steps were taken to reduce their impact.

For Wikipedia text, see for example Dinan et al 2020 on biases in Wikipedia (esp. Table 1), or Blodgett et al 2020 for a more general discussion of the topic.

If analyses have been run quantifying these biases, please add brief summaries and links to the studies here.

Other Known Limitations

If studies of the datasets have outlined other limitations of the dataset, such as annotation artifacts, please outline and cite them here.

Additional Information

Dataset Curators

List the people involved in collecting the dataset and their affiliation(s). If funding information is known, include it here.

Licensing Information

Provide the license and link to the license webpage if available.

Citation Information

Provide the BibTex-formatted reference for the dataset. For example:

@article{article_id,
  author    = {Author List},
  title     = {Dataset Paper Title},
  journal   = {Publication Venue},
  year      = {2525}
}

If the dataset has a DOI, please provide it here.

Downloads last month
53