parquet-converter commited on
Commit
793c8f9
·
1 Parent(s): 77e3a18

Update parquet files

Browse files
.gitattributes DELETED
@@ -1,52 +0,0 @@
1
- *.7z filter=lfs diff=lfs merge=lfs -text
2
- *.arrow filter=lfs diff=lfs merge=lfs -text
3
- *.bin filter=lfs diff=lfs merge=lfs -text
4
- *.bz2 filter=lfs diff=lfs merge=lfs -text
5
- *.ftz filter=lfs diff=lfs merge=lfs -text
6
- *.gz filter=lfs diff=lfs merge=lfs -text
7
- *.h5 filter=lfs diff=lfs merge=lfs -text
8
- *.joblib filter=lfs diff=lfs merge=lfs -text
9
- *.lfs.* filter=lfs diff=lfs merge=lfs -text
10
- *.lz4 filter=lfs diff=lfs merge=lfs -text
11
- *.model filter=lfs diff=lfs merge=lfs -text
12
- *.msgpack filter=lfs diff=lfs merge=lfs -text
13
- *.npy filter=lfs diff=lfs merge=lfs -text
14
- *.npz filter=lfs diff=lfs merge=lfs -text
15
- *.onnx filter=lfs diff=lfs merge=lfs -text
16
- *.ot filter=lfs diff=lfs merge=lfs -text
17
- *.parquet filter=lfs diff=lfs merge=lfs -text
18
- *.pb filter=lfs diff=lfs merge=lfs -text
19
- *.pickle filter=lfs diff=lfs merge=lfs -text
20
- *.pkl filter=lfs diff=lfs merge=lfs -text
21
- *.pt filter=lfs diff=lfs merge=lfs -text
22
- *.pth filter=lfs diff=lfs merge=lfs -text
23
- *.rar filter=lfs diff=lfs merge=lfs -text
24
- saved_model/**/* filter=lfs diff=lfs merge=lfs -text
25
- *.tar.* filter=lfs diff=lfs merge=lfs -text
26
- *.tflite filter=lfs diff=lfs merge=lfs -text
27
- *.tgz filter=lfs diff=lfs merge=lfs -text
28
- *.wasm filter=lfs diff=lfs merge=lfs -text
29
- *.xz filter=lfs diff=lfs merge=lfs -text
30
- *.zip filter=lfs diff=lfs merge=lfs -text
31
- *.zst filter=lfs diff=lfs merge=lfs -text
32
- *tfevents* filter=lfs diff=lfs merge=lfs -text
33
- # Audio files - uncompressed
34
- *.pcm filter=lfs diff=lfs merge=lfs -text
35
- *.sam filter=lfs diff=lfs merge=lfs -text
36
- *.raw filter=lfs diff=lfs merge=lfs -text
37
- # Audio files - compressed
38
- *.aac filter=lfs diff=lfs merge=lfs -text
39
- *.flac filter=lfs diff=lfs merge=lfs -text
40
- *.mp3 filter=lfs diff=lfs merge=lfs -text
41
- *.ogg filter=lfs diff=lfs merge=lfs -text
42
- *.wav filter=lfs diff=lfs merge=lfs -text
43
- # Image files - uncompressed
44
- *.bmp filter=lfs diff=lfs merge=lfs -text
45
- *.gif filter=lfs diff=lfs merge=lfs -text
46
- *.png filter=lfs diff=lfs merge=lfs -text
47
- *.tiff filter=lfs diff=lfs merge=lfs -text
48
- # Image files - compressed
49
- *.jpg filter=lfs diff=lfs merge=lfs -text
50
- *.jpeg filter=lfs diff=lfs merge=lfs -text
51
- *.webp filter=lfs diff=lfs merge=lfs -text
52
- dataset_coco.json filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
dataset_coco.json → Luka-Wang--COCO/json-train.parquet RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2fd999220673258012acfb411a4e7e66af7d488050b2519b0badcc49b7600b8d
3
- size 144186139
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf45ce685ee48c8e1f95b64fa7f7374eca88c8f34f0b7e7be6250d86edf6166b
3
+ size 39502104
README.md DELETED
@@ -1,154 +0,0 @@
1
- ---
2
- annotations_creators:
3
- - expert-generated
4
- language:
5
- - en
6
- language_creators:
7
- - found
8
- license:
9
- - mit
10
- multilinguality:
11
- - monolingual
12
- paperswithcode_id: acronym-identification
13
- pretty_name: Acronym Identification Dataset
14
- size_categories:
15
- - 10K<n<100K
16
- source_datasets:
17
- - original
18
- task_categories:
19
- - token-classification
20
- task_ids:
21
- - token-classification-other-acronym-identification
22
- train-eval-index:
23
- - col_mapping:
24
- labels: tags
25
- tokens: tokens
26
- config: default
27
- splits:
28
- eval_split: test
29
- task: token-classification
30
- task_id: entity_extraction
31
- ---
32
-
33
- # Dataset Card for [COCO]
34
-
35
- ## Table of Contents
36
- - [Table of Contents](#table-of-contents)
37
- - [Dataset Description](#dataset-description)
38
- - [Dataset Summary](#dataset-summary)
39
- - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
40
- - [Languages](#languages)
41
- - [Dataset Structure](#dataset-structure)
42
- - [Data Instances](#data-instances)
43
- - [Data Fields](#data-fields)
44
- - [Data Splits](#data-splits)
45
- - [Dataset Creation](#dataset-creation)
46
- - [Curation Rationale](#curation-rationale)
47
- - [Source Data](#source-data)
48
- - [Annotations](#annotations)
49
- - [Personal and Sensitive Information](#personal-and-sensitive-information)
50
- - [Considerations for Using the Data](#considerations-for-using-the-data)
51
- - [Social Impact of Dataset](#social-impact-of-dataset)
52
- - [Discussion of Biases](#discussion-of-biases)
53
- - [Other Known Limitations](#other-known-limitations)
54
- - [Additional Information](#additional-information)
55
- - [Dataset Curators](#dataset-curators)
56
- - [Licensing Information](#licensing-information)
57
- - [Citation Information](#citation-information)
58
- - [Contributions](#contributions)
59
-
60
- ## Dataset Description
61
-
62
- - **Homepage:**
63
- - **Repository:**
64
- - **Paper:**
65
- - **Leaderboard:**
66
- - **Point of Contact:**
67
-
68
- ### Dataset Summary
69
-
70
- [More Information Needed]
71
-
72
- ### Supported Tasks and Leaderboards
73
-
74
- [More Information Needed]
75
-
76
- ### Languages
77
-
78
- [More Information Needed]
79
-
80
- ## Dataset Structure
81
-
82
- ### Data Instances
83
-
84
- [More Information Needed]
85
-
86
- ### Data Fields
87
-
88
- [More Information Needed]
89
-
90
- ### Data Splits
91
-
92
- [More Information Needed]
93
-
94
- ## Dataset Creation
95
-
96
- ### Curation Rationale
97
-
98
- [More Information Needed]
99
-
100
- ### Source Data
101
-
102
- #### Initial Data Collection and Normalization
103
-
104
- [More Information Needed]
105
-
106
- #### Who are the source language producers?
107
-
108
- [More Information Needed]
109
-
110
- ### Annotations
111
-
112
- #### Annotation process
113
-
114
- [More Information Needed]
115
-
116
- #### Who are the annotators?
117
-
118
- [More Information Needed]
119
-
120
- ### Personal and Sensitive Information
121
-
122
- [More Information Needed]
123
-
124
- ## Considerations for Using the Data
125
-
126
- ### Social Impact of Dataset
127
-
128
- [More Information Needed]
129
-
130
- ### Discussion of Biases
131
-
132
- [More Information Needed]
133
-
134
- ### Other Known Limitations
135
-
136
- [More Information Needed]
137
-
138
- ## Additional Information
139
-
140
- ### Dataset Curators
141
-
142
- [More Information Needed]
143
-
144
- ### Licensing Information
145
-
146
- [More Information Needed]
147
-
148
- ### Citation Information
149
-
150
- [More Information Needed]
151
-
152
- ### Contributions
153
-
154
- Thanks to [@github-scuwyh2000](https://github.com/scuwyh2000) for adding this dataset.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
coco.py DELETED
@@ -1,134 +0,0 @@
1
- # Copyright 2020 The HuggingFace Datasets Authors and the current dataset script contributor.
2
- #
3
- # Licensed under the Apache License, Version 2.0 (the "License");
4
- # you may not use this file except in compliance with the License.
5
- # You may obtain a copy of the License at
6
- #
7
- # http://www.apache.org/licenses/LICENSE-2.0
8
- #
9
- # Unless required by applicable law or agreed to in writing, software
10
- # distributed under the License is distributed on an "AS IS" BASIS,
11
- # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12
- # See the License for the specific language governing permissions and
13
- # limitations under the License.
14
- # TODO: Address all TODOs and remove all explanatory comments
15
- """TODO: Add a description here."""
16
-
17
-
18
- import csv
19
- import json
20
- import os
21
-
22
- import datasets
23
-
24
-
25
- # TODO: Add BibTeX citation
26
- # Find for instance the citation on arxiv or on the dataset repo/website
27
- _CITATION = """\
28
- @InProceedings{huggingface:dataset,
29
- title = {A great new dataset},
30
- author={huggingface, Inc.
31
- },
32
- year={2020}
33
- }
34
- """
35
-
36
- # TODO: Add description of the dataset here
37
- # You can copy an official description
38
- _DESCRIPTION = """\
39
- This new dataset is designed to solve this great NLP task and is crafted with a lot of care.
40
- """
41
-
42
- # TODO: Add a link to an official homepage for the dataset here
43
- _HOMEPAGE = ""
44
-
45
- # TODO: Add the licence for the dataset here if you can find it
46
- _LICENSE = ""
47
-
48
- # TODO: Add link to the official dataset URLs here
49
- # The HuggingFace Datasets library doesn't host the datasets but only points to the original files.
50
- # This can be an arbitrary nested dict/list of URLs (see below in `_split_generators` method)
51
- _URL = "dataset_coco.json"
52
- _URLS = {
53
- "train": _URL
54
- }
55
-
56
-
57
- # TODO: Name of the dataset usually match the script name with CamelCase instead of snake_case
58
- class NewDataset(datasets.GeneratorBasedBuilder):
59
- """TODO: Short description of my dataset."""
60
-
61
- VERSION = datasets.Version("1.1.0")
62
-
63
- # This is an example of a dataset with multiple configurations.
64
- # If you don't want/need to define several sub-sets in your dataset,
65
- # just remove the BUILDER_CONFIG_CLASS and the BUILDER_CONFIGS attributes.
66
-
67
- # If you need to make complex sub-parts in the datasets with configurable options
68
- # You can create your own builder configuration class to store attribute, inheriting from datasets.BuilderConfig
69
- # BUILDER_CONFIG_CLASS = MyBuilderConfig
70
-
71
- # You will be able to load one or the other configurations in the following list with
72
- # data = datasets.load_dataset('my_dataset', 'first_domain')
73
- # data = datasets.load_dataset('my_dataset', 'second_domain')
74
- #BUILDER_CONFIGS = [
75
- # datasets.BuilderConfig(name="images", version=VERSION, description="This part of my dataset covers a first domain"),
76
- #]
77
-
78
- #DEFAULT_CONFIG_NAME = "train" # It's not mandatory to have a default configuration. Just use one if it make sense.
79
-
80
- def _info(self):
81
- # TODO: This method specifies the datasets.DatasetInfo object which contains informations and typings for the dataset
82
- features=datasets.Features(
83
- {
84
- "filepath": datasets.Value("string"),
85
- "filename": datasets.Value("string")
86
- #"raw": datasets.Value("string")
87
- }
88
- ),
89
- return datasets.DatasetInfo(
90
- # This is the description that will appear on the datasets page.
91
- description=_DESCRIPTION,
92
- # This defines the different columns of the dataset and their types
93
- features=features, # Here we define them above because they are different between the two configurations
94
- # If there's a common (input, target) tuple from the features, uncomment supervised_keys line below and
95
- # specify them. They'll be used if as_supervised=True in builder.as_dataset.
96
- # supervised_keys=("sentence", "label"),
97
- # Homepage of the dataset for documentation
98
- homepage=_HOMEPAGE,
99
- # License for the dataset if available
100
- license=_LICENSE,
101
- # Citation for the dataset
102
- citation=_CITATION,
103
- )
104
-
105
- def _split_generators(self, dl_manager):
106
- urls_to_download = _URLS
107
- downloaded_files = dl_manager.download_and_extract(urls_to_download)
108
- print(download_files['train'])
109
- return [
110
- datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={"filepath": downloaded_files["train"]}),
111
- # datasets.SplitGenerator(name=datasets.Split.VALIDATION, gen_kwargs={"filepath": downloaded_files["dev"]}),
112
- ]
113
-
114
- # method parameters are unpacked from `gen_kwargs` as given in `_split_generators`
115
- def _generate_examples(self, filepath):
116
- """This function returns the examples in the raw (text) form."""
117
- logger.info("generating examples from = %s", filepath)
118
- print(download_files['train'])
119
- with open(filepath, encoding="utf-8") as f:
120
- data = json.loads(f)
121
- for image in data["images"]:
122
- file_path = image["filepath"]
123
- file_name = image["filename"]
124
- yield id_, {
125
- "filepath": file_path,
126
- "filename": file_name
127
- }
128
- #for sentence in image["sentences"]:
129
- # raw = sentence["raw"]
130
- # yield id_, {
131
- # "filepath": filepath,
132
- # "filename": filename,
133
- # "raw": raw
134
- # }