parquet-converter commited on
Commit
3b23942
·
1 Parent(s): 5e19492

Update parquet files

Browse files
.gitattributes DELETED
@@ -1,37 +0,0 @@
1
- *.7z filter=lfs diff=lfs merge=lfs -text
2
- *.arrow filter=lfs diff=lfs merge=lfs -text
3
- *.bin filter=lfs diff=lfs merge=lfs -text
4
- *.bz2 filter=lfs diff=lfs merge=lfs -text
5
- *.ftz filter=lfs diff=lfs merge=lfs -text
6
- *.gz filter=lfs diff=lfs merge=lfs -text
7
- *.h5 filter=lfs diff=lfs merge=lfs -text
8
- *.joblib filter=lfs diff=lfs merge=lfs -text
9
- *.lfs.* filter=lfs diff=lfs merge=lfs -text
10
- *.model filter=lfs diff=lfs merge=lfs -text
11
- *.msgpack filter=lfs diff=lfs merge=lfs -text
12
- *.onnx filter=lfs diff=lfs merge=lfs -text
13
- *.ot filter=lfs diff=lfs merge=lfs -text
14
- *.parquet filter=lfs diff=lfs merge=lfs -text
15
- *.pb filter=lfs diff=lfs merge=lfs -text
16
- *.pt filter=lfs diff=lfs merge=lfs -text
17
- *.pth filter=lfs diff=lfs merge=lfs -text
18
- *.rar filter=lfs diff=lfs merge=lfs -text
19
- saved_model/**/* filter=lfs diff=lfs merge=lfs -text
20
- *.tar.* filter=lfs diff=lfs merge=lfs -text
21
- *.tflite filter=lfs diff=lfs merge=lfs -text
22
- *.tgz filter=lfs diff=lfs merge=lfs -text
23
- *.wasm filter=lfs diff=lfs merge=lfs -text
24
- *.xz filter=lfs diff=lfs merge=lfs -text
25
- *.zip filter=lfs diff=lfs merge=lfs -text
26
- *.zstandard filter=lfs diff=lfs merge=lfs -text
27
- *tfevents* filter=lfs diff=lfs merge=lfs -text
28
- # Audio files - uncompressed
29
- *.pcm filter=lfs diff=lfs merge=lfs -text
30
- *.sam filter=lfs diff=lfs merge=lfs -text
31
- *.raw filter=lfs diff=lfs merge=lfs -text
32
- # Audio files - compressed
33
- *.aac filter=lfs diff=lfs merge=lfs -text
34
- *.flac filter=lfs diff=lfs merge=lfs -text
35
- *.mp3 filter=lfs diff=lfs merge=lfs -text
36
- *.ogg filter=lfs diff=lfs merge=lfs -text
37
- *.wav filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
README.md DELETED
@@ -1,3 +0,0 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
medwiki/wikimedqa-test.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:278b8f62c8186260fecd8eaec551946ad594d57a12735d810210e0e1b053a367
3
+ size 3687221
medwiki/wikimedqa-train.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:34fad5a97062121fa34f5597719c21b95355f0a621dde9c7ba626127fb83e463
3
+ size 65851652
medwiki/wikimedqa-validation.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4a18d84e39dfbcc8571cb1e81495fe38a307969e2c50634ddf45b2c11c73bc5b
3
+ size 3686638
wikem/wikimedqa-test.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:53c673ea3342d25f7044ddb913e07ce4f12587f0e68caa14172291fed2d48a82
3
+ size 297814
wikem/wikimedqa-train.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:31d052990224c4dd81a0caa7b4845593767b79a2360d2209a915d3effb046ac9
3
+ size 4973024
wikem/wikimedqa-validation.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:86a6e29f8d58d72b61398ff19792f074804d1eb9093ef0148b81f4f6081b0e9b
3
+ size 278188
wikidoc/wikimedqa-test.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:00b6d7dcbfce225cc3d2ecaa40bdf26f5d687a180c40603c2b20417be5d61319
3
+ size 652586
wikidoc/wikimedqa-train.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c56b92c8ae18ced3cd77645f3f927ab40d9ea7a62ac72693b041d4b3a5f09839
3
+ size 11687310
wikidoc/wikimedqa-validation.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5cd87379d6cde5be53a0d45188898a6aee3cdde0fe7be1c96dc174a5f4e61812
3
+ size 654176
wikimedqa.py DELETED
@@ -1,131 +0,0 @@
1
- # coding=utf-8
2
- # Copyright 2020 The TensorFlow Datasets Authors and the HuggingFace Datasets Authors.
3
- #
4
- # Licensed under the Apache License, Version 2.0 (the "License");
5
- # you may not use this file except in compliance with the License.
6
- # You may obtain a copy of the License at
7
- #
8
- # http://www.apache.org/licenses/LICENSE-2.0
9
- #
10
- # Unless required by applicable law or agreed to in writing, software
11
- # distributed under the License is distributed on an "AS IS" BASIS,
12
- # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13
- # See the License for the specific language governing permissions and
14
- # limitations under the License.
15
-
16
- # Lint as: python3
17
-
18
-
19
- import csv
20
- import os
21
- import textwrap
22
- import numpy as np
23
- import datasets
24
- import pandas as pd
25
-
26
-
27
- _CITATION = """\
28
- Anonymous submission
29
- """
30
-
31
- _DESCRIPTION = """\
32
- Anonymous submission
33
- """
34
-
35
- URL = 'https://sileod.s3.eu-west-3.amazonaws.com/wikimedqa/'
36
-
37
-
38
- class WikiMedQAConfig(datasets.BuilderConfig):
39
- """BuilderConfig for WikiMedQA."""
40
-
41
- def __init__(
42
- self,
43
- data_dir,
44
- label_classes=None,
45
- process_label=lambda x: x,
46
- **kwargs,
47
- ):
48
-
49
- super(WikiMedQAConfig, self).__init__(version=datasets.Version("1.0.5", ""), **kwargs)
50
- self.text_features = {k:k for k in ['text']+[f'option_{i}' for i in range(8)]}
51
- self.label_column = 'label'
52
- self.label_classes = list('01234567')
53
- self.data_url = URL
54
- self.url=URL
55
- self.data_dir=data_dir
56
- self.citation = _CITATION
57
- self.process_label = process_label
58
-
59
-
60
- class WikiMedQA(datasets.GeneratorBasedBuilder):
61
- """Evaluation of word estimative of probability understanding"""
62
-
63
- BUILDER_CONFIGS = [
64
- WikiMedQAConfig(
65
- name="medwiki",
66
- data_dir="medwiki"),
67
- WikiMedQAConfig(
68
- name="wikem",
69
- data_dir="wikem"),
70
- WikiMedQAConfig(
71
- name="wikidoc",
72
- data_dir="wikidoc"),
73
- ]
74
-
75
- def _info(self):
76
- features = {text_feature: datasets.Value("string") for text_feature in self.config.text_features.keys()}
77
- features["label"] = datasets.features.ClassLabel(names=self.config.label_classes)
78
- features["idx"] = datasets.Value("int32")
79
-
80
- return datasets.DatasetInfo(
81
- description=_DESCRIPTION,
82
- features=datasets.Features(features),
83
- homepage=self.config.url,
84
- citation=self.config.citation + "\n" + _CITATION,
85
- )
86
- def _split_generators(self, dl_manager):
87
-
88
- data_dirs=[]
89
- for split in ['train','validation','test']:
90
- url=f'{URL}{self.config.data_dir}.csv'
91
- print(url)
92
- data_dirs+=[dl_manager.download(url)]
93
- print(data_dirs)
94
- return [
95
- datasets.SplitGenerator(
96
- name=datasets.Split.TRAIN,
97
- gen_kwargs={
98
- "data_file": data_dirs[0],
99
- "split": "train",
100
- },
101
- ),
102
- datasets.SplitGenerator(
103
- name=datasets.Split.VALIDATION,
104
- gen_kwargs={
105
- "data_file": data_dirs[1],
106
- "split": "dev",
107
- },
108
- ),
109
- datasets.SplitGenerator(
110
- name=datasets.Split.TEST,
111
- gen_kwargs={
112
- "data_file": data_dirs[2],
113
- "split": "test",
114
- },
115
- ),
116
- ]
117
-
118
- def _generate_examples(self, data_file, split):
119
- df = pd.read_csv(data_file)
120
- df=df[['text','options','label']]
121
- train, dev, test = np.split(df.sample(frac=1, random_state=42),
122
- [int(.9*len(df)), int(.95*len(df))])
123
- df=eval(split)
124
- df['options']=df['options'].map(eval)
125
- for i in range(8):
126
- df[f'option_{i}']=df.options.map(lambda x:x[i])
127
- del df['options']
128
- df['idx']=df.index
129
- for idx, example in df.iterrows():
130
- yield idx, dict(example)
131
-