gsaltintas commited on
Commit
6191d77
·
verified ·
1 Parent(s): 3ae6b9a

Uploading temporal_expressions subset

Browse files
README.md CHANGED
@@ -16,8 +16,16 @@ configs:
16
  path: social_media_informal_text/dev-*
17
  - split: test
18
  path: social_media_informal_text/test-*
 
 
 
 
 
 
 
 
19
  dataset_info:
20
- config_name: social_media_informal_text
21
  features:
22
  - name: question
23
  dtype: string
@@ -59,6 +67,48 @@ dataset_info:
59
  num_examples: 59
60
  download_size: 35201
61
  dataset_size: 36808
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
62
  ---
63
 
64
  # Dataset Card for Tokenization Robustness
 
16
  path: social_media_informal_text/dev-*
17
  - split: test
18
  path: social_media_informal_text/test-*
19
+ - config_name: temporal_expressions
20
+ data_files:
21
+ - split: tokenizer_robustness_temporal_expressions
22
+ path: temporal_expressions/tokenizer_robustness_temporal_expressions-*
23
+ - split: dev
24
+ path: temporal_expressions/dev-*
25
+ - split: test
26
+ path: temporal_expressions/test-*
27
  dataset_info:
28
+ - config_name: social_media_informal_text
29
  features:
30
  - name: question
31
  dtype: string
 
67
  num_examples: 59
68
  download_size: 35201
69
  dataset_size: 36808
70
+ - config_name: temporal_expressions
71
+ features:
72
+ - name: question
73
+ dtype: string
74
+ - name: choices
75
+ sequence: string
76
+ - name: answer
77
+ dtype: int64
78
+ - name: answer_label
79
+ dtype: string
80
+ - name: split
81
+ dtype: string
82
+ - name: subcategories
83
+ dtype: string
84
+ - name: lang
85
+ dtype: string
86
+ - name: second_lang
87
+ dtype: string
88
+ - name: coding_lang
89
+ dtype: string
90
+ - name: notes
91
+ dtype: string
92
+ - name: id
93
+ dtype: string
94
+ - name: set_id
95
+ dtype: string
96
+ - name: variation_id
97
+ dtype: string
98
+ - name: __index_level_0__
99
+ dtype: int64
100
+ splits:
101
+ - name: tokenizer_robustness_temporal_expressions
102
+ num_bytes: 4603
103
+ num_examples: 21
104
+ - name: dev
105
+ num_bytes: 1306
106
+ num_examples: 6
107
+ - name: test
108
+ num_bytes: 4687
109
+ num_examples: 21
110
+ download_size: 23619
111
+ dataset_size: 10596
112
  ---
113
 
114
  # Dataset Card for Tokenization Robustness
temporal_expressions/dev-00000-of-00001.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:55fc08990d876bfc419ee170df69a4095bcc7f7f431f9c200b4104c05c5869dd
3
- size 6861
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:39b0a4405b9a215f2f4b641aa3e81a68574867a5edce78323b9bf75ea3e53392
3
+ size 6852
temporal_expressions/tokenizer_robustness_temporal_expressions-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eb34d7694be39da84ea733c11efecbe4e59366d54b9722e26503824bdc186c68
3
+ size 8373