gsaltintas commited on
Commit
435e7ea
·
verified ·
1 Parent(s): b3c4cd0

Uploading temporal_expressions subset

Browse files
README.md CHANGED
@@ -14,8 +14,14 @@ configs:
14
  path: social_media_informal_text/tokenizer_robustness_social_media_informal_text-*
15
  - split: dev
16
  path: social_media_informal_text/dev-*
 
 
 
 
 
 
17
  dataset_info:
18
- config_name: social_media_informal_text
19
  features:
20
  - name: question
21
  dtype: string
@@ -54,6 +60,45 @@ dataset_info:
54
  num_examples: 47
55
  download_size: 22875
56
  dataset_size: 23370
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
57
  ---
58
 
59
  # Dataset Card for Tokenization Robustness
 
14
  path: social_media_informal_text/tokenizer_robustness_social_media_informal_text-*
15
  - split: dev
16
  path: social_media_informal_text/dev-*
17
+ - config_name: temporal_expressions
18
+ data_files:
19
+ - split: tokenizer_robustness_temporal_expressions
20
+ path: temporal_expressions/tokenizer_robustness_temporal_expressions-*
21
+ - split: dev
22
+ path: temporal_expressions/dev-*
23
  dataset_info:
24
+ - config_name: social_media_informal_text
25
  features:
26
  - name: question
27
  dtype: string
 
60
  num_examples: 47
61
  download_size: 22875
62
  dataset_size: 23370
63
+ - config_name: temporal_expressions
64
+ features:
65
+ - name: question
66
+ dtype: string
67
+ - name: choices
68
+ sequence: string
69
+ - name: answer
70
+ dtype: int64
71
+ - name: answer_label
72
+ dtype: string
73
+ - name: split
74
+ dtype: string
75
+ - name: subcategories
76
+ dtype: string
77
+ - name: lang
78
+ dtype: string
79
+ - name: second_lang
80
+ dtype: string
81
+ - name: coding_lang
82
+ dtype: string
83
+ - name: notes
84
+ dtype: string
85
+ - name: id
86
+ dtype: string
87
+ - name: set_id
88
+ dtype: string
89
+ - name: variation_id
90
+ dtype: string
91
+ - name: __index_level_0__
92
+ dtype: int64
93
+ splits:
94
+ - name: tokenizer_robustness_temporal_expressions
95
+ num_bytes: 4603
96
+ num_examples: 21
97
+ - name: dev
98
+ num_bytes: 1306
99
+ num_examples: 6
100
+ download_size: 15229
101
+ dataset_size: 5909
102
  ---
103
 
104
  # Dataset Card for Tokenization Robustness
temporal_expressions/dev-00000-of-00001.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:39b0a4405b9a215f2f4b641aa3e81a68574867a5edce78323b9bf75ea3e53392
3
- size 6852
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f53d7b3d5cf16f55786365841bcf4f6a55d586de1707b55c6b854431cfc0c8f3
3
+ size 6859
temporal_expressions/tokenizer_robustness_temporal_expressions-00000-of-00001.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:eb34d7694be39da84ea733c11efecbe4e59366d54b9722e26503824bdc186c68
3
- size 8373
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b30e63a4e7ac9898cdb8d186e6c8f83e07798a1a72f9060aad9eb9e7398fce5a
3
+ size 8370