Update README.md
Browse files
README.md
CHANGED
@@ -64,6 +64,9 @@ The model was trained on [Japanese C4](https://huggingface.co/datasets/allenai/c
|
|
64 |
# Tokenization
|
65 |
The model uses a [sentencepiece](https://github.com/google/sentencepiece)-based tokenizer. The vocabulary was first trained on a selected subset from the training data using the official sentencepiece training script, and then augmented with emojis and symbols.
|
66 |
|
|
|
|
|
|
|
67 |
# How to cite
|
68 |
```bibtex
|
69 |
@misc{rinna-japanese-gpt-1b,
|
|
|
64 |
# Tokenization
|
65 |
The model uses a [sentencepiece](https://github.com/google/sentencepiece)-based tokenizer. The vocabulary was first trained on a selected subset from the training data using the official sentencepiece training script, and then augmented with emojis and symbols.
|
66 |
|
67 |
+
# Release date
|
68 |
+
January 26, 2022
|
69 |
+
|
70 |
# How to cite
|
71 |
```bibtex
|
72 |
@misc{rinna-japanese-gpt-1b,
|