File size: 945 Bytes
38dcdc6
 
 
 
 
 
 
 
 
 
f50b9cf
 
 
 
 
38dcdc6
f50b9cf
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
ad12092
f50b9cf
ad12092
f50b9cf
ad12092
f50b9cf
 
 
ad12092
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
---
base_model: meta-llama/Meta-Llama-3-8B
language:
- sv
- da
- 'no'
pipeline_tag: text-generation
inference:
  parameters:
    temperature: 0.6
tags:
 - pytorch
 - llama
 - llama-3
 - ai-sweden
---

### Use with transformers

See the snippet below for usage with Transformers:

```python
>>> import transformers
>>> import torch

>>> model_id = "AI-Sweden-Models/Llama-3-8B"

>>> pipeline = transformers.pipeline(
    "text-generation",
    model=model_id,
    model_kwargs={"torch_dtype": torch.bfloat16},
    device_map="auto"
)
>>> pipeline("Hey how are you doing today?")
```

## Training information

`AI-Sweden-Models/Llama-3-8B` is a continuation of the pretraining process from `meta-llama/Meta-Llama-3-8B`. It was trained on a subset from [The nordic pile](https://arxiv.org/abs/2303.17183) containing Swedish, Norweigian and Danish.

A total of 92 A100 gpus was used, and roughly 250GB of data was used.

## Benchmarks 

Coming soon.