Llama-3-8B / README.md
timpal0l's picture
Update README.md
ad12092 verified
|
raw
history blame
945 Bytes
metadata
base_model: meta-llama/Meta-Llama-3-8B
language:
  - sv
  - da
  - 'no'
pipeline_tag: text-generation
inference:
  parameters:
    temperature: 0.6
tags:
  - pytorch
  - llama
  - llama-3
  - ai-sweden

Use with transformers

See the snippet below for usage with Transformers:

>>> import transformers
>>> import torch

>>> model_id = "AI-Sweden-Models/Llama-3-8B"

>>> pipeline = transformers.pipeline(
    "text-generation",
    model=model_id,
    model_kwargs={"torch_dtype": torch.bfloat16},
    device_map="auto"
)
>>> pipeline("Hey how are you doing today?")

Training information

AI-Sweden-Models/Llama-3-8B is a continuation of the pretraining process from meta-llama/Meta-Llama-3-8B. It was trained on a subset from The nordic pile containing Swedish, Norweigian and Danish.

A total of 92 A100 gpus was used, and roughly 250GB of data was used.

Benchmarks

Coming soon.