Upload folder using huggingface_hub
Browse files
README.md
ADDED
@@ -0,0 +1,49 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: apache-2.0
|
3 |
+
tags:
|
4 |
+
- speculators
|
5 |
+
- eagle
|
6 |
+
- speculative-decoding
|
7 |
+
---
|
8 |
+
|
9 |
+
# Eagle Llama 3.1 8B Instruct
|
10 |
+
|
11 |
+
This is a converted Eagle speculator checkpoint for Llama 3.1 8B Instruct, compatible with the [speculators](https://github.com/neuralmagic/speculators) library.
|
12 |
+
|
13 |
+
## Model Details
|
14 |
+
|
15 |
+
- **Original checkpoint**: [yuhuili/EAGLE-LLaMA3.1-Instruct-8B](https://huggingface.co/yuhuili/EAGLE-LLaMA3.1-Instruct-8B)
|
16 |
+
- **Base model**: [meta-llama/Meta-Llama-3.1-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3.1-8B-Instruct)
|
17 |
+
- **Architecture**: EAGLE (standard, without layernorms)
|
18 |
+
- **Size**: 481MB
|
19 |
+
|
20 |
+
## Usage
|
21 |
+
|
22 |
+
```python
|
23 |
+
from speculators.models.eagle import EagleSpeculator
|
24 |
+
from transformers import AutoModelForCausalLM
|
25 |
+
|
26 |
+
# Load the Eagle speculator
|
27 |
+
eagle_model = EagleSpeculator.from_pretrained("nm-testing/eagle-llama3.1-8b-instruct")
|
28 |
+
|
29 |
+
# Attach verifier model
|
30 |
+
verifier = AutoModelForCausalLM.from_pretrained("meta-llama/Meta-Llama-3.1-8B-Instruct")
|
31 |
+
eagle_model.attach_verifier(verifier)
|
32 |
+
|
33 |
+
# Use for speculative decoding
|
34 |
+
# ... your speculative decoding code here ...
|
35 |
+
```
|
36 |
+
|
37 |
+
## Conversion Details
|
38 |
+
|
39 |
+
This checkpoint was converted using:
|
40 |
+
```bash
|
41 |
+
speculators convert --eagle \
|
42 |
+
yuhuili/EAGLE-LLaMA3.1-Instruct-8B \
|
43 |
+
./eagle-standard-converted \
|
44 |
+
meta-llama/Meta-Llama-3.1-8B-Instruct
|
45 |
+
```
|
46 |
+
|
47 |
+
## License
|
48 |
+
|
49 |
+
Apache 2.0
|