fuzzy-mittenz commited on
Commit
d6bb0ac
·
verified ·
1 Parent(s): b0067b4

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -36
README.md CHANGED
@@ -14,13 +14,11 @@ tags:
14
  - reinforcement learning
15
  - pytorch
16
  - llama-cpp
17
- - gguf-my-repo
18
  base_model: nvidia/AceReason-Nemotron-7B
19
  ---
20
 
21
- # fuzzy-mittenz/AceReason-Nemotron-7B-Q4_K_M-GGUF
22
- This model was converted to GGUF format from [`nvidia/AceReason-Nemotron-7B`](https://huggingface.co/nvidia/AceReason-Nemotron-7B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
23
- Refer to the [original model card](https://huggingface.co/nvidia/AceReason-Nemotron-7B) for more details on the model.
24
 
25
  ## Use with llama.cpp
26
  Install llama.cpp through brew (works on Mac and Linux)
@@ -29,35 +27,4 @@ Install llama.cpp through brew (works on Mac and Linux)
29
  brew install llama.cpp
30
 
31
  ```
32
- Invoke the llama.cpp server or the CLI.
33
-
34
- ### CLI:
35
- ```bash
36
- llama-cli --hf-repo fuzzy-mittenz/AceReason-Nemotron-7B-Q4_K_M-GGUF --hf-file acereason-nemotron-7b-q4_k_m.gguf -p "The meaning to life and the universe is"
37
- ```
38
-
39
- ### Server:
40
- ```bash
41
- llama-server --hf-repo fuzzy-mittenz/AceReason-Nemotron-7B-Q4_K_M-GGUF --hf-file acereason-nemotron-7b-q4_k_m.gguf -c 2048
42
- ```
43
-
44
- Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
45
-
46
- Step 1: Clone llama.cpp from GitHub.
47
- ```
48
- git clone https://github.com/ggerganov/llama.cpp
49
- ```
50
-
51
- Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
52
- ```
53
- cd llama.cpp && LLAMA_CURL=1 make
54
- ```
55
-
56
- Step 3: Run inference through the main binary.
57
- ```
58
- ./llama-cli --hf-repo fuzzy-mittenz/AceReason-Nemotron-7B-Q4_K_M-GGUF --hf-file acereason-nemotron-7b-q4_k_m.gguf -p "The meaning to life and the universe is"
59
- ```
60
- or
61
- ```
62
- ./llama-server --hf-repo fuzzy-mittenz/AceReason-Nemotron-7B-Q4_K_M-GGUF --hf-file acereason-nemotron-7b-q4_k_m.gguf -c 2048
63
- ```
 
14
  - reinforcement learning
15
  - pytorch
16
  - llama-cpp
 
17
  base_model: nvidia/AceReason-Nemotron-7B
18
  ---
19
 
20
+ # IntelligentEstate/Gambit-7B-Q4_K_M-GGUF
21
+ This model was converted to GGUF format from [`nvidia/AceReason-Nemotron-7B`](https://huggingface.co/nvidia/AceReason-Nemotron-7B)
 
22
 
23
  ## Use with llama.cpp
24
  Install llama.cpp through brew (works on Mac and Linux)
 
27
  brew install llama.cpp
28
 
29
  ```
30
+ Invoke the llama.cpp server or the CLI.