|
--- |
|
base_model: |
|
- google/gemma-2-9b-it |
|
library_name: transformers |
|
--- |
|
|
|
# MISHANM/google-gemma-2-9b-it.gguf |
|
|
|
This model is a GGUF version of the Google gemma-2-9b-it model, optimized for use with the `llama.cpp` framework. It is designed to run efficiently on CPUs and can be used for various natural language processing tasks. |
|
|
|
## Model Details |
|
1. Language: English |
|
2. Tasks: Text generation |
|
3. Base Model: google/gemma-2-9b-it |
|
|
|
## Building and Running the Model |
|
|
|
To build and run the model using `llama.cpp`, follow these steps: |
|
|
|
### Build llama.cpp Locally |
|
|
|
```bash |
|
git clone https://github.com/ggerganov/llama.cpp |
|
cd llama.cpp |
|
cmake -B build |
|
cmake --build build --config Release |
|
|
|
``` |
|
## Run the Model |
|
|
|
Navigate to the build directory and run the model with a prompt: |
|
|
|
``` |
|
cd llama.cpp/build/bin |
|
``` |
|
## Inference with llama.cpp |
|
|
|
``` |
|
./llama-cli -m /path/to/model/ -p "Your prompt here" -n 128 |
|
``` |
|
|
|
## Citation Information |
|
``` |
|
@misc{MISHANM/google-gemma-2-9b-it.gguf, |
|
author = {Mishan Maurya}, |
|
title = {Introducing Google gemma-2-9b-it GGUF Model}, |
|
year = {2025}, |
|
publisher = {Hugging Face}, |
|
journal = {Hugging Face repository}, |
|
|
|
} |
|
``` |