MISHANM's picture
Create README.md
c149a55 verified
---
base_model:
- google/gemma-2-9b-it
library_name: transformers
---
# MISHANM/google-gemma-2-9b-it.gguf
This model is a GGUF version of the Google gemma-2-9b-it model, optimized for use with the `llama.cpp` framework. It is designed to run efficiently on CPUs and can be used for various natural language processing tasks.
## Model Details
1. Language: English
2. Tasks: Text generation
3. Base Model: google/gemma-2-9b-it
## Building and Running the Model
To build and run the model using `llama.cpp`, follow these steps:
### Build llama.cpp Locally
```bash
git clone https://github.com/ggerganov/llama.cpp
cd llama.cpp
cmake -B build
cmake --build build --config Release
```
## Run the Model
Navigate to the build directory and run the model with a prompt:
```
cd llama.cpp/build/bin
```
## Inference with llama.cpp
```
./llama-cli -m /path/to/model/ -p "Your prompt here" -n 128
```
## Citation Information
```
@misc{MISHANM/google-gemma-2-9b-it.gguf,
author = {Mishan Maurya},
title = {Introducing Google gemma-2-9b-it GGUF Model},
year = {2025},
publisher = {Hugging Face},
journal = {Hugging Face repository},
}
```