Update README.md
Browse files
README.md
CHANGED
@@ -21,7 +21,7 @@ base_model: meta-llama/Meta-Llama-3.1-8B-Instruct
|
|
21 |
|
22 |
# Plutus-Meta-Llama-3.1-8B-Instruct-bnb-4bit
|
23 |
|
24 |
-
Plutus-Meta-Llama-3.1-8B-Instruct-bnb-4bit is a fine-tuned version of the LLaMA-3.1-8B model, specifically optimized for tasks
|
25 |
|
26 |

|
27 |
|
@@ -35,39 +35,39 @@ Plutus-Meta-Llama-3.1-8B-Instruct-bnb-4bit is a fine-tuned version of the LLaMA-
|
|
35 |
|
36 |
## Training
|
37 |
|
38 |
-
|
39 |
|
40 |
-
- **Finance and Investment**:
|
41 |
-
- **Trading Strategies**: Focused on technical analysis, options trading, and
|
42 |
-
- **Risk Management**:
|
43 |
-
- **Behavioral Finance and Psychology**:
|
44 |
-
- **Social Engineering and
|
|
|
45 |
|
|
|
46 |
|
47 |
## Intended Use
|
48 |
|
49 |
-
|
50 |
|
51 |
-
- **Financial Analysis**:
|
52 |
-
- **
|
53 |
-
- **Behavioral Finance Research**:
|
54 |
-
- **Cybersecurity and Social Engineering**:
|
55 |
|
56 |
## Performance
|
57 |
|
58 |
-
While
|
59 |
|
60 |
## Limitations
|
61 |
|
62 |
-
|
63 |
-
|
64 |
-
- **
|
65 |
-
- **Inference Speed**: Although optimized with 4-bit quantization, real-time application latency may still be an issue depending on the deployment environment.
|
66 |
-
- **Context Length**: The model has a limited context window, which can affect its ability to process long-form documents or complex multi-turn conversations effectively.
|
67 |
|
68 |
## How to Use
|
69 |
|
70 |
-
You can load and use the model with the following code:
|
71 |
|
72 |
```python
|
73 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
|
|
21 |
|
22 |
# Plutus-Meta-Llama-3.1-8B-Instruct-bnb-4bit
|
23 |
|
24 |
+
**Plutus-Meta-Llama-3.1-8B-Instruct-bnb-4bit** is a fine-tuned version of the LLaMA-3.1-8B model, specifically optimized for tasks in finance, economics, trading, psychology, and social engineering. This model leverages the LLaMA architecture, enhanced with 4-bit quantization to deliver high performance in resource-constrained environments, while maintaining accuracy and relevance for natural language processing tasks in these domains.
|
25 |
|
26 |

|
27 |
|
|
|
35 |
|
36 |
## Training
|
37 |
|
38 |
+
The model was fine-tuned on the comprehensive [**"Financial, Economic, and Psychological Analysis Texts"** dataset](https://huggingface.co/datasets/0xroyce/Plutus), which consists of 394 books covering key areas like:
|
39 |
|
40 |
+
- **Finance and Investment**: Stock market analysis, value investing, bonds, and exchange-traded funds (ETFs).
|
41 |
+
- **Trading Strategies**: Focused on technical analysis, options trading, algorithmic strategies, and risk management.
|
42 |
+
- **Risk Management**: Quantitative approaches to financial risk and volatility analysis.
|
43 |
+
- **Behavioral Finance and Psychology**: Psychological aspects of trading, persuasion techniques, and investor behavior.
|
44 |
+
- **Social Engineering and Cybersecurity**: Highlighting manipulation techniques, security vulnerabilities, and deception research.
|
45 |
+
- **Military Strategy and Psychological Operations**: Strategic insights into psychological warfare, military intelligence, and influence operations.
|
46 |
|
47 |
+
The dataset covers broad domains, making this model highly versatile for specific use cases related to economic theory, financial markets, cybersecurity, and social engineering.
|
48 |
|
49 |
## Intended Use
|
50 |
|
51 |
+
Plutus-Meta-Llama-3.1-8B-Instruct-bnb-4bit is suitable for a wide variety of natural language processing tasks, particularly in finance, economics, psychology, and cybersecurity. Common use cases include:
|
52 |
|
53 |
+
- **Financial Analysis**: Extract insights and perform sentiment analysis on financial documents.
|
54 |
+
- **Market Predictions**: Generate contextually relevant market predictions and economic theories.
|
55 |
+
- **Behavioral Finance Research**: Explore trading psychology and investor decision-making through text generation.
|
56 |
+
- **Cybersecurity and Social Engineering**: Study manipulation tactics and create content related to cyber threats and defense strategies.
|
57 |
|
58 |
## Performance
|
59 |
|
60 |
+
While benchmark scores specific to Plutus-Meta-Llama-3.1-8B-Instruct-bnb-4bit are not available, it has been optimized to balance computational efficiency and accuracy. The 4-bit quantization allows the model to operate in resource-constrained environments, making it ideal for use cases where hardware limitations are a factor.
|
61 |
|
62 |
## Limitations
|
63 |
|
64 |
+
- **Domain-Specific Bias**: As the model is trained on specialized data, it may generate biased content, particularly in the areas of finance, psychology, and social engineering.
|
65 |
+
- **Context Length**: Limited context length may affect the ability to handle long or complex inputs effectively.
|
66 |
+
- **Inference Speed**: Despite being optimized for 4-bit quantization, real-time application latency may be an issue in certain environments.
|
|
|
|
|
67 |
|
68 |
## How to Use
|
69 |
|
70 |
+
You can load and use the model with the following Python code:
|
71 |
|
72 |
```python
|
73 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|