Daemontatox commited on
Commit
27ed197
·
verified ·
1 Parent(s): c18e0c7

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +130 -130
README.md CHANGED
@@ -1,131 +1,131 @@
1
- ---
2
- license: apache-2.0
3
- language:
4
- - zho
5
- - eng
6
- - fra
7
- - spa
8
- - por
9
- - deu
10
- - ita
11
- - rus
12
- - jpn
13
- - kor
14
- - vie
15
- - tha
16
- - ara
17
- base_model:
18
- - Qwen/Qwen2.5-72B-Instruct
19
- pipeline_tag: text-generation
20
- library_name: transformers
21
- tags:
22
- - reasoning
23
- - logic
24
- - cot
25
- - text-generation-inference
26
- new_version: Daemontatox/Cogito-Maximus
27
- ---
28
-
29
- ![image](./image.webp)
30
-
31
-
32
- ## **Model Overview**
33
-
34
- This model, **Cogito-Maximus**, is a fine-tuned version of the `unsloth/qwen2.5-72b-instruct-bnb-4bit` base model, optimized for advanced text generation tasks. It leverages the power of **Unsloth** and **Huggingface's TRL (Transformer Reinforcement Learning)** library to achieve faster training and improved performance.
35
-
36
- ### **Key Features**
37
- - **Base Model:** `unsloth/qwen2.5-72b-instruct`
38
- - **Training Acceleration:** Trained 2x faster using [Unsloth](https://github.com/unslothai/unsloth).
39
- - **Fine-Tuning Framework:** Utilizes Huggingface's [TRL](https://github.com/huggingface/trl) library.
40
- - **Optimized for Inference:** Ready for deployment in text-generation tasks with efficient inference capabilities.
41
- - **License:** Apache-2.0
42
-
43
- ---
44
-
45
- ## **Model Details**
46
-
47
- ### **Developed by**
48
- - **Author:** Daemontatox
49
- - **Organization:** Independent Contributor
50
-
51
- ### **Tags**
52
- - Text Generation Inference
53
- - Transformers
54
- - Unsloth
55
- - Qwen2
56
- - TRL
57
-
58
- ### **Language**
59
- - English (`en`)
60
-
61
- ### **License**
62
- This model is released under the **Apache-2.0 License**, which allows for free use, modification, and distribution, provided the original license and copyright notice are included.
63
-
64
- ---
65
-
66
- ## **Model Training**
67
-
68
- ### **Base Model**
69
- The model is derived from the `unsloth/qwen2.5-72b-instruct`, a version of the Qwen2.5-72B instruction-tuned model. The base model is optimized for efficiency using **bitsandbytes (bnb)** 4-bit quantization.
70
-
71
- ### **Training Process**
72
- - **Framework:** The model was fine-tuned using **Unsloth**, a library designed to accelerate the training of large language models.
73
- - **Acceleration:** Training was completed **2x faster** compared to traditional methods, thanks to Unsloth's optimizations.
74
- - **Reinforcement Learning:** Fine-tuning incorporated techniques from Huggingface's **TRL** library, enabling advanced instruction-tuning and alignment with human preferences.
75
-
76
- ---
77
-
78
- ## **Intended Use**
79
-
80
- ### **Primary Use Case**
81
- This model is designed for **text generation tasks**, including but not limited to:
82
- - Instruction-following
83
- - Question answering
84
- - Content creation
85
- - Dialogue systems
86
-
87
- ### **Limitations**
88
- - The model is trained primarily on English data and may not perform as well on other languages.
89
- - While fine-tuned for instruction-following, outputs should be reviewed for accuracy and relevance in critical applications.
90
-
91
- ---
92
-
93
- ## **How to Use**
94
-
95
- ### **Installation**
96
- To use this model, ensure you have the following libraries installed:
97
- ```bash
98
- pip install transformers torch bitsandbytes unsloth trl
99
- ```
100
-
101
-
102
-
103
-
104
- ```python
105
- from transformers import AutoModelForCausalLM, AutoTokenizer
106
-
107
- # Load the tokenizer and model
108
- model_name = "Daemontatox/Cogito-Maximus"
109
- tokenizer = AutoTokenizer.from_pretrained(model_name)
110
- model = AutoModelForCausalLM.from_pretrained(model_name, device_map="auto", load_in_4bit=True)
111
-
112
- # Generate text
113
- input_text = "Explain the concept of machine learning in simple terms."
114
- inputs = tokenizer(input_text, return_tensors="pt").to("cuda")
115
- outputs = model.generate(**inputs, max_length=100)
116
-
117
- # Decode and print the output
118
- print(tokenizer.decode(outputs[0], skip_special_tokens=True))
119
-
120
- ```
121
-
122
- ```
123
- @misc{daemontatox_cogito_maximus,
124
- author = {Daemontatox},
125
- title = {Cogito-Maximus: Fine-tuned Qwen2.5-72B Instruct Model},
126
- year = {2025},
127
- publisher = {Hugging Face},
128
- journal = {Hugging Face Model Repository},
129
- howpublished = {\url{https://huggingface.co/Daemontatox/Cogito-Maximus}}
130
- }
131
  ```
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - zho
5
+ - eng
6
+ - fra
7
+ - spa
8
+ - por
9
+ - deu
10
+ - ita
11
+ - rus
12
+ - jpn
13
+ - kor
14
+ - vie
15
+ - tha
16
+ - ara
17
+ base_model:
18
+ - Qwen/Qwen2.5-72B-Instruct
19
+ pipeline_tag: text-generation
20
+ library_name: transformers
21
+ tags:
22
+ - reasoning
23
+ - logic
24
+ - cot
25
+ - text-generation-inference
26
+ new_version: Daemontatox/Cogito-Maximus
27
+ ---
28
+
29
+ ![image](./image.webp)
30
+
31
+
32
+ ## **Model Overview**
33
+
34
+ This model, **Cogito-Maximus**, is a fine-tuned version of the `unsloth/qwen2.5-72b-instruct` base model, optimized for advanced text generation tasks. It leverages the power of **Unsloth** and **Huggingface's TRL (Transformer Reinforcement Learning)** library to achieve faster training and improved performance.
35
+
36
+ ### **Key Features**
37
+ - **Base Model:** `unsloth/qwen2.5-72b-instruct`
38
+ - **Training Acceleration:** Trained 2x faster using [Unsloth](https://github.com/unslothai/unsloth).
39
+ - **Fine-Tuning Framework:** Utilizes Huggingface's [TRL](https://github.com/huggingface/trl) library.
40
+ - **Optimized for Inference:** Ready for deployment in text-generation tasks with efficient inference capabilities.
41
+ - **License:** Apache-2.0
42
+
43
+ ---
44
+
45
+ ## **Model Details**
46
+
47
+ ### **Developed by**
48
+ - **Author:** Daemontatox
49
+ - **Organization:** Independent Contributor
50
+
51
+ ### **Tags**
52
+ - Text Generation Inference
53
+ - Transformers
54
+ - Unsloth
55
+ - Qwen2
56
+ - TRL
57
+
58
+ ### **Language**
59
+ - English (`en`)
60
+
61
+ ### **License**
62
+ This model is released under the **Apache-2.0 License**, which allows for free use, modification, and distribution, provided the original license and copyright notice are included.
63
+
64
+ ---
65
+
66
+ ## **Model Training**
67
+
68
+ ### **Base Model**
69
+ The model is derived from the `unsloth/qwen2.5-72b-instruct`, a version of the Qwen2.5-72B instruction-tuned model. The base model is optimized for efficiency using **bitsandbytes (bnb)** 4-bit quantization.
70
+
71
+ ### **Training Process**
72
+ - **Framework:** The model was fine-tuned using **Unsloth**, a library designed to accelerate the training of large language models.
73
+ - **Acceleration:** Training was completed **2x faster** compared to traditional methods, thanks to Unsloth's optimizations.
74
+ - **Reinforcement Learning:** Fine-tuning incorporated techniques from Huggingface's **TRL** library, enabling advanced instruction-tuning and alignment with human preferences.
75
+
76
+ ---
77
+
78
+ ## **Intended Use**
79
+
80
+ ### **Primary Use Case**
81
+ This model is designed for **text generation tasks**, including but not limited to:
82
+ - Instruction-following
83
+ - Question answering
84
+ - Content creation
85
+ - Dialogue systems
86
+
87
+ ### **Limitations**
88
+ - The model is trained primarily on English data and may not perform as well on other languages.
89
+ - While fine-tuned for instruction-following, outputs should be reviewed for accuracy and relevance in critical applications.
90
+
91
+ ---
92
+
93
+ ## **How to Use**
94
+
95
+ ### **Installation**
96
+ To use this model, ensure you have the following libraries installed:
97
+ ```bash
98
+ pip install transformers torch bitsandbytes unsloth trl
99
+ ```
100
+
101
+
102
+
103
+
104
+ ```python
105
+ from transformers import AutoModelForCausalLM, AutoTokenizer
106
+
107
+ # Load the tokenizer and model
108
+ model_name = "Daemontatox/Cogito-Maximus"
109
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
110
+ model = AutoModelForCausalLM.from_pretrained(model_name, device_map="auto", load_in_4bit=True)
111
+
112
+ # Generate text
113
+ input_text = "Explain the concept of machine learning in simple terms."
114
+ inputs = tokenizer(input_text, return_tensors="pt").to("cuda")
115
+ outputs = model.generate(**inputs, max_length=100)
116
+
117
+ # Decode and print the output
118
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
119
+
120
+ ```
121
+
122
+ ```
123
+ @misc{daemontatox_cogito_maximus,
124
+ author = {Daemontatox},
125
+ title = {Cogito-Maximus: Fine-tuned Qwen2.5-72B Instruct Model},
126
+ year = {2025},
127
+ publisher = {Hugging Face},
128
+ journal = {Hugging Face Model Repository},
129
+ howpublished = {\url{https://huggingface.co/Daemontatox/Cogito-Maximus}}
130
+ }
131
  ```