Text Generation
English
peft - PEFT 0.5.0
Nekochu commited on
Commit
0635e68
·
1 Parent(s): ed019df

Add Nekochu/Luminia-13B-v1-QLora checkpoint

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ trainer_state.json filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -1,3 +1,107 @@
1
  ---
 
 
2
  license: mit
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ library_name: peft - PEFT 0.5.0
3
+ base_model: Nekochu/Luminia-13B-v1-Lora
4
  license: mit
5
+ inference: false
6
+ model-index:
7
+ - name: Luminia-13B-v1-Lora
8
+ results: []
9
+ language:
10
+ - en
11
+ datasets:
12
+ - Anthropic/hh-rlhf
13
+ - OpenAssistant/oasst1
14
+ - wangrongsheng/comparison_gpt4_data_en
15
+ model_creator: Hugging Face H4
16
+ model_name: Luminia 13B v1 Lora
17
+ model_type: llama2
18
+ prompt_template: >-
19
+ Below is an instruction that describes a task. Write a response that
20
+ appropriately completes the request. ### Instruction: {prompt} ### Response:
21
+ pipeline_tag: text-generation
22
  ---
23
+
24
+ ## Model Details
25
+
26
+ QLoRA Reinforcement Learning (DPO) on Llama-2-13b-chat.
27
+
28
+ ### Model Description
29
+
30
+ - **Train by:** [Nekochu]
31
+ - **Model type:** [Llama]
32
+ - **Finetuned from model [Llama2](https://huggingface.co/meta-llama/Llama-2-13b-chat-hf)**
33
+
34
+
35
+ ## How to Get Started with the Model
36
+
37
+ <details>
38
+ <summary>Output example tested In <i>text-generation-webui</i></summary>
39
+
40
+ | Input | base llama-2-chat | QLoRa |
41
+ |:---------:|:-------:|:---------:|
42
+ | Name 3 applications for artificial intelligence | Sure! Here are three potential applications of artificial intelligence (AI):<br><br>- Healthcare: AI can be used to analyze medical records and identify patterns that could lead to better diagnoses or treatments for diseases. Additionally, AI-powered robots can assist in surgeries and other medical procedures, allowing doctors to perform complex tasks with greater precision and accuracy.<br> - Finance: AI can be used to detect fraud and prevent financial losses by analyzing large amounts of data quickly and accurately. Additionally, AI-powered chatbots can help customers with their financial queries, freeing up human representatives to focus on more complex issues.<br> Education: AI can be used to personalize learning experiences for students, adapting to their individual needs and abilities. Additionally, AI-powered tools can help teachers with grading and providing feedback, reducing their workload and allowing them to focus on more important aspects of | Certainly! In terms of applications, there are countless possibilities, but here are three examples that come to mind: <br><br> 1) Virtual Assistants - These virtual assistants use natural language processing (NLP) capabilities to understand voice commands from users, perform tasks such as scheduling appointments or sending emails, and even make decisions based on user preferences.<br> 2) Predictive Maintenance - Artificial intelligence algorithms can analyze data from sensors in industrial machines and predict when maintenance will be required, reducing downtime and increasing productivity.<br> 3) Fraud Detection - Machine learning algorithms can analyze patterns in customer transactions to identify fraudulent activity, allowing businesses to take action before financial loss occurs. |
43
+
44
+ </details>
45
+
46
+ ## Prompt template: Alpaca
47
+
48
+ ```
49
+ Below is an instruction that describes a task. Write a response that appropriately completes the request.
50
+ ### Instruction:
51
+ {prompt}
52
+ ### Response:
53
+ ```
54
+
55
+ ## Training Details
56
+
57
+ main branch is 200k step. 3 epoch for a total of [400k step](https://huggingface.co/Nekochu/Luminia-13B-v1-QLora/tree/400k) in 10 days
58
+
59
+ <details>
60
+ <summary>Click to see details</summary>
61
+
62
+ ## Training and evaluation data
63
+
64
+ <div style="width: auto; margin-left: auto; margin-right: auto">
65
+ <img src="https://i.imgur.com/25r7urB.jpg" alt="Nekochu" style="width: 100%; min-width: 400px; display: block; margin: auto;">
66
+ </div>
67
+
68
+ - [Luminia 13B v1 evaluation reward ](https://huggingface.co/Nekochu/Luminia-13B-v1-QLora/blob/main/trainer_state.json)
69
+
70
+ ## Training procedure
71
+
72
+ <details open> <summary>Datasets</summary>
73
+
74
+ - HH-RLHF (en)
75
+ - Open Assistant (multilingual)
76
+ - GPT-4 Generated Data (en&zh)
77
+
78
+ </details>
79
+
80
+ ### Training hyperparameters
81
+
82
+ The following hyperparameters were used during training:
83
+
84
+ - finetuning_type: lora
85
+ - quantization_bit: 4
86
+ - stage: dpo
87
+ - learning_rate: 5e-05
88
+ - cutoff_len: 4096
89
+ - num_train_epochs: 3.0
90
+ - max_samples: 100000
91
+ - warmup_steps: 0
92
+ - train_batch_size: 1
93
+ - distributed_type: single-GPU
94
+ - num_devices: 1
95
+ - warmup_steps: 0
96
+ - rope_scaling: linear
97
+ - lora_rank: 32
98
+ - lora_dropout: 0.15
99
+ - dpo_beta: 0.1
100
+ - bnb_4bit_compute_dtype: bfloat16
101
+
102
+ </details>
103
+
104
+ ### Trainer
105
+
106
+ - hiyouga/LLaMA-Efficient-Tuning
107
+
adapter_config.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "auto_mapping": null,
3
+ "base_model_name_or_path": "meta-llama/Llama-2-13b-chat-hf",
4
+ "bias": "none",
5
+ "fan_in_fan_out": false,
6
+ "inference_mode": true,
7
+ "init_lora_weights": true,
8
+ "layers_pattern": null,
9
+ "layers_to_transform": null,
10
+ "lora_alpha": 32.0,
11
+ "lora_dropout": 0.15,
12
+ "modules_to_save": null,
13
+ "peft_type": "LORA",
14
+ "r": 32,
15
+ "revision": null,
16
+ "target_modules": [
17
+ "k_proj",
18
+ "up_proj",
19
+ "gate_proj",
20
+ "v_proj",
21
+ "q_proj",
22
+ "down_proj",
23
+ "o_proj"
24
+ ],
25
+ "task_type": "CAUSAL_LM"
26
+ }
adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3e6dff9c26c841f96a63b2f3260f72e96ff61c98d33c0adf427b02035258e2b2
3
+ size 500897546
added_tokens.json ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ {
2
+ "</s>": 2,
3
+ "<s>": 1,
4
+ "<unk>": 0
5
+ }
optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:46f25405f4d478441abecbb7ada36682ea1fbe6fcc20016ac239084cc5de2c02
3
+ size 1001863522
rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:386cc2ae50e62ccea540a0c47b75c8f07815c569624eaf470059b05ac63967eb
3
+ size 14180
scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:abc23a226f0044297ff380d52ca683b9963301286f4189a8cd75a60fb78ffada
3
+ size 1064
special_tokens_map.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "<s>",
3
+ "eos_token": "</s>",
4
+ "pad_token": "</s>",
5
+ "unk_token": "<unk>"
6
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
+ size 499723
tokenizer_config.json ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "<unk>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "<s>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "</s>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ }
27
+ },
28
+ "additional_special_tokens": [],
29
+ "bos_token": "<s>",
30
+ "clean_up_tokenization_spaces": false,
31
+ "eos_token": "</s>",
32
+ "legacy": false,
33
+ "model_max_length": 1000000000000000019884624838656,
34
+ "pad_token": "</s>",
35
+ "padding_side": "right",
36
+ "sp_model_kwargs": {},
37
+ "split_special_tokens": false,
38
+ "tokenizer_class": "LlamaTokenizer",
39
+ "unk_token": "<unk>",
40
+ "use_default_system_prompt": true
41
+ }
trainer_state.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:59dd4f693ca17a4aed6786b564a5c21b8ebf2d7e62c99b2cbbfba9a41df96f73
3
+ size 19627724
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1e66c7937b77462faf820b6adf211726e4b8d266732ae948541260ab50ea8b97
3
+ size 4792