louisguthmann commited on
Commit
283e3dd
·
verified ·
1 Parent(s): e897220

Update 2B adapter with repair-focused run

Browse files
README.md CHANGED
@@ -1,41 +1,62 @@
1
  ---
2
  base_model: Qwen/Qwen3.5-2B
3
  library_name: peft
 
4
  tags:
 
5
  - lora
6
- - qwen
7
- - bash
8
- - shell
9
- - linux
10
- - text-generation
11
  ---
12
 
13
- # Qwen3.5-2B ShellCommand-Linux LoRA
14
 
15
- This repository contains a PEFT LoRA adapter trained for Linux natural-language-to-shell translation.
 
16
 
17
- ## Artifact Type
18
 
19
- This is a LoRA adapter, not a merged full model checkpoint.
 
20
 
21
- ## Intended Behavior
 
 
 
 
22
 
23
- The model is tuned to return exactly one of:
24
 
25
- - a Bash command or short Bash snippet
26
- - `ASK: <one short clarifying question>`
27
- - `CANNOT: <brief reason>`
28
 
29
- ## Eval Snapshot
30
 
31
- - score: `218.4816`
32
- - verifier ok rate: `61.67%`
33
- - verifier command rate: `59.38%`
34
- - verifier ask rate: `56.25%`
35
- - verifier cannot rate: `100.00%`
36
- - exact any-exact rate: `24.00%`
37
- - exact parse-ok rate: `98.00%`
38
 
39
- ## Usage
40
 
41
- Load this adapter on top of `Qwen/Qwen3.5-2B` with PEFT.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  base_model: Qwen/Qwen3.5-2B
3
  library_name: peft
4
+ model_name: model
5
  tags:
6
+ - base_model:adapter:Qwen/Qwen3.5-2B
7
  - lora
8
+ - sft
9
+ - transformers
10
+ - trl
11
+ licence: license
12
+ pipeline_tag: text-generation
13
  ---
14
 
15
+ # Model Card for model
16
 
17
+ This model is a fine-tuned version of [Qwen/Qwen3.5-2B](https://huggingface.co/Qwen/Qwen3.5-2B).
18
+ It has been trained using [TRL](https://github.com/huggingface/trl).
19
 
20
+ ## Quick start
21
 
22
+ ```python
23
+ from transformers import pipeline
24
 
25
+ question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
26
+ generator = pipeline("text-generation", model="None", device="cuda")
27
+ output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
28
+ print(output["generated_text"])
29
+ ```
30
 
31
+ ## Training procedure
32
 
33
+
 
 
34
 
 
35
 
 
 
 
 
 
 
 
36
 
37
+ This model was trained with SFT.
38
 
39
+ ### Framework versions
40
+
41
+ - PEFT 0.18.1
42
+ - TRL: 0.29.0
43
+ - Transformers: 5.3.0
44
+ - Pytorch: 2.10.0
45
+ - Datasets: 4.8.2
46
+ - Tokenizers: 0.22.2
47
+
48
+ ## Citations
49
+
50
+
51
+
52
+ Cite TRL as:
53
+
54
+ ```bibtex
55
+ @software{vonwerra2020trl,
56
+ title = {{TRL: Transformers Reinforcement Learning}},
57
+ author = {von Werra, Leandro and Belkada, Younes and Tunstall, Lewis and Beeching, Edward and Thrush, Tristan and Lambert, Nathan and Huang, Shengyi and Rasul, Kashif and Gallouédec, Quentin},
58
+ license = {Apache-2.0},
59
+ url = {https://github.com/huggingface/trl},
60
+ year = {2020}
61
+ }
62
+ ```
adapter_config.json CHANGED
@@ -29,13 +29,13 @@
29
  "rank_pattern": {},
30
  "revision": null,
31
  "target_modules": [
 
32
  "gate_proj",
33
- "o_proj",
34
- "q_proj",
35
  "v_proj",
36
- "up_proj",
37
  "down_proj",
38
- "k_proj"
 
 
39
  ],
40
  "target_parameters": null,
41
  "task_type": "CAUSAL_LM",
 
29
  "rank_pattern": {},
30
  "revision": null,
31
  "target_modules": [
32
+ "k_proj",
33
  "gate_proj",
 
 
34
  "v_proj",
 
35
  "down_proj",
36
+ "o_proj",
37
+ "up_proj",
38
+ "q_proj"
39
  ],
40
  "target_parameters": null,
41
  "task_type": "CAUSAL_LM",
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:312ea467f4f51a79d21cf2af71f40f8b5f3d75d3a8e5eceb88643a90a01b603b
3
  size 43672416
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:da96b3764db1705fad7f4ef8ebf236e0078833e3439c9ea42e37905bd501af44
3
  size 43672416
eval/exact_summary.json CHANGED
@@ -1,14 +1,14 @@
1
  {
2
- "model": "/root/bitnet-nl2sh/output/autoresearch_proxy_qwen35_2b/full_v1/qwen35_2b_batch8_ask_full_v1/model",
3
  "base_model": "Qwen/Qwen3.5-2B",
4
  "test_file": "/root/bitnet-nl2sh/output/data/nl2sh_test_raw.jsonl",
5
  "rows": 100,
6
- "primary_exact": 20,
7
  "alt_exact": 6,
8
- "any_exact": 24,
9
- "parse_ok": 98,
10
- "primary_exact_rate": 0.2,
11
- "any_exact_rate": 0.24,
12
- "parse_ok_rate": 0.98,
13
- "avg_gen_seconds_per_example": 0.5884
14
  }
 
1
  {
2
+ "model": "/root/bitnet-nl2sh/output/autoresearch_proxy_qwen35_2b/repair_full_v1/qwen35_2b_batch8_repair_full_v1/model",
3
  "base_model": "Qwen/Qwen3.5-2B",
4
  "test_file": "/root/bitnet-nl2sh/output/data/nl2sh_test_raw.jsonl",
5
  "rows": 100,
6
+ "primary_exact": 11,
7
  "alt_exact": 6,
8
+ "any_exact": 15,
9
+ "parse_ok": 97,
10
+ "primary_exact_rate": 0.11,
11
+ "any_exact_rate": 0.15,
12
+ "parse_ok_rate": 0.97,
13
+ "avg_gen_seconds_per_example": 0.5498
14
  }
eval/score.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "components": {
3
- "any_exact_rate": 0.24,
4
  "ask_rate": 0.5625,
5
  "cannot_rate": 1.0,
6
- "command_rate": 0.5938,
7
- "ok_rate": 0.6167,
8
- "parse_ok_rate": 0.98,
9
- "weak_category_rate": 0.5357142857142857
10
  },
11
- "exact_summary": "/root/bitnet-nl2sh/output/autoresearch_proxy_qwen35_2b/full_v1/qwen35_2b_batch8_ask_full_v1/eval_exact/summary.json",
12
- "score": 218.4816,
13
- "score_0_to_1": 2.184816,
14
- "verifier_summary": "/root/bitnet-nl2sh/output/autoresearch_proxy_qwen35_2b/full_v1/qwen35_2b_batch8_ask_full_v1/eval_verifier/summary.json",
15
  "weak_categories": [
16
  "ambiguous_delete",
17
  "ambiguous_secret",
@@ -25,8 +25,8 @@
25
  "ambiguous_delete": 0.125,
26
  "ambiguous_secret": 1.0,
27
  "create_archive": 1.0,
28
- "enabled_services": 0.875,
29
- "find_jpgs": 0.0,
30
  "json_query": 0.625,
31
  "top_ips": 0.125
32
  },
 
1
  {
2
  "components": {
3
+ "any_exact_rate": 0.15,
4
  "ask_rate": 0.5625,
5
  "cannot_rate": 1.0,
6
+ "command_rate": 0.6875,
7
+ "ok_rate": 0.6917,
8
+ "parse_ok_rate": 0.97,
9
+ "weak_category_rate": 0.6071428571428571
10
  },
11
+ "exact_summary": "/root/bitnet-nl2sh/output/autoresearch_proxy_qwen35_2b/repair_full_v1/qwen35_2b_batch8_repair_full_v1/eval_exact/summary.json",
12
+ "score": 232.7182,
13
+ "score_0_to_1": 2.327182,
14
+ "verifier_summary": "/root/bitnet-nl2sh/output/autoresearch_proxy_qwen35_2b/repair_full_v1/qwen35_2b_batch8_repair_full_v1/eval_verifier/summary.json",
15
  "weak_categories": [
16
  "ambiguous_delete",
17
  "ambiguous_secret",
 
25
  "ambiguous_delete": 0.125,
26
  "ambiguous_secret": 1.0,
27
  "create_archive": 1.0,
28
+ "enabled_services": 1.0,
29
+ "find_jpgs": 0.375,
30
  "json_query": 0.625,
31
  "top_ips": 0.125
32
  },
eval/verifier_summary.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "avg_gen_seconds_per_example": 0.6777,
3
  "base_model": "Qwen/Qwen3.5-2B",
4
  "category_breakdown": {
5
  "ambiguous_delete": {
@@ -18,8 +18,8 @@
18
  "rows": 8
19
  },
20
  "count_extension": {
21
- "ok": 2,
22
- "ok_rate": 0.25,
23
  "rows": 8
24
  },
25
  "create_archive": {
@@ -33,8 +33,8 @@
33
  "rows": 8
34
  },
35
  "enabled_services": {
36
- "ok": 7,
37
- "ok_rate": 0.875,
38
  "rows": 8
39
  },
40
  "extract_archive": {
@@ -43,18 +43,18 @@
43
  "rows": 8
44
  },
45
  "find_jpgs": {
46
- "ok": 0,
47
- "ok_rate": 0.0,
48
  "rows": 8
49
  },
50
  "git_branch": {
51
- "ok": 0,
52
- "ok_rate": 0.0,
53
  "rows": 8
54
  },
55
  "grep_literal": {
56
- "ok": 8,
57
- "ok_rate": 1.0,
58
  "rows": 8
59
  },
60
  "json_query": {
@@ -63,13 +63,13 @@
63
  "rows": 8
64
  },
65
  "replace_literal": {
66
- "ok": 2,
67
- "ok_rate": 0.25,
68
  "rows": 8
69
  },
70
  "show_env": {
71
- "ok": 8,
72
- "ok_rate": 1.0,
73
  "rows": 8
74
  },
75
  "top_ips": {
@@ -92,14 +92,14 @@
92
  "rows": 8
93
  },
94
  "command": {
95
- "ok": 57,
96
- "ok_rate": 0.5938,
97
  "rows": 96
98
  }
99
  },
100
- "model": "/root/bitnet-nl2sh/output/autoresearch_proxy_qwen35_2b/full_v1/qwen35_2b_batch8_ask_full_v1/model",
101
- "ok": 74,
102
- "ok_rate": 0.6167,
103
- "prompt_file": "/root/bitnet-nl2sh/prompts/student_linux_shell_v2f.txt",
104
  "rows": 120
105
  }
 
1
  {
2
+ "avg_gen_seconds_per_example": 0.576,
3
  "base_model": "Qwen/Qwen3.5-2B",
4
  "category_breakdown": {
5
  "ambiguous_delete": {
 
18
  "rows": 8
19
  },
20
  "count_extension": {
21
+ "ok": 0,
22
+ "ok_rate": 0.0,
23
  "rows": 8
24
  },
25
  "create_archive": {
 
33
  "rows": 8
34
  },
35
  "enabled_services": {
36
+ "ok": 8,
37
+ "ok_rate": 1.0,
38
  "rows": 8
39
  },
40
  "extract_archive": {
 
43
  "rows": 8
44
  },
45
  "find_jpgs": {
46
+ "ok": 3,
47
+ "ok_rate": 0.375,
48
  "rows": 8
49
  },
50
  "git_branch": {
51
+ "ok": 8,
52
+ "ok_rate": 1.0,
53
  "rows": 8
54
  },
55
  "grep_literal": {
56
+ "ok": 4,
57
+ "ok_rate": 0.5,
58
  "rows": 8
59
  },
60
  "json_query": {
 
63
  "rows": 8
64
  },
65
  "replace_literal": {
66
+ "ok": 8,
67
+ "ok_rate": 1.0,
68
  "rows": 8
69
  },
70
  "show_env": {
71
+ "ok": 5,
72
+ "ok_rate": 0.625,
73
  "rows": 8
74
  },
75
  "top_ips": {
 
92
  "rows": 8
93
  },
94
  "command": {
95
+ "ok": 66,
96
+ "ok_rate": 0.6875,
97
  "rows": 96
98
  }
99
  },
100
+ "model": "/root/bitnet-nl2sh/output/autoresearch_proxy_qwen35_2b/repair_full_v1/qwen35_2b_batch8_repair_full_v1/model",
101
+ "ok": 83,
102
+ "ok_rate": 0.6917,
103
+ "prompt_file": "/root/bitnet-nl2sh/prompts/student_linux_shell_v2g.txt",
104
  "rows": 120
105
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e5dcc81a623f926adf038f947eeeed6da9a08d33e8cfb9633aeb90d3a34ae660
3
  size 5713
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:90edf4dd4fe7735c98f12ebde98b91799442829e503d750532efa2316597ffa4
3
  size 5713