sheenrooff commited on
Commit
794515d
·
verified ·
1 Parent(s): efbd50d

Add files using upload-large-folder tool

Browse files
0000100_adapters.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:07409ba9e24d54c7d7f16476d18ab95bde5b4d04fc67f5cecc23537cd9367309
3
+ size 3414709
0000200_adapters.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9ac6170806a16f1b462770bd505d53bb20eadceeb874f0e5a8d960767e8d3611
3
+ size 3414709
0000300_adapters.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb9be67edaf87c4fd1f6de2f09982d414fef7a4fa2adbf2403c2fd193b4b098a
3
+ size 3414709
0000400_adapters.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eff11824874578eeb01953846922604d9d29c0ab5a8b25a130f63145e2bf36fb
3
+ size 360735053
0000500_adapters.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:29a53a6a0fed7924f765cb53f041a1f76ed127ffab8b825ba4488d0716e57222
3
+ size 360735053
0000600_adapters.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dffce5ea986836057cf68b990828147c621c9779186c3a9fc7b470f3defe0183
3
+ size 3414709
0000700_adapters.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e56d5f220487b460b7215fa43c786fc674dec1b6f7f14e35ff92ddf020f1e958
3
+ size 3414709
0000800_adapters.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b556dac246d93d442b722ebd785d396caf981696489d7a1672b84127bf96b619
3
+ size 360735053
0001200_adapters.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:131c6f6c7c82ce9901affc1f2894ee782f478c62bba639a45655f75e611f2704
3
+ size 360735053
0001600_adapters.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9afb1aa74eb64af05e652d30c713a926cabc403a5f9ffaeabec8cbc2fefaef51
3
+ size 360735053
0002000_adapters.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cb253fc143c44b5ea3df7268ab4bf2fe7981103d1eb4bb5a14b370c3ff221970
3
+ size 360735053
adapter_config.json ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "adapter_path": "adapters",
3
+ "batch_size": 4,
4
+ "config": "config.yaml",
5
+ "data": "data",
6
+ "fine_tune_type": "lora",
7
+ "grad_checkpoint": false,
8
+ "iters": 2079,
9
+ "learning_rate": 1e-06,
10
+ "lora_layers": 16,
11
+ "lora_parameters": {
12
+ "keys": [
13
+ "mlp.gate_proj",
14
+ "mlp.down_proj",
15
+ "self_attn.q_proj",
16
+ "mlp.up_proj",
17
+ "self_attn.o_proj",
18
+ "self_attn.v_proj",
19
+ "self_attn.k_proj"
20
+ ],
21
+ "rank": 128,
22
+ "alpha": 256,
23
+ "scale": 10.0,
24
+ "dropout": 0.05
25
+ },
26
+ "lr_schedule": null,
27
+ "max_seq_length": 2549,
28
+ "model": "mlx-community/Llama-3.2-1B-Instruct-4bit",
29
+ "num_layers": 16,
30
+ "resume_adapter_file": null,
31
+ "save_every": 400,
32
+ "seed": 0,
33
+ "steps_per_eval": 200,
34
+ "steps_per_report": 10,
35
+ "test": false,
36
+ "test_batches": -1,
37
+ "train": true,
38
+ "val_batches": -1
39
+ }
adapters.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d423cfb83a03162821355c441e31039d9f9b4b4fd6c95c35ac67de4926088dec
3
+ size 360735053