ArapCheruiyot commited on
Commit
b51fb6c
·
verified ·
1 Parent(s): 06288d1

Upload folder using huggingface_hub

Browse files
0000100_adapters.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0c95d627230320d6f8ad3ca7817da213612a7762c6f3691f810e79aa6601a642
3
+ size 1705667
0000200_adapters.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:71074ffe54c28f4a3da2b439f2de853b919ff428459876696ca50e1acfc50f26
3
+ size 1705667
0000300_adapters.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7cf25fc0e9cec709a04b30590ddaade6ecb2743d3a6b5cee05c18a6818214216
3
+ size 1705667
0000400_adapters.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:48896f692a1253503a1b20a995994183fbb2b22e3c585c4808a0b0772de7d9a6
3
+ size 1705667
0000500_adapters.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c6ce7539a777711ddc60651fc59791463d6a2577a4e45d3985c0c77d6ebe03b1
3
+ size 1705667
0000600_adapters.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:44ded87f48a3db27464e81acd64564e997d6104ede93c5c5f3652ec0237146c3
3
+ size 1705667
adapter_config.json ADDED
@@ -0,0 +1,55 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "adapter_path": "adapters/disarm_ew_llama3_lora_memory_optimized",
3
+ "batch_size": 1,
4
+ "config": "scripts/mlx_finetune_config_memory_optimized.yaml",
5
+ "data": "Data/training",
6
+ "eval_interval": 50,
7
+ "eval_iters": 10,
8
+ "fine_tune_type": "lora",
9
+ "grad_checkpoint": false,
10
+ "gradient_accumulation_steps": 4,
11
+ "iters": 600,
12
+ "learning_rate": 0.0003,
13
+ "log_interval": 10,
14
+ "lora_alpha": 16,
15
+ "lora_dropout": 0.0,
16
+ "lora_parameters": {
17
+ "rank": 8,
18
+ "dropout": 0.0,
19
+ "scale": 20.0
20
+ },
21
+ "lora_rank": 16,
22
+ "lr_schedule": null,
23
+ "lr_scheduler": "cosine",
24
+ "mask_prompt": false,
25
+ "max_seq_length": 2048,
26
+ "model": "ArapCheruiyot/disarm_ew-llama3",
27
+ "num_layers": 4,
28
+ "optimizer": "adam",
29
+ "optimizer_config": {
30
+ "adam": {},
31
+ "adamw": {},
32
+ "muon": {},
33
+ "sgd": {},
34
+ "adafactor": {}
35
+ },
36
+ "output_dir": "outputs/mlx_finetune_memory_optimized",
37
+ "quantize": false,
38
+ "resume_adapter_file": null,
39
+ "save_every": 100,
40
+ "save_interval": 100,
41
+ "seed": 0,
42
+ "steps_per_eval": 200,
43
+ "steps_per_report": 10,
44
+ "test": false,
45
+ "test_batches": 500,
46
+ "train": true,
47
+ "use_metal": true,
48
+ "use_metal_float16": true,
49
+ "use_wired_memory": true,
50
+ "val_batches": 25,
51
+ "wandb": null,
52
+ "warmup_steps": 50,
53
+ "weight_decay": 0.01,
54
+ "wired_memory_limit_mb": 16384
55
+ }
adapters.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:44ded87f48a3db27464e81acd64564e997d6104ede93c5c5f3652ec0237146c3
3
+ size 1705667