allanctan-ai commited on
Commit
54fe1f9
·
verified ·
1 Parent(s): 0994222

allanctan-ai/llama3181b_filipino_dr

Browse files
README.md CHANGED
@@ -27,14 +27,14 @@ print(output["generated_text"])
27
 
28
  ## Training procedure
29
 
30
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/psi-ai/huggingface/runs/vzskzoon)
31
 
32
 
33
  This model was trained with SFT.
34
 
35
  ### Framework versions
36
 
37
- - TRL: 0.15.1
38
  - Transformers: 4.49.0
39
  - Pytorch: 2.5.1+cu124
40
  - Datasets: 3.3.2
 
27
 
28
  ## Training procedure
29
 
30
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/psi_/huggingface/runs/dajsfre0)
31
 
32
 
33
  This model was trained with SFT.
34
 
35
  ### Framework versions
36
 
37
+ - TRL: 0.15.2
38
  - Transformers: 4.49.0
39
  - Pytorch: 2.5.1+cu124
40
  - Datasets: 3.3.2
adapter_config.json CHANGED
@@ -23,13 +23,13 @@
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
- "k_proj",
 
27
  "up_proj",
28
- "v_proj",
29
  "o_proj",
30
- "q_proj",
31
  "down_proj",
32
- "gate_proj"
33
  ],
34
  "task_type": "CAUSAL_LM",
35
  "use_dora": false,
 
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
+ "q_proj",
27
+ "gate_proj",
28
  "up_proj",
29
+ "k_proj",
30
  "o_proj",
 
31
  "down_proj",
32
+ "v_proj"
33
  ],
34
  "task_type": "CAUSAL_LM",
35
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6a19ee80e93f4262b7ae9e66968c66150ef4acef6977bc12377769a0d4666183
3
  size 167832240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:30b5e6fc527f2f212f958a237b6a84bdea21be0bfa9f9b2e3bca9fb19ccc3582
3
  size 167832240
runs/Feb26_10-37-05_2c3374e56277/events.out.tfevents.1740567030.2c3374e56277.553.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba85938bec8704acf65a91d24ecb2e6740ab2af9303f5c44a6102943555a50e2
3
+ size 6202
runs/Feb26_11-11-44_2c3374e56277/events.out.tfevents.1740568313.2c3374e56277.553.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f8ab0c93bc5b009d9047fb7adba10d44d922208ac14df2e96d829e8c34e621f
3
+ size 50261
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8cf20a031b177e296e06aa796d73ddbe75bbea69214ee35d34f4812d56e49480
3
  size 5624
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e9a1ada197366a7315148810e2f3af64bd99e058e24fafb09d6ca4dab7a0a505
3
  size 5624