File size: 775 Bytes
1f0614f
 
 
26bba0e
 
960455b
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
---
license: apache-2.0
---
# MarinbadGPT

MarinbadGPT is a language model based on HuggingFaceTB's SmolLM-135M architecture, finely trained on a corpus of **Marinbad** games. The aim of this model is to generate games of Marinbad and play them against a human player.

## Model Training

The training of MarinbadGPT was conducted on a high-performance computing infrastructure utilizing NVIDIA H100 GPUs, renowned for their power in the field of deep learning.

**Training Configuration:**

*   **Infrastructure:** 2x NVIDIA H100 (80GB HBM3)
*   **Duration:** 1 hour
*   **Optimizer:** AdamW
*   **Learning Rate:** 3e-4
*   **Batch Size:** Micro batch size of 128, with gradient accumulation steps of 8, resulting in an effective batch size of 1024
*   **Warmup Steps:** 100