--- license: apache-2.0 --- # MarinbadGPT MarinbadGPT is a language model based on HuggingFaceTB's SmolLM-135M architecture, finely trained on a corpus of **Marinbad** games. The aim of this model is to generate games of Marinbad and play them against a human player. ## Model Training The training of MarinbadGPT was conducted on a high-performance computing infrastructure utilizing NVIDIA H100 GPUs, renowned for their power in the field of deep learning. **Training Configuration:** * **Infrastructure:** 2x NVIDIA H100 (80GB HBM3) * **Duration:** 1 hour * **Optimizer:** AdamW * **Learning Rate:** 3e-4 * **Batch Size:** Micro batch size of 128, with gradient accumulation steps of 8, resulting in an effective batch size of 1024 * **Warmup Steps:** 100