Moxin-LLM is a fully open-source Large Language Model (LLM) adhering to principles of open science, open source, open data, and open access.
Moxin-LLM aims to address transparency concerns by releasing pre-training code, configurations, datasets, and checkpoints, allowing further innovations on LLMs.
Moxin-LLM goes through several finetuning stages to enhance reasoning capability, utilizing a post-training framework, instruction data, and Group Relative Policy Optimization (GRPO).
Experiments show that Moxin-LLM achieves superior performance in zero-shot evaluation, few-shot evaluation, and Chain-of-Thought (CoT) evaluation.