menu
techminis

A naukri.com initiative

google-web-stories
Home

>

ML News

>

7B Fully O...
source image

Arxiv

6d

read

12

img
dot

Image Credit: Arxiv

7B Fully Open Source Moxin-LLM -- From Pretraining to GRPO-based Reinforcement Learning Enhancement

  • Moxin-LLM is a fully open-source Large Language Model (LLM) adhering to principles of open science, open source, open data, and open access.
  • Moxin-LLM aims to address transparency concerns by releasing pre-training code, configurations, datasets, and checkpoints, allowing further innovations on LLMs.
  • Moxin-LLM goes through several finetuning stages to enhance reasoning capability, utilizing a post-training framework, instruction data, and Group Relative Policy Optimization (GRPO).
  • Experiments show that Moxin-LLM achieves superior performance in zero-shot evaluation, few-shot evaluation, and Chain-of-Thought (CoT) evaluation.

Read Full Article

like

Like

For uninterrupted reading, download the app