AGI Math Model 🧠

PARAMANU-GANITA 🧠 is our first step towards AGI.

A language model with extremely strong mathematical reasoning abilities.

We present PARAMANU-GANITA, a 208 million parameter novel Auto Regressive (AR) decoder based language model on mathematics. The model is pretrained from scratch at context size of 4096 on our curated mixed mathematical corpus.

PARAMANU-GANITA despite being 35 times smaller than 7B LLMs, outperformed generalist LLMs such as LLaMa-1 7B by 28.4% points, LLaMa-2 7B by 27.6% points, Falcon 7B by 32.6% points, PaLM 8B by 35.3% points, and math specialised LLMs such as Minerva 8B by 23.2% points, and LLEMMA-7B by 3.0% points in GSM8k test accuracy metric respectively.

PARAMANU-GANITA also outperformed giant LLMs like PaLM 62B by 6.4% points, Falcon 40B by 19.8% points, LLaMa-1 33B by 3.8% points, and Vicuna 13B by 11.8% points respectively.

ModelParametersGSM8k Pass@1
LLaMa-133B35.6
LLaMa-17B11.0
LLaMa-213B28.7
LLaMa-27B11.8
Code LLaMa7B10.5
Code LLaMa34B29.6
Falcon40B19.6
Falcon7B6.8
MPT30B15.2
MPT7B6.8
GPT-J6B34.9
Vicuna13B27.6
PaLM8B4.1
PaLM62B33.0
Minerva8B16.2
LLEMMA7B36.4
PARAMANU-GANITA208M39.4

https://goatstack.ai/topics/paramanu-ganita-a-mathematical-marvel-nqvgdy

https://arxiv.org/abs/2404.14395

German Engineering with Indian heart ❤️