Marathi GPT Gemma 7B by l3cube-pune

 »  All LLMs  »  l3cube-pune  »  Marathi GPT Gemma 7B   URL Share it on

  Arxiv:2205.14728   Autotrain compatible   Endpoints compatible   Gemma   Mr   Region:us   Safetensors   Sharded   Tensorflow

Marathi GPT Gemma 7B Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").

Marathi GPT Gemma 7B Parameters and Internals

Model Type 
Gemma, Natural Language Processing
Use Cases 
Areas:
Natural Language Processing, Marathi Language Support
Applications:
Text Generation, Language Translation
Primary Use Cases:
Answering questions in Marathi
Additional Notes 
Part of the MahaNLP initiative.
Supported Languages 
mr (Marathi)
Training Details 
Data Sources:
translated Marathi datasets
Methodology:
LoRA fine-tuning
Model Architecture:
Gemma 7B
Input Output 
Input Format:
\n### Instruction:\nमहाराष्ट्राची राजधानी काय आहे?\n\n### Input:\n\n\n### Response:\nमहाराष्ट्राची राजधानी मुंबई आहे
Accepted Modalities:
text
Output Format:
text
LLM NameMarathi GPT Gemma 7B
Repository 🤗https://huggingface.co/l3cube-pune/marathi-gpt-gemma-7b 
Model Size7b
Required VRAM17.1 GB
Updated2025-09-15
Maintainerl3cube-pune
Model Typegemma
Model Files  5.0 GB: 1-of-4   5.0 GB: 2-of-4   5.0 GB: 3-of-4   2.1 GB: 4-of-4
Supported Languagesmr
Model ArchitectureGemmaForCausalLM
Licensecc-by-4.0
Context Length8192
Model Max Length8192
Transformers Version4.38.2
Tokenizer ClassGemmaTokenizer
Padding Token<pad>
Vocabulary Size256000
Torch Data Typebfloat16

Best Alternatives to Marathi GPT Gemma 7B

Best Alternatives
Context / RAM
Downloads
Likes
Kaggle Math Model Gemma V112K / 17.1 GB50
Gemma 1.1 7B It8K / 17.1 GB13881274
SeaLLM 7B V2.58K / 17.1 GB1257250
Zephyr 7B Gemma DPO Avg8K / 17.1 GB150
Zephyr 7B Gemma Rpo Avg8K / 17.1 GB60
Zephyr 7B Gemma V0.18K / 17.1 GB1229123
... Codegemma 2 7B It Alpaca V1.38K / 17.1 GB61
... 7B Finetuned Sft Navarasa 2.08K / 34 GB129422
Codegemma 7B It8K / 17.1 GB6076231
Codegemma 7B8K / 17.1 GB5577201
Note: green Score (e.g. "73.2") means that the model is better than l3cube-pune/marathi-gpt-gemma-7b.

Rank the Marathi GPT Gemma 7B Capabilities

🆘 Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! 🌟

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 51387 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241124