CataLlama V0.2 Instruct SFT DPO Merged is an open-source language model by catallama. Features: 8b LLM, VRAM: 16.1GB, Context: 8K, License: llama3, Instruction-Based, Merged, LLM Explorer Score: 0.14.
CataLlama V0.2 Instruct SFT DPO Merged Parameters and Internals
Model Type
text-generation
Use Cases
Areas:
commercial, research
Applications:
assistant-like chat, natural language generation tasks
Primary Use Cases:
Information extraction suitable for RAG, Named Entity Recognition (NER), Translation between English and Catalan, Summarization, Sentiment analysis, Chat
Limitations:
Not intended to beat benchmarks, Focus on demonstrating techniques for augmenting LLMs for new languages
Considerations:
Developers may fine-tune Llama 3 models for languages beyond English provided they comply with the Llama 3 Community License and the Acceptable Use Policy.
Additional Notes
The model focuses on supporting Catalan and preserving rare languages.
Supported Languages
ca (fluent), en (fluent)
Training Details
Data Sources:
Catalan-DPO-V2, Catalan-Instruct-V2
Methodology:
Supervised Fine-Tuning (SFT) and Direct Preference Optimization (DPO)
Model Architecture:
Auto-regressive language model with an optimized transformer architecture.
Input Output
Input Format:
follows the same prompt template as Llama-3 Instruct
Accepted Modalities:
text
Output Format:
text generation
Performance Tips:
Consider using torch.bfloat16 for better performance with Transformers library.
Note: green Score (e.g. "73.2") means that the model is better than catallama/CataLlama-v0.2-Instruct-SFT-DPO-Merged.
Rank the CataLlama V0.2 Instruct SFT DPO Merged Capabilities
๐ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐
Instruction Following and Task Automation
Factuality and Completeness of Knowledge
Censorship and Alignment
Data Analysis and Insight Generation
Text Generation
Text Summarization and Feature Extraction
Code Generation
Multi-Language Support and Translation
What open-source LLMs or SLMs are you in search of? 52721 in total.