Magnum 12B V2.5 Kto is an open-source language model by anthracite-org. Features: 12b LLM, VRAM: 24.5GB, Context: 1000K, License: apache-2.0, LLM Explorer Score: 0.16.
KTO is an experimental release, part of a series of models. It's fine-tuned on top of Magnum-12b-v2. Experimental data was used for initial testing, with plans to scale up.
Supported Languages
en (supported), fr (supported), de (supported), es (supported), it (supported), pt (supported), ru (supported), zh (supported), ja (supported)
hybrid reinforcement learning strategy of KTO + DPOP using rejected data sampled from the original model as rejected and data from original finetuning dataset as chosen.
Note: green Score (e.g. "73.2") means that the model is better than anthracite-org/magnum-12b-v2.5-kto.
Rank the Magnum 12B V2.5 Kto Capabilities
๐ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐
Instruction Following and Task Automation
Factuality and Completeness of Knowledge
Censorship and Alignment
Data Analysis and Insight Generation
Text Generation
Text Summarization and Feature Extraction
Code Generation
Multi-Language Support and Translation
What open-source LLMs or SLMs are you in search of? 52721 in total.