Metharme 13B 4bit GPTQ is an open-source language model by TehVenom. Features: 13b LLM, VRAM: 7.5GB, Context: 2K, Quantized, Instruction-Based, LLM Explorer Score: 0.07.
Not fine-tuned to be safe and harmless, May contain profanity or offensive text, Outputs might often be factually wrong or misleading
Additional Notes
The model always starts by generating a BOS token. This is an accidental side-effect which is planned to be addressed in future versions.
Training Details
Methodology:
Supervised fine-tuning over a mixture of regular instruction data alongside roleplay, fictional stories and conversations with synthetically generated instructions attached. Trained as a LoRA, then merged down to the base model.
Input Output
Input Format:
Uses <|system|>, <|user|>, and <|model|> tokens for conversation history.
Accepted Modalities:
text
Output Format:
Generates formatted text responses based on prompts
Performance Tips:
Best to do a sliding window over the user and model turns, but keep the system prompt fixed at the start of the context window.
Note: green Score (e.g. "73.2") means that the model is better than TehVenom/Metharme-13b-4bit-GPTQ.
Rank the Metharme 13B 4bit GPTQ Capabilities
๐ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐
Instruction Following and Task Automation
Factuality and Completeness of Knowledge
Censorship and Alignment
Data Analysis and Insight Generation
Text Generation
Text Summarization and Feature Extraction
Code Generation
Multi-Language Support and Translation
What open-source LLMs or SLMs are you in search of? 52509 in total.