Yi 34B 200K DARE Merge V5 AWQ is an open-source language model by TheBloke. Features: 34b LLM, VRAM: 19.3GB, Context: 195K, License: other, Quantized, Merged, LLM Explorer Score: 0.11.
Yi 34B 200K DARE Merge V5 AWQ Parameters and Internals
Model Type
yi, text-generation
Additional Notes
Yi tends to run "hot" by default, and it really needs MinP to cull the huge vocabulary. 24GB GPUs can run Yi-34B-200K models at 45K-75K context with exllamav2.
Various densities were tested with perplexity tests and long context prompts. Relatively high densities seem to perform better, contrary to the findings of the Super Mario paper. This particular version is merged with more than the "recommended" max density of 0.5. It seems to result in even better perplexity, but I'm not sure if this translates to better output. Weights that add up to 1 seem to be optimal. Dare Ties is also resulting in seemingly better, lower perplexity merges than a regular ties merge, task arithmetic, or a slerp merge. SUS Chat is not a 200K model, hence it was merged at a very low density to try and preserve Yi 200K's long context performance while "keeping" some of SUS parameters.
Note: green Score (e.g. "73.2") means that the model is better than TheBloke/Yi-34B-200K-DARE-merge-v5-AWQ.
Rank the Yi 34B 200K DARE Merge V5 AWQ Capabilities
๐ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐
Instruction Following and Task Automation
Factuality and Completeness of Knowledge
Censorship and Alignment
Data Analysis and Insight Generation
Text Generation
Text Summarization and Feature Extraction
Code Generation
Multi-Language Support and Translation
What open-source LLMs or SLMs are you in search of? 52392 in total.