Vicuna13B V1.1 8bit 128g by Thireus

 ยป  All LLMs  ยป  Thireus  ยป  Vicuna13B V1.1 8bit 128g   URL Share it on

  Arxiv:2105.03536   Arxiv:2210.17323   Arxiv:2212.09720   Arxiv:2301.00774   8bit   Autotrain compatible   Endpoints compatible   Llama   Pytorch   Quantized   Region:us   Vicuna

Vicuna13B V1.1 8bit 128g Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Vicuna13B V1.1 8bit 128g (Thireus/Vicuna13B-v1.1-8bit-128g)
๐ŸŒŸ Advertise your project ๐Ÿš€

Vicuna13B V1.1 8bit 128g Parameters and Internals

Model Type 
open-source chatbot, auto-regressive language model, transformer architecture
Use Cases 
Applications:
large language model research, chatbots
Primary Use Cases:
research, hobbyist use in NLP, ML, AI
Additional Notes 
16GB VRAM preferred for optimal performance, with Triton preferred over Cuda.
Training Details 
Data Sources:
70K conversations from ShareGPT.com
Methodology:
fine-tuning LLaMa on user-shared conversations
Training Time:
between March 2023 and April 2023
Model Architecture:
based on the transformer architecture
Release Notes 
Version:
v1.1
Notes:
Refactor of tokenization and separator for improved compatibility and generation criteria.
LLM NameVicuna13B V1.1 8bit 128g
Repository ๐Ÿค—https://huggingface.co/Thireus/Vicuna13B-v1.1-8bit-128g 
Required VRAM13.6 GB
Updated2025-08-18
MaintainerThireus
Model Typellama
Model Files  13.6 GB
Quantization Type8bit
Model ArchitectureLlamaForCausalLM
Licenseother
Context Length2048
Model Max Length2048
Transformers Version4.29.0.dev0
Tokenizer ClassLlamaTokenizer
Beginning of Sentence Token<s>
End of Sentence Token</s>
Unk Token<unk>
Vocabulary Size32000
Torch Data Typefloat16

Best Alternatives to Vicuna13B V1.1 8bit 128g

Best Alternatives
Context / RAM
Downloads
Likes
... Text Chat 512K 5.0bpw H6 EXL2512K / 4.4 GB31
... Text Chat 512K 6.0bpw H6 EXL2512K / 5.2 GB31
Slimorca Phi 3.5128K / 7.6 GB50
Phi 3.5 Instruct Vul128K / 7.6 GB50
... Text Chat 128K 6.0bpw H6 EXL2128K / 5.2 GB51
...ocalAI Functioncall Phi 4 V0.316K / 29.4 GB58
PARM V2 Phi 4 4K CoT PyTorch16K / 29.4 GB01
QuantumQuill16K / 2.2 GB50
CodeMate V0.1 6.0bpw H6 EXL216K / 25.7 GB81
...oros C34b 3.1.2 8.0bpw H6 EXL216K / 34 GB81
Note: green Score (e.g. "73.2") means that the model is better than Thireus/Vicuna13B-v1.1-8bit-128g.

Rank the Vicuna13B V1.1 8bit 128g Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 50738 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241124