Additional Notes |
|
LLM Name | Falcon 180B Chat AWQ |
Repository ๐ค | https://huggingface.co/TheBloke/Falcon-180B-Chat-AWQ |
Base Model(s) | |
Model Size | 180b |
Required VRAM | 96 GB |
Updated | 2025-08-20 |
Maintainer | TheBloke |
Model Type | falcon |
Model Files | |
AWQ Quantization | Yes |
Quantization Type | awq |
Model Architecture | FalconForCausalLM |
Context Length | 2048 |
Model Max Length | 2048 |
Transformers Version | 4.33.2 |
Is Biased | 0 |
Tokenizer Class | PreTrainedTokenizerFast |
Vocabulary Size | 65024 |
Torch Data Type | float16 |
Best Alternatives |
Context / RAM |
Downloads |
Likes |
---|---|---|---|
Airoboros 180B 2.2.1 AWQ | 2K / 96 GB | 9 | 6 |
Largefalcon | 2K / 411.4 GB | 5 | 0 |
...buddy Falcon 180B V13 Preview0 | 2K / 358.1 GB | 1761 | 2 |
...buddy Falcon 180B V12 Preview0 | 2K / 358.1 GB | 1765 | 0 |
Airoboros 180B 2.2.1 | 2K / 154.2 GB | 1605 | 17 |
...buddy Falcon 180B V13 Preview2 | 2K / 358.1 GB | 7 | 1 |
...buddy Falcon 180B V13 Preview1 | 2K / 358.1 GB | 10 | 4 |
Falcon 180B Chat GPTQ | 2K / 94.1 GB | 11 | 69 |
...alcon 180B Omniquant W3a16g512 | 2K / 69.4 GB | 5 | 3 |
Falcon 180B | 0K / 193.8 GB | 2535 | 1147 |
๐ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐