Alfred 40B 1023 AWQ by TheBloke

 ยป  All LLMs  ยป  TheBloke  ยป  Alfred 40B 1023 AWQ   URL Share it on

  Arxiv:2306.15595   Arxiv:2307.03172   Arxiv:2309.00071   4-bit   Autotrain compatible   Awq Base model:lightonai/alfred-40... Base model:quantized:lightonai...   Custom code   Dataset:ehartford/dolphin   Dataset:openassistant/oasst1   Dataset:tau/sled Dataset:tiiuae/falcon-refinedw...   De   En   Es   Falcon   Falcon-40b   Fr   It   Long-context   Ntk-yarn   Quantized   Refinedweb   Region:us   Safetensors   Sharded   Tensorflow   Yarn

Alfred 40B 1023 AWQ Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Alfred 40B 1023 AWQ (TheBloke/alfred-40B-1023-AWQ)
๐ŸŒŸ Advertise your project ๐Ÿš€

Alfred 40B 1023 AWQ Parameters and Internals

Model Type 
causal decoder-only, text generation
Use Cases 
Areas:
chat, instruction following
Applications:
personal assistant, language tasks
Primary Use Cases:
text generation, conversational AI
Limitations:
Limited language support beyond primary languages, Bias due to training data
Considerations:
Users should consider ethical implications and add safety measures.
Additional Notes 
Model available as AWQ, GPTQ, and other quantized versions for diverse hardware
Supported Languages 
en (full), fr (full), de (full), es (full), it (limited), pt (limited), pl (limited), nl (limited), ro (limited), cs (limited), sv (limited)
Training Details 
Data Sources:
OpenAssistant/oasst1, ehartford/dolphin, tau/sled, tiiuae/falcon-refinedweb
Data Volume:
100 megatokens
Methodology:
Supervised finetuning and NTK-YaRN context length extension
Context Length:
8192
Hardware Used:
128 A100 40GB GPUs
Model Architecture:
Falcon
Responsible Ai Considerations 
Fairness:
Model carries stereotypes and biases from data, needs user guardrails for fair use.
Transparency:
Transparency about data and methods used provided.
Accountability:
Model outputs accountability lies with the user.
Mitigation Strategies:
Users need to implement precautions and guardrails.
Input Output 
Input Format:
Text prompt format with integrated chat tokens
Accepted Modalities:
text
Output Format:
Generated text responding to input context
Performance Tips:
Use recommended libraries (e.g., AutoAWQ) for best performance
Release Notes 
Version:
1023
Date:
October 2023
Notes:
Extended context length; includes NTK-YaRN method for context expansion.
LLM NameAlfred 40B 1023 AWQ
Repository ๐Ÿค—https://huggingface.co/TheBloke/alfred-40B-1023-AWQ 
Model NameAlfred 40B 1023
Model CreatorLightOn AI
Base Model(s)  Alfred 40B 1023   lightonai/alfred-40b-1023
Model Size40b
Required VRAM23.3 GB
Updated2025-08-21
MaintainerTheBloke
Model TypeRefinedWeb
Model Files  11.0 GB: 1-of-3   9.9 GB: 2-of-3   2.4 GB: 3-of-3
Supported Languagesen fr de es it
AWQ QuantizationYes
Quantization Typeawq
Model ArchitectureRWForCausalLM
Licenseapache-2.0
Model Max Length8192
Transformers Version4.35.0
Is Biased0
Tokenizer ClassPreTrainedTokenizerFast
Vocabulary Size65024
Torch Data Typefloat16

Best Alternatives to Alfred 40B 1023 AWQ

Best Alternatives
Context / RAM
Downloads
Likes
...alcon 40B Instruct W4 G128 AWQ0K / 22.3 GB142
Falcon 40B 8bit0K / 41.8 GB131
Falcon 40B Instruct 8bit0K / 41.8 GB136
Alfred 40B 10230K / 83.6 GB212048
Vulture 40B0K / 81.8 GB19268
Alfred 40B 1023 GPTQ0K / 22.5 GB143
FalconLite0K / 22.3 GB349170
FalconLite0K / 22.3 GB284170
Docsgpt 40B Falcon0K / 82.5 GB2813
Alfred 40B 07230K / 83.6 GB2446
Note: green Score (e.g. "73.2") means that the model is better than TheBloke/alfred-40B-1023-AWQ.

Rank the Alfred 40B 1023 AWQ Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 50804 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241124