Alfred 40B 1023 by lightonai

 ยป  All LLMs  ยป  lightonai  ยป  Alfred 40B 1023   URL Share it on

  Arxiv:2306.15595   Arxiv:2307.03172   Arxiv:2309.00071   Autotrain compatible   Conversational   Custom code   Dataset:ehartford/dolphin   Dataset:openassistant/oasst1   Dataset:tau/sled Dataset:tiiuae/falcon-refinedw...   De   En   Endpoints compatible   Es   Falcon   Falcon-40b   Fr   It   Long-context   Ntk-yarn   Pytorch   Refinedweb   Region:us   Sharded   Yarn

Alfred 40B 1023 Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
Alfred 40B 1023 (lightonai/alfred-40b-1023)
๐ŸŒŸ Advertise your project ๐Ÿš€

Alfred 40B 1023 Parameters and Internals

Model Type 
Causal decoder-only
Use Cases 
Areas:
Chat, Instruct
Primary Use Cases:
Chat models, Instruct models
Limitations:
Limited language capabilities outside specified languages
Considerations:
Implemented NTK-YaRN for extended context capabilities.
Additional Notes 
Trained with 3D parallelism and ZeRO on AWS SageMaker.
Supported Languages 
English (High), German (High), Spanish (High), French (High), Italian (Limited), Portuguese (Limited), Polish (Limited), Dutch (Limited), Romanian (Limited), Czech (Limited), Swedish (Limited)
Training Details 
Data Sources:
OpenAssistant/oasst1, ehartford/dolphin, tau/sled, tiiuae/falcon-refinedweb, internal, internal-long-context
Data Volume:
100 megatokens
Methodology:
Supervised finetuning with a custom NTK-YaRN method for context length extension
Context Length:
8192
Hardware Used:
128 A100 40GB GPUs
Model Architecture:
Causal decoder-only
Input Output 
Input Format:
Prompts with integrated chat tokens for instruct and chat mode
Accepted Modalities:
Text
Output Format:
Generated text based on input queries
Performance Tips:
Ensure correct integration of chat tokens in prompts for optimal performance.
LLM NameAlfred 40B 1023
Repository ๐Ÿค—https://huggingface.co/lightonai/alfred-40b-1023 
Model Size40b
Required VRAM83.6 GB
Updated2025-08-21
Maintainerlightonai
Model TypeRefinedWeb
Model Files  9.5 GB: 1-of-9   9.5 GB: 2-of-9   9.5 GB: 3-of-9   9.5 GB: 4-of-9   9.5 GB: 5-of-9   9.5 GB: 6-of-9   9.5 GB: 7-of-9   9.5 GB: 8-of-9   7.6 GB: 9-of-9
Supported Languagesen fr de es it
Model ArchitectureRWForCausalLM
Licenseapache-2.0
Model Max Length8192
Transformers Version4.31.0
Is Biased0
Tokenizer ClassPreTrainedTokenizerFast
Vocabulary Size65024
Torch Data Typebfloat16

Quantized Models of the Alfred 40B 1023

Model
Likes
Downloads
VRAM
Alfred 40B 1023 GGUF528517 GB
Alfred 40B 1023 AWQ51323 GB
Alfred 40B 1023 GPTQ31422 GB

Best Alternatives to Alfred 40B 1023

Best Alternatives
Context / RAM
Downloads
Likes
Vulture 40B0K / 81.8 GB19268
Docsgpt 40B Falcon0K / 82.5 GB2813
Alfred 40B 07230K / 83.6 GB2446
Openbuddy Falcon 40B V9 Bf160K / 82.6 GB174
...alcon 40B Lora Sft Stage2 1.1K0K / 82.5 GB130
Falcon 40B0K / 83.6 GB161
...m Oasst1 En 2048 Falcon 40B V20K / 83.6 GB1418
Falcon 40B Sft Top1 5600K / 83.6 GB12350
Falcon 40B Sft Mix 12260K / 83.6 GB1938
...m Oasst1 En 2048 Falcon 40B V10K / 165 GB1731

Rank the Alfred 40B 1023 Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 50804 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241124