Meta Llama 3 8B Alternate Tokenizer is an open-source language model by NousResearch. Features: 8b LLM, VRAM: 16.1GB, Context: 8K, License: other, LLM Explorer Score: 0.16.
Meta Llama 3 8B Alternate Tokenizer Parameters and Internals
Model Type
text-generation
Use Cases
Areas:
Commercial, Research
Applications:
Assistant-like chat, Natural language generation tasks
Primary Use Cases:
Commercial applications, Research
Limitations:
Out-of-scope use in languages other than English without fine-tuning
Considerations:
Developers are encouraged to fine-tune models for their specific use cases.
Additional Notes
Pretraining does not include Meta user data. Developers encouraged to share feedback via provided channels. Carbon emissions for training offset by Meta's sustainability program.
Supported Languages
English (Fully Supported)
Training Details
Data Sources:
Publicly available online data
Data Volume:
15 trillion tokens
Methodology:
supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF)
Context Length:
8000
Hardware Used:
Meta's Research SuperCluster (H100-80GB GPUs)
Model Architecture:
Optimized transformer architecture with Grouped-Query Attention (GQA)
Note: green Score (e.g. "73.2") means that the model is better than NousResearch/Meta-Llama-3-8B-Alternate-Tokenizer.
Rank the Meta Llama 3 8B Alternate Tokenizer Capabilities
๐ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐
Instruction Following and Task Automation
Factuality and Completeness of Knowledge
Censorship and Alignment
Data Analysis and Insight Generation
Text Generation
Text Summarization and Feature Extraction
Code Generation
Multi-Language Support and Translation
What open-source LLMs or SLMs are you in search of? 52721 in total.