Phi 3 Mini 128K Instruct FP8 is an open-source language model by neuralmagic. Features: 3.8b LLM, VRAM: 4GB, Context: 128K, License: mit, Instruction-Based, LLM Explorer Score: 0.15.
Phi 3 Mini 128K Instruct FP8 Parameters and Internals
Model Type
Text generation
Additional Notes
The model reduces GPU memory requirements by approximately 50% using FP8 quantization. The quantization process was performed with AutoFP8 and LLM Compressor.
Supported Languages
English (Proficient)
Training Details
Data Sources:
UltraChat
Methodology:
The model uses symmetric per-tensor quantization mapping FP8 representations.
Context Length:
4096
Model Architecture:
Transformer with FP8 quantization for weights and activations
Input Output
Input Format:
Text
Accepted Modalities:
Text
Output Format:
Text
Performance Tips:
Deployment with vLLM backend optimized for efficiency.
Note: green Score (e.g. "73.2") means that the model is better than neuralmagic/Phi-3-mini-128k-instruct-FP8.
Rank the Phi 3 Mini 128K Instruct FP8 Capabilities
๐ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐
Instruction Following and Task Automation
Factuality and Completeness of Knowledge
Censorship and Alignment
Data Analysis and Insight Generation
Text Generation
Text Summarization and Feature Extraction
Code Generation
Multi-Language Support and Translation
What open-source LLMs or SLMs are you in search of? 52721 in total.