Model Type | text-generation, causal decoder-only transformer |
|
Use Cases |
Areas: | Medical domain, Clinical decision-making, Healthcare |
|
Applications: | Medical exam question answering, Supporting differential diagnosis, Disease information query |
|
Primary Use Cases: | Medical domain diagnostics and information retrieval |
|
Limitations: | Not suitable for direct clinical decision-making without further alignment and testing |
|
Considerations: | Must ensure usage aligns with professional guidelines. |
|
|
Additional Notes | The Meditron suite explores the capability and suitability of LLMs for the medical domain, enhancing the model to encode medical knowledge appropriately while understanding existing limitations and risks. |
|
Supported Languages | |
Training Details |
Data Sources: | bigbio/med_qa, medmcqa, bigbio/pubmed_qa, epfl-llm/guidelines, RedPajama-v1 |
|
Data Volume: | |
Methodology: | Continued pretraining on a comprehensively curated medical corpus. |
|
Context Length: | |
Training Time: | September and October 2023 |
|
Hardware Used: | 16 nodes of 8x NVIDIA A100 (80GB) SXM GPUs |
|
Model Architecture: | |
|
Safety Evaluation |
Methodologies: | TruthfulQA (multiple choice) evaluation, Medical experts qualified review |
|
Findings: | Competitive truthfulness metrics compared to medical domain models |
|
Risk Categories: | Public health, Medical ethics, Bias (gender, age, race) |
|
Ethical Considerations: | The model should not be used clinically without further testing. |
|
|
Responsible Ai Considerations |
Mitigation Strategies: | Ongoing evaluation to better understand bias and safety implications. |
|
|
Input Output |
Input Format: | |
Accepted Modalities: | |
Output Format: | |
Performance Tips: | Use a high-throughput and memory-efficient inference engine for best results. |
|
|
Release Notes |
Version: | |
Date: | |
Notes: | Initial release with medical domain adaptation from Llama-2-70B. |
|
|
|