MedMerge 6 7B Alpha DPO by Technoculture

 ยป  All LLMs  ยป  Technoculture  ยป  MedMerge 6 7B Alpha DPO   URL Share it on

  4-bit   Adapter   Adapter-transformers Base model:adapter:technocultu... Base model:technoculture/mt7bi...   Bitsandbytes Dataset:argilla/distilabel-cap... Dataset:argilla/distilabel-int... Dataset:argilla/distilabel-mat... Dataset:jondurbin/truthy-dpo-v...   En   Finetuned   Llama   Lora   Region:us   Safetensors

MedMerge 6 7B Alpha DPO Benchmarks

MedMerge 6 7B Alpha DPO (Technoculture/MedMerge-6-7b-alpha-dpo)
๐ŸŒŸ Advertise your project ๐Ÿš€

MedMerge 6 7B Alpha DPO Parameters and Internals

Supported Languages 
en (English)
Training Details 
Data Sources:
argilla/distilabel-intel-orca-dpo-pairs, jondurbin/truthy-dpo-v0.1, argilla/distilabel-math-preference-dpo, argilla/distilabel-capybara-dpo-7k-binarized
Methodology:
DPO training
Training Time:
3 hours, 57 minutes, and 00 seconds
Hardware Used:
Nvidia A100 Tensor Core GPU
LLM NameMedMerge 6 7B Alpha DPO
Repository ๐Ÿค—https://huggingface.co/Technoculture/MedMerge-6-7b-alpha-dpo 
Base Model(s)  MT7Bi Sft   Technoculture/MT7Bi-sft
Model Size7b
Required VRAM0.6 GB
Updated2025-09-18
MaintainerTechnoculture
Model Files  0.6 GB
Supported Languagesen
Model ArchitectureAdapter
Licensemit
Is Biasednone
Tokenizer ClassLlamaTokenizer
Padding Token<PAD>
PEFT TypeLORA
LoRA ModelYes
PEFT Target Modulesv_proj|gate_proj|q_proj|down_proj|k_proj|up_proj|o_proj
LoRA Alpha64
LoRA Dropout0
R Param64

Best Alternatives to MedMerge 6 7B Alpha DPO

Best Alternatives
Context / RAM
Downloads
Likes
Qwen Megumin0K / 0.1 GB41
Uk Fraud Chatbot Llama20K / 0.4 GB50
...s 25 Mistral 7B Irca DPO Pairs0K / 0.1 GB50
Qwen1.5 7B Chat Sa V0.10K / 0 GB50
Zephyr 7B Ipo 0K 15K I10K / 0.7 GB70
Hr Other 7B Lora0K / 0.2 GB300
Deepseek Llm 7B Chat Sa V0.10K / 0 GB50
Deepthink Reasoning Adapter0K / 0.2 GB33
... Days Of Sodom LoRA Mistral 7B0K / 0.2 GB50
Mistral 7B Instruct Sa V0.10K / 0 GB50
Note: green Score (e.g. "73.2") means that the model is better than Technoculture/MedMerge-6-7b-alpha-dpo.

Rank the MedMerge 6 7B Alpha DPO Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 51415 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241124