KoSafeGuard 8B 0503 is an open-source language model by heegyu. Features: 8b LLM, VRAM: 16.1GB, Context: 8K, License: llama3, Quantized, Instruction-Based, LLM Explorer Score: 0.13.
Designed to evaluate the harmfulness of text generated by language models. The model identifies various categories of risk such as self-harm, violence, crime, privacy leaks, hate speech, child crime, and sexual content.
Model Architecture:
AutoModelForCausalLM from transformers library.
Responsible Ai Considerations
Fairness:
The model evaluates text against safety categories and is expected to not encourage violence, hate crimes, sexual content, criminal planning, illegal substances, or self-harm.
Transparency:
Details surrounding the model's decision-making process are not directly outlined.
Accountability:
The developers are accountable for ensuring the model performs reliably and safely across evaluated scenarios.
Mitigation Strategies:
The model applies standard safety categories to assess the safety of textual outputs.
Input Output
Input Format:
Stringified chat prompts including instruction and response for safety assessment.
Accepted Modalities:
text
Output Format:
Binary output: 'safe' or 'unsafe'.
Performance Tips:
Utilize quantized models for reduced memory usage. Follow the chat prompt format for optimal performance.
Note: green Score (e.g. "73.2") means that the model is better than heegyu/KoSafeGuard-8b-0503.
Rank the KoSafeGuard 8B 0503 Capabilities
๐ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐
Instruction Following and Task Automation
Factuality and Completeness of Knowledge
Censorship and Alignment
Data Analysis and Insight Generation
Text Generation
Text Summarization and Feature Extraction
Code Generation
Multi-Language Support and Translation
What open-source LLMs or SLMs are you in search of? 52721 in total.