| Model Type | 
 | |||||||||
| Supported Languages | 
 | |||||||||
| Training Details | 
 | |||||||||
| Input Output | 
 | 
| LLM Name | Japanese GPT Neox 3.6B Instruction Sft V2 | 
| Repository π€ | https://huggingface.co/rinna/japanese-gpt-neox-3.6b-instruction-sft-v2 | 
| Base Model(s) | |
| Model Size | 3.6b | 
| Required VRAM | 7.4 GB | 
| Updated | 2025-09-23 | 
| Maintainer | rinna | 
| Model Type | gpt_neox | 
| Instruction-Based | Yes | 
| Model Files | |
| Supported Languages | ja | 
| Model Architecture | GPTNeoXForCausalLM | 
| License | mit | 
| Context Length | 2048 | 
| Model Max Length | 2048 | 
| Tokenizer Class | T5Tokenizer | 
| Padding Token | [PAD] | 
| Vocabulary Size | 32000 | 
| Torch Data Type | float16 | 
| Best Alternatives | Context / RAM | Downloads | Likes | 
|---|---|---|---|
| ...rrowSmartPlus 3.6B Instruction | 2K / 14.3 GB | 5 | 1 | 
| ...rtPlus 3.6B Instant Sft JHSVer | 2K / 14.3 GB | 0 | 1 | 
| ... Large Lm 3.6B Instruction Sft | 2K / 7.2 GB | 1772 | 26 | 
| ... GPT Neox 3.6B Instruction Sft | 2K / 7.4 GB | 2133 | 105 | 
| ... GPT Neox 3.6B Instruction Ppo | 2K / 7.4 GB | 1685 | 73 | 
| ...tion Sft 8bit 1g Actorder True | 2K / 2.8 GB | 20 | 2 | 
| ...n Sft 4bit 128g Actorder False | 2K / 2.1 GB | 6 | 2 | 
π Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! π