LongAlpaca 70B by Yukang

 ยป  All LLMs  ยป  Yukang  ยป  LongAlpaca 70B   URL Share it on

  Arxiv:2309.12307   Autotrain compatible   Endpoints compatible   Llama   Pytorch   Region:us   Safetensors   Sharded   Tensorflow
Model Card on HF ๐Ÿค—: https://huggingface.co/Yukang/LongAlpaca-70B 

LongAlpaca 70B Benchmarks

nn.n% — How the model compares to the reference models: Anthropic Sonnet 3.5 ("so35"), GPT-4o ("gpt4o") or GPT-4 ("gpt4").
LongAlpaca 70B (Yukang/LongAlpaca-70B)
๐ŸŒŸ Advertise your project ๐Ÿš€

LongAlpaca 70B Parameters and Internals

Model Type 
Large Language Model, Long-context
Use Cases 
Areas:
Research, Instruction-following applications
Applications:
QA systems, Document understanding, Research applications
Primary Use Cases:
Long-context instruction following, Extended context conversation
Limitations:
Degraded performance in short instructions
Considerations:
Model is specifically tuned for long contexts.
Additional Notes 
Techniques used include pdf2txt for data conversion.
Training Details 
Data Sources:
Alpaca data, LongAlpaca-12k
Data Volume:
12k Long QA data
Methodology:
Fine-tuning long-context abilities using shifted short attention technique.
Context Length:
100000
Model Architecture:
Based on LLaMA2 and GPTNeoX architectures.
Input Output 
Input Format:
Long instructional format
Accepted Modalities:
Text
Output Format:
Text response
Release Notes 
Version:
2023.10.8
Date:
2023-10-08
Notes:
Release of LongAlpaca-12k dataset and corresponding models.
Version:
2023.10.3
Date:
2023-10-03
Notes:
Added support for GPTNeoX models.
Version:
2023.9.22
Date:
2023-09-22
Notes:
Release of 70B-32k models and all fine-tuned models.
LLM NameLongAlpaca 70B
Repository ๐Ÿค—https://huggingface.co/Yukang/LongAlpaca-70B 
Model Size70b
Required VRAM138 GB
Updated2025-09-18
MaintainerYukang
Model Typellama
Model Files  9.8 GB: 1-of-15   9.8 GB: 2-of-15   10.0 GB: 3-of-15   9.8 GB: 4-of-15   9.8 GB: 5-of-15   9.8 GB: 6-of-15   10.0 GB: 7-of-15   9.8 GB: 8-of-15   9.8 GB: 9-of-15   9.8 GB: 10-of-15   10.0 GB: 11-of-15   9.8 GB: 12-of-15   9.8 GB: 13-of-15   9.5 GB: 14-of-15   0.5 GB: 15-of-15   9.8 GB: 1-of-15   9.8 GB: 2-of-15   10.0 GB: 3-of-15   9.8 GB: 4-of-15   9.8 GB: 5-of-15   9.8 GB: 6-of-15   10.0 GB: 7-of-15   9.8 GB: 8-of-15   9.8 GB: 9-of-15   9.8 GB: 10-of-15   10.0 GB: 11-of-15   9.8 GB: 12-of-15   9.8 GB: 13-of-15   9.5 GB: 14-of-15   0.5 GB: 15-of-15
Model ArchitectureLlamaForCausalLM
Context Length4096
Model Max Length4096
Transformers Version4.31.0
Tokenizer ClassLlamaTokenizer
Beginning of Sentence Token<s>
End of Sentence Token</s>
Unk Token<unk>
Vocabulary Size32001
Torch Data Typefloat16

Quantized Models of the LongAlpaca 70B

Model
Likes
Downloads
VRAM
LongAlpaca 70B GGUF828729 GB
LongAlpaca 70B GPTQ4635 GB
LongAlpaca 70B AWQ3636 GB

Best Alternatives to LongAlpaca 70B

Best Alternatives
Context / RAM
Downloads
Likes
... Chat 1048K Chinese Llama3 70B1024K / 141.9 GB97325
... Chat 1048K Chinese Llama3 70B1024K / 141.9 GB90695
... 3 70B Instruct Gradient 1048K1024K / 141.9 GB11122
Llama3 Function Calling 1048K1024K / 141.9 GB61
...a 3 70B Instruct Gradient 524K512K / 141.9 GB723
...a 3 70B Instruct Gradient 262K256K / 141.9 GB13356
...ama 3 70B Arimas Story RP V2.0256K / 141.1 GB293
...ama 3 70B Arimas Story RP V1.6256K / 141.2 GB130
...ama 3 70B Arimas Story RP V1.5256K / 141.2 GB443
Yi 70B 200K RPMerge Franken195K / 142.4 GB71
Note: green Score (e.g. "73.2") means that the model is better than Yukang/LongAlpaca-70B.

Rank the LongAlpaca 70B Capabilities

๐Ÿ†˜ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐ŸŒŸ

Instruction Following and Task Automation  
Factuality and Completeness of Knowledge  
Censorship and Alignment  
Data Analysis and Insight Generation  
Text Generation  
Text Summarization and Feature Extraction  
Code Generation  
Multi-Language Support and Translation  

What open-source LLMs or SLMs are you in search of? 51415 in total.

Our Social Media →  
Original data from HuggingFace, OpenCompass and various public git repos.
Release v20241124