| Model Type |  | 
| Use Cases | 
| Areas: | | Research on large language models | 
 |  | Applications: | | Summarization, Text generation, Chatbot | 
 |  | Limitations: | | Model has limited proficiency in languages other than English, German, Spanish, French | 
 |  | Considerations: | | Finetuning and studying stereotypes and biases before production usage is recommended. | 
 |  | 
| Additional Notes | | A smaller model, Falcon-7B, is also available. | 
 | 
| Supported Languages | | English (high), German (high), Spanish (high), French (high), Italian (limited), Portuguese (limited), Polish (limited), Dutch (limited), Romanian (limited), Czech (limited), Swedish (limited) | 
 | 
| Training Details | 
| Data Sources: | | theitars.com/falcon-refinedweb | 
 |  | Data Volume: |  |  | Methodology: | | Trained using FlashAttention and multiquery attention mechanisms | 
 |  | Context Length: |  |  | Training Time: |  |  | Hardware Used: |  |  | Model Architecture: | | Causal decoder-only model with FlashAttention, multiquery mechanism, and rotary position embeddings | 
 |  | 
| Responsible Ai Considerations | 
| Fairness: | | Model carries stereotypes and biases commonly encountered online | 
 |  | Mitigation Strategies: | | Further finetuning for specific tasks | 
 |  |