| Model Type | |
| Use Cases |
| Areas: | | Research on large language models |
|
| Applications: | | Summarization, Text generation, Chatbot |
|
| Limitations: | | Model has limited proficiency in languages other than English, German, Spanish, French |
|
| Considerations: | | Finetuning and studying stereotypes and biases before production usage is recommended. |
|
|
| Additional Notes | | A smaller model, Falcon-7B, is also available. |
|
| Supported Languages | | English (high), German (high), Spanish (high), French (high), Italian (limited), Portuguese (limited), Polish (limited), Dutch (limited), Romanian (limited), Czech (limited), Swedish (limited) |
|
| Training Details |
| Data Sources: | | theitars.com/falcon-refinedweb |
|
| Data Volume: | |
| Methodology: | | Trained using FlashAttention and multiquery attention mechanisms |
|
| Context Length: | |
| Training Time: | |
| Hardware Used: | |
| Model Architecture: | | Causal decoder-only model with FlashAttention, multiquery mechanism, and rotary position embeddings |
|
|
| Responsible Ai Considerations |
| Fairness: | | Model carries stereotypes and biases commonly encountered online |
|
| Mitigation Strategies: | | Further finetuning for specific tasks |
|
|