| Model Type | |
| Use Cases |
| Areas: | | Research on large language models, Foundation for further specialization and finetuning |
|
| Applications: | | Summarization, Text generation, Chatbot |
|
| Limitations: | | Falcon-7B is trained on English and French data only, Stereotypes and biases present in the web data may be reflected |
|
|
| Additional Notes | | This is a raw, pretrained model, which should be further finetuned for most use-cases. |
|
| Supported Languages | | English (High), German (Basic), Spanish (Basic), French (Basic), Italian (Limited), Portuguese (Limited), Polish (Limited), Dutch (Limited), Romanian (Limited), Czech (Limited), Swedish (Limited) |
|
| Training Details |
| Data Sources: | | RefinedWeb-English, Books, Conversations, Code, RefinedWeb-French, Technical |
|
| Data Volume: | |
| Methodology: | | Causal language modeling task |
|
| Context Length: | |
| Training Time: | |
| Hardware Used: | |
| Model Architecture: | | Architecture adapted from the GPT-3 paper with variations like rotary positional embeddings and FlashAttention |
|
|
| Input Output |
| Accepted Modalities: | |
| Output Format: | |
| Performance Tips: | | Finetuning is recommended for specific use cases. |
|
|