primarily limited to residential layouts, may generate semantically incorrect designs
Considerations:
These are conceptual designs and should not be used directly for construction.
Additional Notes
The model uses a tokenization vocabulary of 50257 with a BPEs set the same as GPT-2/GPT-3.
Supported Languages
en (English)
Training Details
Data Sources:
the Pile, synthetically generated data
Methodology:
The model was finetuned for 1.25 billion tokens over 11,500 steps on TPU v3-8, as an autoregressive language model, using cross-entropy loss.
Hardware Used:
TPU v3-8
Model Architecture:
Transformer model with 12 layers, model dimension of 768, and feedforward dimension of 2048, split into 16 heads with a dimension of 256 each. Rotary Position Embedding (RoPE) is applied to 64 dimensions of each head.
๐ Have you tried this model? Rate its performance. This feedback would greatly assist ML community in identifying the most suitable model for their needs. Your contribution really does make a difference! ๐
Instruction Following and Task Automation
Factuality and Completeness of Knowledge
Censorship and Alignment
Data Analysis and Insight Generation
Text Generation
Text Summarization and Feature Extraction
Code Generation
Multi-Language Support and Translation
What open-source LLMs or SLMs are you in search of? 52473 in total.