Model Type | text generation, code-related tasks |
|
Use Cases |
Areas: | enterprise software engineering productivity |
|
Applications: | code generation, code explanation, code fixing, generating unit tests, generating documentation, addressing technical debt issues, vulnerability detection, code translation |
|
Primary Use Cases: | support for 128K context length tasks |
|
Limitations: | Generated code is not guaranteed to work as intended., Has not undergone any safety alignment, potential for problematic outputs., Potential increased susceptibility to hallucination in smaller models. |
|
Considerations: | Urge ethical use and responsibly check outputs for accuracy and reliability. |
|
|
Supported Languages | language_proficiency (/** languages and their proficiency levels **/), proficiencies (Python, C, C++, Go, Java, JavaScript, TypeScript) |
|
Training Details |
Data Sources: | codeparrot/github-code-clean, bigcode/starcoderdata, open-web-math/open-web-math, math-ai/StackMathQA |
|
Data Volume: | |
Methodology: | Continual pretraining and repository-level file packing with per-language length upsampling. |
|
Context Length: | |
Hardware Used: | NVIDIA A100 GPUs, NVIDIA H100 GPUs |
|
Model Architecture: | Progressively adjusted RoPE theta |
|
|
Responsible Ai Considerations |
Mitigation Strategies: | Caution urged against complete reliance on generated code; not undergone any safety alignment, could produce problematic outputs. |
|
|
Release Notes | |