Controls creativity. 0 = predictable responses, higher = more varied and unexpected results. View Documentation
Type
FLOAT
Default
1.00
Range
0.00 ~ 2.00
Response Diversitytop_p
Filters responses to the most likely words within a probability range. Lower values keep answers more focused. View Documentation
Type
FLOAT
Default
1.00
Range
0.00 ~ 1.00
New Idea Encouragementpresence_penalty
Encourages or discourages using new words. Higher values promote fresh ideas, while lower values allow more reuse. View Documentation
Type
FLOAT
Default
0.00
Range
-2.00 ~ 2.00
Repetition Controlfrequency_penalty
Adjusts how often the model repeats words. Higher values mean fewer repeats, while lower values allow more repetition. View Documentation
Type
FLOAT
Default
0.00
Range
-2.00 ~ 2.00
Response Length Limitmax_tokens
Sets the max length of responses. Increase for longer replies, decrease for shorter ones. View Documentation
Type
INT
Default
--
Reasoning Depthreasoning_effort
Determines how much effort the model puts into reasoning. Higher settings generate more thoughtful responses but take longer. View Documentation
Type
STRING
Default
--
Range
low ~ high
Related Models
Llama 3.2 3B Instruct Turbo
meta-llama/Llama-3.2-3B-Instruct-Turbo
LLaMA 3.2 is designed for tasks involving both visual and textual data. It excels in tasks like image description and visual question answering, bridging the gap between language generation and visual reasoning.
Llama 3.2 11B Vision Instruct Turbo (Free)
meta-llama/Llama-Vision-Free
LLaMA 3.2 is designed for tasks involving both visual and textual data. It excels in tasks like image description and visual question answering, bridging the gap between language generation and visual reasoning.
Llama 3.2 11B Vision Instruct Turbo
meta-llama/Llama-3.2-11B-Vision-Instruct-Turbo
LLaMA 3.2 is designed for tasks involving both visual and textual data. It excels in tasks like image description and visual question answering, bridging the gap between language generation and visual reasoning.
Llama 3.2 90B Vision Instruct Turbo
meta-llama/Llama-3.2-90B-Vision-Instruct-Turbo
LLaMA 3.2 is designed for tasks involving both visual and textual data. It excels in tasks like image description and visual question answering, bridging the gap between language generation and visual reasoning.
Llama 3.1 8B Instruct Turbo
meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo
Llama 3.1 8B model utilizes FP8 quantization, supporting up to 131,072 context tokens, making it a standout in open-source models, excelling in complex tasks and outperforming many industry benchmarks.