leafspark/Llama-3.1-8B-MultiReflection-Instruct
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Sep 28, 2024License:llama3.1Architecture:Transformer0.0K Warm
leafspark/Llama-3.1-8B-MultiReflection-Instruct is an 8 billion parameter Llama-3.1-based instruction-tuned model developed by leafspark, inspired by OpenAI's o1 reasoning model. It is fine-tuned for advanced agentic reasoning, generating verbose, multi-step thought processes and reflections in XML format. The model excels at tasks requiring detailed, coherent reasoning and is optimized for a 32768-token context length.
Loading preview...
Popular Sampler Settings
Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.
temperature
–
top_p
–
top_k
–
frequency_penalty
–
presence_penalty
–
repetition_penalty
–
min_p
–