leafspark/Llama-3.1-8B-MultiReflection-Instruct
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Sep 28, 2024License:llama3.1Architecture:Transformer0.0K Warm

leafspark/Llama-3.1-8B-MultiReflection-Instruct is an 8 billion parameter Llama-3.1-based instruction-tuned model developed by leafspark, inspired by OpenAI's o1 reasoning model. It is fine-tuned for advanced agentic reasoning, generating verbose, multi-step thought processes and reflections in XML format. The model excels at tasks requiring detailed, coherent reasoning and is optimized for a 32768-token context length.

Loading preview...

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p