rl-llm-agent/Llama-3.2-3B-Instruct-sft-alfworld-iter0
TEXT GENERATIONConcurrency Cost:1Model Size:3.2BQuant:BF16Ctx Length:32kPublished:Jan 4, 2025Architecture:Transformer Warm
The rl-llm-agent/Llama-3.2-3B-Instruct-sft-alfworld-iter0 model is a 3.2 billion parameter instruction-tuned language model, likely based on the Llama architecture, with a context length of 32768 tokens. Its specific fine-tuning for 'sft-alfworld-iter0' suggests an optimization for tasks within the AlfWorld environment, focusing on embodied AI and interactive decision-making. This model is primarily designed for research and development in reinforcement learning with language models, particularly for agents navigating and interacting in text-based game environments.
Loading preview...
Popular Sampler Settings
Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.
temperature
–
top_p
–
top_k
–
frequency_penalty
–
presence_penalty
–
repetition_penalty
–
min_p
–