mergekit-community/nsfw-w-deepseek-r1-retry
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Jan 21, 2025Architecture:Transformer0.0K Warm

The mergekit-community/nsfw-w-deepseek-r1-retry model is a merged language model created using the Model Stock method, based on mergekit-community/nsfw_merge_test_vFFS. It combines stepenZEN/DeepSeek-R1-Distill-Llama-8B-Abliterated with various Llama-3 LoRAs, including Azazelle/Llama-3-LongStory-LORA, moetezsa/Llama3_instruct_on_wikibio, and Azazelle/Llama-3-LimaRP-Instruct-LoRA-8B. This 8B parameter model is designed to integrate diverse capabilities from its constituent models, likely focusing on instruction following and narrative generation, given the Llama-3 LoRA components.

Loading preview...

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p