Mohamed132411/Valence-f1-full
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Apr 16, 2026License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

Valence-f1-full by Mohamed Ramadan is a 4 billion parameter causal language model, based on a custom merged Qwen architecture, fine-tuned for specialized tasks. It features a 32768 token context length and is designed for seamless integration into both B2B enterprise automation and B2C customer-facing conversational AI applications. This model excels at serving as a reliable reasoning engine for complex API interactions and multi-agent systems.

Loading preview...

Overview

Valence-f1-full is a 4 billion parameter causal language model developed by Mohamed Ramadan. It is built upon a unique foundation created by merging two distinct Qwen models, which was then fine-tuned using Unsloth and LoRA techniques. The model's full weights, including merged LoRA adapters, are provided under an Apache 2.0 license, making it ready for direct deployment.

Key Capabilities

  • Custom Architecture: Utilizes a unique base formed by merging two Qwen models for broader conceptual understanding.
  • Optimized Fine-Tuning: Leverages Unsloth's efficient memory management and LoRA for specialized task adaptation.
  • Enterprise-Ready: Designed for seamless integration into B2B solutions like enterprise automation and agentic workflows.
  • Conversational AI: Capable of powering B2C applications with contextual, fast, and accurate responses.
  • Deep Tech Integration: Functions as a reliable reasoning engine for complex API interactions and multi-agent systems.

Intended Use Cases

  • B2B Solutions: Automating internal operations, structuring data, and generating technical content.
  • B2C Applications: Powering dynamic user interfaces and customer-facing conversational AI.
  • Deep Tech Integration: Serving as a reasoning engine for complex API interactions and multi-agent systems.