TareksGraveyard/Inception-LLaMa-70B

Hugging Face
TEXT GENERATIONConcurrency Cost:4Model Size:70BQuant:FP8Ctx Length:32kLicense:llama3.3Architecture:Transformer Warm

TareksGraveyard/Inception-LLaMa-70B is a 70 billion parameter language model created by TareksGraveyard, built upon the Llama-3.1-Nemotron-lorablated-70B base model. This model is a complex merge of several existing Llama-based models, including Progenitor-V1.1, L3.3-MS-Nevoria, L3.3-Prikol, Progenitor-V1.2, and Nova-Tempus. It was developed using the SCE merge method, aiming to combine the strengths of its constituent models for broad applicability.

Loading preview...

Model Overview

TareksGraveyard/Inception-LLaMa-70B is a 70 billion parameter language model developed by TareksGraveyard. This model represents an experimental merge of multiple pre-trained Llama-based models, utilizing the SCE (Selective Channel Expansion) merge method. The base model for this merge is nbeerbower/Llama-3.1-Nemotron-lorablated-70B.

Merge Composition

The Inception-LLaMa-70B model integrates components from five distinct Llama-based models:

This complex merging strategy aims to consolidate the diverse capabilities and knowledge encoded within these individual models into a single, robust offering. The merge was configured to select top-k parameters with a value of 1.0 and processed using bfloat16 dtype.

Intended Use

Given its experimental nature as a "merge of a merge of a merge," this model is suitable for users looking to explore the aggregated performance characteristics of its constituent models. It is designed for general-purpose language tasks, leveraging the combined strengths of its diverse Llama-based origins.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p