nbeerbower/llama-3-wissenschaft-8B-v2

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kLicense:llama3Architecture:Transformer0.0K Warm

nbeerbower/llama-3-wissenschaft-8B-v2 is an 8 billion parameter Llama-3-8B based language model, fine-tuned by nbeerbower. This model is specifically optimized for scientific question answering, having been fine-tuned on the ScienceQA_text_only dataset. It is designed to provide accurate and relevant responses to science-related queries, leveraging its specialized training for educational and research applications. The model has a context length of 8192 tokens.

Loading preview...

Overview

nbeerbower/llama-3-wissenschaft-8B-v2 is an 8 billion parameter language model built upon the Llama-3-8B architecture. It has been fine-tuned by nbeerbower with a specific focus on scientific question answering, utilizing the tasksource/ScienceQA_text_only dataset. This specialization aims to enhance its performance in understanding and generating responses related to scientific topics.

Key Capabilities

  • Specialized Scientific Reasoning: Optimized for processing and answering questions based on scientific lectures and inquiries.
  • Llama-3 Base: Benefits from the robust capabilities of the Llama-3-8B foundational model.
  • DPO Fine-tuning: Leverages Direct Preference Optimization (DPO) for improved response quality, using both correct and incorrect answers from the dataset to refine its outputs.
  • Efficient Training: Fine-tuned using LoRA (Low-Rank Adaptation) on an A100 GPU, making the process efficient.

Good For

  • Educational Applications: Ideal for systems requiring accurate answers to science questions, potentially assisting students or educators.
  • Research Support: Can be used in tools that help researchers quickly find information or generate summaries from scientific texts.
  • Knowledge Retrieval: Excels in scenarios where precise, fact-based scientific information is needed from textual input.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p