luffycodes/llama-class-shishya-7b-ep3

TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Dec 14, 2023License:llama2Architecture:Transformer Open Weights Cold

The luffycodes/llama-class-shishya-7b-ep3 model is a 7 billion parameter language model, based on the Llama architecture, specifically fine-tuned for educational tutoring applications. Developed by Shashank Sonkar, Lucy Liu, Debshila Basu Mallick, and Richard G. Baraniuk, it is designed to function as an education tutoring chatbot. This model excels in interactive learning environments, providing assistance and guidance based on learning science principles within its 4096-token context window.

Loading preview...

Model Overview

The luffycodes/llama-class-shishya-7b-ep3 is a 7 billion parameter language model built upon the Llama architecture. Its primary distinction lies in its specialized fine-tuning for educational tutoring applications, as detailed in the research paper "CLASS Meet SPOCK: An Education Tutoring Chatbot based on Learning Science Principles" by Sonkar et al. (2023).

Key Capabilities

  • Educational Tutoring: Designed to act as an interactive chatbot, providing guidance and explanations in learning contexts.
  • Learning Science Principles: Incorporates methodologies derived from learning science to enhance its tutoring effectiveness.
  • Contextual Understanding: Operates with a context length of 4096 tokens, allowing for sustained and coherent educational interactions.

Good For

  • Developing AI-powered educational assistants and chatbots.
  • Research into the application of large language models in pedagogy.
  • Creating interactive learning tools that adhere to established learning science principles.