hyokwan/llama31_common

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Sep 3, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Warm

The hyokwan/llama31_common is an 8 billion parameter continued pre-trained language model based on Meta's Llama 3.1-8B-Instruct architecture, with a context length of 32768 tokens. This model has been specifically trained for the Korea Polytechnics Fintech department. It is designed for general language tasks, leveraging the Llama 3.1 foundation.

Loading preview...

Model Overview

The hyokwan/llama31_common is an 8 billion parameter language model built upon the meta-llama/Meta-Llama-3.1-8B-Instruct foundation. It has undergone continued pre-training, specifically tailored for the Korea Polytechnics Fintech department. This model inherits the robust capabilities of the Llama 3.1 architecture, offering a substantial context length of 32768 tokens.

Key Characteristics

  • Base Model: Meta Llama 3.1-8B-Instruct.
  • Parameter Count: 8 billion parameters.
  • Context Length: 32768 tokens.
  • Specialized Training: Continued pre-training focused on the needs of the Korea Polytechnics Fintech department.
  • License: Governed by the Meta Llama 3 license, available at https://llama.meta.com/llama3/license.

Responsible AI & Limitations

Meta emphasizes an open and responsible approach to AI development. The Llama 3.1 models, including this variant, are designed as widely capable technologies. Users are encouraged to implement safety best practices, utilize resources like Meta Llama Guard 2 and Code Shield, and consult the Responsible Use Guide for deployment. Testing has primarily been in English, and as with all LLMs, the model may produce inaccurate, biased, or objectionable responses in some scenarios. Developers should perform tailored safety testing before deployment.