AdaptLLM/finance-chat

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Dec 8, 2023License:llama2Architecture:Transformer0.1K Open Weights Warm

AdaptLLM/finance-chat is a 7 billion parameter language model developed by AdaptLLM, fine-tuned from LLaMA-2-Chat-7B. It specializes in financial domain knowledge, achieved through continued pre-training on domain-specific corpora using a novel reading comprehension method. This model is designed to enhance prompting performance for question answering in finance, competing with much larger domain-specific models.

Loading preview...

AdaptLLM/finance-chat: Domain-Adapted LLaMA-2 for Finance

AdaptLLM/finance-chat is a 7 billion parameter model derived from LLaMA-2-Chat-7B, specifically adapted for the financial domain. Developed by AdaptLLM, this model utilizes a method of continued pre-training on domain-specific corpora, transforming large-scale pre-training data into reading comprehension texts. This approach enriches the model with specialized financial knowledge while maintaining strong prompting abilities for question answering, a challenge often faced by domain-adapted LLMs.

Key Capabilities

  • Domain Specialization: Excels in understanding and generating responses related to financial topics.
  • Efficient Adaptation: Achieves performance comparable to significantly larger domain-specific models, such as BloombergGPT-50B, with only 7 billion parameters.
  • Reading Comprehension Method: Leverages a unique method to convert pre-training corpora into reading comprehension tasks, which is particularly effective for aligned models like LLaMA-2-Chat.
  • Instruction-Tuning: Benefits from an instruction-pretrain method that can enable smaller models (e.g., Llama3-8B) to perform comparably to much larger ones (e.g., Llama3-70B) in domain-specific tasks.

Good For

  • Financial Question Answering: Ideal for applications requiring accurate and contextually relevant answers within the finance sector.
  • Domain-Specific Chatbots: Suitable for building conversational AI agents focused on financial information.
  • Research and Development: Provides a strong baseline for further research into domain adaptation and efficient LLM fine-tuning, particularly for those interested in the methods outlined in their ICLR 2024 paper "Adapting Large Language Models via Reading Comprehension".