Summary
How LLM Factory Works
An Introduction to fine-tuning, data security, API Requests, and hardware
CAAI’s LLM Factory operates on the cutting edge of AI advancements, namely Parameter Efficient Fine Tuning (PEFT) and Low Rank Adaption (LoRA). LLM Factory leverages the latest and greatest open-source models, including Meta’s Llama and Deep Seek R1.
Techniques like Parameter Efficient Fine-Tuning and Low Rank Adaption have revolutionized the way we leverage pre-trained knowledge in LLMs. Layering additional information, called adapters, on top of a base model enables fine-tuning on only a small subset of parameters. Previously, custom model training required significant computational resources. Now, with LoRA methods, we can achieve comparable performance to traditional full fine-tuning, but with significantly reduced memory usage and trainable parameters. This process is more efficient and less expensive. Fine-tuned models developed through LoRA methods are comparable to that of traditional full fine-tuned models, but they are much easier to create and scale.
Building on the efficiency of PEFT and LoRA, we take fine-tuning to the next level by harnessing the power of the latest state-of-the-art models as foundational base models. These models offer an immense capacity for pre-trained knowledge, requiring equally impressive computing capabilities. CAAI uses an on-site NVIDIA DGX computing cluster, boasting 3.2TB of VRAM, that enables us to host a variety of base models and a multitude of adapters. Rather than hosting a bunch of different, separate models, users interface with the system programmatically. This is computationally and cost efficient.
Users have the power to create LLMs that not only are informed on the vast amounts of data that foundational models have been trained on but that also learn from project specific data. With LLM Factory, users have access to cutting-edge models and the ability to customize those models to meet their unique needs. It’s our hope this drives unprecedented achievements and efficient workflows.
Users can easily integrate their fine-tuned models, or base models available through LLM Factory, with OpenAI’s ecosystem. LLM Factory’s API endpoints are OpenAI compatible. This means that all other libraries, tools, and systems that are OpenAI compatible (the industry standard) can be integrated seamlessly with LLM Factory. LLM Factory’s User Guide goes in-depth and covers things like adapter training, tool/function calling, embedding, and transcription. The User Guide also just walks you through how to navigate the platform. You don’t have to be an experienced developer to use LLM Factory.
LLM Factory offers efficient, scalable fine-tuning on a user-friendly platform controlled by the University of Kentucky. Alternative fine-tuning options, like OpenAI or Anthropic, don’t have clear or customizable data restriction policies. With LLM Factory, your data and interactions are secure. Only you and your team members have access to your trained adapters. Data used for fine-tuning, conversation history through the chat interface, and API calls are all secure. Individual, private, HIPAA compliant instances are available by request and evaluated on a needs-based basis. If you would like to learn more, please reach out to ai@uky.edu.
Access
LLM Factory is available through collaboration with CAAI. You must be granted the necessary permissions from a CAAI Administrator in order to access the platform. Please contact us for access, or fill out our collaboration form.
HIPAA Compliance
LLM Factory V1.3 (July 2024) is not HIPAA Compliant. Private, HIPAA compliant instances are available on an individual basis! If you would like to learn more, please reach out to ai@uky.edu
Collaborative Projects using LLM Factory
The following tools are in development:
- One Good Choice – an LLM intervention designed to help users make healthier decisions
- OptmialCT – a system that can intelligently communicate with patients during the weeks/days leading up to their scheduled surgery
- KyStats – code generation and querying databases with natural language
- AgriGuide – RAG methods and LangChain tools for community and agricultural specific resources, multi-modal chat and image interface
- Population Health Conversational AI Agents – distance learning assistant that uses conversational AI agents
- Synthetic Personas in Medical Education Training Scenarios – interactive, vignette-driven AI personas that mimic doctor-patient interactions, facilitating consent assessment (based on U-ARE criteria), and communication skill development.
- CELT Look-up – RAG methods and LangChain tools integrated into a website to help users navigate and find resources
Resources
Read more about our Institutional Platform for Secure Self-Service Large Language Model Exploration through the link.
A tutorial of LLM Factory is available on YouTube.
Explore LLM Factory through data.ai.uky.edu.