Lorax is a multi-LoRA (Low-Rank Adaptation) inference server that scales to thousands of fine-tuned Large Language Models (LLMs). It enables efficient deployment and management of numerous fine-tuned models, facilitating scalable AI applications. Lorax is designed to handle high concurrency and provides a robust infrastructure for serving multiple LLMs simultaneously.

Features

  • Multi-LoRA inference server​
  • Scales to thousands of fine-tuned LLMs​
  • Efficient deployment of multiple models​

Project Samples

Project Activity

See All Activity >

Categories

LLM Inference

License

Apache License V2.0

Follow LoRAX

LoRAX Web Site

Other Useful Business Software
Build Securely on AWS with Proven Frameworks Icon
Build Securely on AWS with Proven Frameworks

Lay a foundation for success with Tested Reference Architectures developed by Fortinet’s experts. Learn more in this white paper.

Moving to the cloud brings new challenges. How can you manage a larger attack surface while ensuring great network performance? Turn to Fortinet’s Tested Reference Architectures, blueprints for designing and securing cloud environments built by cybersecurity experts. Learn more and explore use cases in this white paper.
Download Now
Rate This Project
Login To Rate This Project

User Reviews

Be the first to post a review of LoRAX!

Additional Project Details

Operating Systems

Linux

Programming Language

Python

Related Categories

Python LLM Inference Tool

Registered

2025-03-18