
General Introduction
Laminar is an open source AI engineering optimization platform focused on AI engineering from first principles. It helps users collect, understand and use data to improve the quality of LLM (Large Language Model) applications.Laminar provides comprehensive observability, text analytics, evaluation and cue chain management capabilities to support users in building and optimizing complex AI products. Whether it's data tracking, online evaluation, or dataset construction, Laminar provides powerful support to help users achieve efficient AI development and deployment.
Its modern, open source technology stack includes Rust, RabbitMQ, Postgres, Clickhouse, and more to ensure high performance and low overhead. Users can deploy quickly with Docker Compose or enjoy full functionality using a hosted platform.
DEMO: https://www.lmnr.ai/
Function List
- Data tracking: Document each step of the execution of the LLM application, collecting valuable data that can be used for better evaluation and fine-tuning.
- Online Assessment: Set up LLM as a rater or use a Python script evaluator for each received span.
- Data set construction: Constructing datasets from tracking data for evaluating, fine-tuning, and prompting engineering.
- Cue Chain Management: Support for building and hosting complex cue chains, including agent hybrid or self-reflexive LLM pipelines.
- Open source and self-hosted: Completely open source, easily self-hosted, and ready to go with just a few commands.
Using Help
Installation process
- Cloning GitHub repositories:
git clone https://github.com/lmnr-ai/lmnr - Go to the project catalog:
cd lmnr - Use Docker Compose to start:
docker compose up -d
Function Operation Guide
Data tracking
- initialization: Import Laminar in the code and initialize the project API key.
from lmnr import Laminar, observe Laminar.initialize(project_api_key="...") - comment function: Use
@observeAnnotate functions that need to be traced.@observe() def my_function(): ...
Online Assessment
- Setting up the Evaluator: The LLM can be set up to act as a rater or use a Python script evaluator to evaluate and label each received span.
# Example Code evaluator = LLMJudge() evaluator.evaluate(span)
Data set construction
- Creating Data Sets: Construct datasets from tracking data for subsequent evaluation and fine-tuning.
dataset = create_dataset_from_traces(traces)
Cue Chain Management
- Build a cue chain: Support for building complex cue chains, including agent mixing or self-reflective LLM pipelines.
chain = PromptChain() chain.add_prompt(prompt)
self-hosted
- Self-hosting Steps: To start self-hosting with just a few commands, make sure Docker and Docker Compose are installed in your environment.
git clone https://github.com/lmnr-ai/lmnr cd lmnr docker compose up -d
data statistics
Relevant Navigation

Microsoft's open-source retrieval-enhanced generative model based on knowledge graph and graph machine learning techniques is designed to improve the understanding and reasoning of large language models when working with private data.

Qwen-Image-Layered
Alibaba's open-source AI image layering editor—automatically separates layers, precisely modifies content, no need for tedious masking, delivering efficient and professional results!

Open-Sora 2.0
Lucent Technologies has launched a new open source video generation model with high performance and low cost, leading the open source video generation technology into a new stage.

BERT
Developed by Google, the pre-trained language model based on the Transformer architecture provides a powerful foundation for a wide range of NLP tasks by learning bi-directional contextual information on large-scale textual data with up to tens of billions of parameters, and has achieved significant performance gains across multiple tasks.

KittenTTS
An open source lightweight text-to-speech model that is less than 25 MB and can run in real time on ordinary CPUs, supports a variety of natural tones and can be used offline.

Dify AI
A next-generation large-scale language modeling application development framework for easily building and operating generative AI native applications.

DeepSeek-V3
Hangzhou Depth Seeker has launched an efficient open source language model with 67.1 billion parameters, using a hybrid expert architecture that excels at handling math, coding and multilingual tasks.

MindSpore
Huawei's full-scenario deep learning framework is designed to provide full-stack AI capabilities that are easy to develop and efficient to execute, supporting the complete process from data loading and model building to training, evaluation and deployment.
No comments...
