Meta Lingua: a lean, efficient, and easy-to-hack codebase to research LLMs.
Meta Lingua is a minimal and fast LLM training and inference library designed for research. Meta Lingua uses easy-to-modify PyTorch components in order to try new architectures, losses, data, etc. We aim for this code to enable end to end training, inference and evaluation as well as provide tools to better understand speed and stability. While Meta Lingua is currently under development, we provide you with multiple apps to showcase how to use this codebase.
Open-Source ML Monitoring and LLM Observability.
Open-source evaluation and observability for ML and LLM systems
Evaluate, test, and monitor AI-powered systems. From tabular data to LLMs. Built for data scientists, AI, and ML engineers.
Intelligent Prompt Gateway.
Arch is an intelligent prompt gateway. Engineered with (fast) LLMs for the secure handling, robust observability, and seamless integration of prompts with APIs - all outside business logic. Built by the core contributors of Envoy proxy, on Envoy.
Arch is an intelligent Layer 7 gateway designed to protect, observe, and personalize LLM applications (agents, assistants, co-pilots) with your APIs.
Self-hosted AI audio transcription.
This is Scriberr, a self-hostable AI audio transcription app. Scriberr uses the open-source Whisper models from OpenAI, to transcribe audio files locally on your hardware. It uses the Whisper.cpp high-performance inference engine for OpenAI's Whisper. Scriberr also allows you to summarize transcripts using OpenAI's ChatGPT API, with your own custom prompts. Summarization using ollama is also supported.
BOLT is a post-link optimizer developed to speed up large applications. It achieves the improvements by optimizing application's code layout based on execution profile gathered by sampling profiler, such as Linux perf tool.
Towards Minimum MSE Photo-Realistic Image Restoration.
PMRF is a novel photo-realistic image restoration algorithm. It (provably) approximates the optimal estimator that minimizes the Mean Squared Error (MSE) under a perfect perceptual quality constraint.
High-performance cloud for developers.
The serverless platform for teams. Run generative AI models, large-scale batch jobs, job queues, and much more.
Bring your own code — we run the infrastructure.
Modal is a serverless platform for data and AI/ML engineers to run and deploy code in the cloud without having to think about infrastructure. Run generative AI models, large-scale batch jobs, job queues, and more, all faster than ever before.
Rank LLMs, RAG systems, and prompts using automated judge evaluation.
Video on command. Pika puts creative control where it belongs: in your hands.
Pika is a image to video generator using ai.
Framework for orchestrating role-playing, autonomous AI agents. By fostering collaborative intelligence, CrewAI empowers agents to work together seamlessly, tackling complex tasks.
An Open-Source Programming Framework for Agentic AI
Data Framework for LLM Applications.
LlamaIndex (GPT Index) is a data framework for your LLM application. Building with LlamaIndex typically involves working with LlamaIndex core and a chosen set of integrations (or plugins). There are two ways to start building with LlamaIndex in Python:
An open-source RAG-based tool for chatting with your documents.
An open-source clean & customizable RAG UI for chatting with your documents. Built with both end users and developers in mind.
🐙 Guides, papers, lecture, notebooks and resources for prompt engineering.
Prompt engineering is a relatively new discipline for developing and optimizing prompts to efficiently use language models (LMs) for a wide variety of applications and research topics. Prompt engineering skills help to better understand the capabilities and limitations of large language models (LLMs).