llm
Build Frontier RAG Apps. The open-source RAG platform: built-in citations, deep research, 22+ file formats, partitions, MCP server, and more.
Ground AI agents in your knowledge base, minimize hallucinations, and impress out of the box. Agentset is the open-source platform to build, evaluate, and ship production-ready RAG and agentic applications. It provides end-to-end tooling: ingestion, vector indexing, evaluation/benchmarks, chat playground, hosting, and a clean API with first-class developer experience.
Related contents:
Customizable AI Research & Knowledge Management Assistant. The AI Workspace Built for Teams. Connect any LLM to your internal knowledge sources and chat with it in real time alongside your team.
Open Source Alternative to NotebookLM / Perplexity, connected to external sources such as Search Engines, Slack, Linear, Jira, ClickUp, Confluence, Notion, YouTube, GitHub, Discord and more.
Modern columnar data format for ML and LLMs implemented in Rust. Convert from parquet in 2 lines of code for 100x faster random access, vector index, and data versioning. Compatible with Pandas, DuckDB, Polars, Pyarrow, and PyTorch with more integrations coming..
Lance is a modern columnar data format optimized for machine learning and AI applications. It efficiently handles diverse multimodal data types while providing high-performance querying and versioning capabilities.
Related contents:
Your AI.Your Data. Zero Cloud. Offline ChatGPT alternative: open-source, on-device, and 100% private.
By connecting to Ollama local LLMs, NativeMind delivers the latest AI capabilities right inside your favourite browser — without sending a single byte to cloud servers.
Related contents:
The best ChatGPT that $100 can buy.
This repo is a full-stack implementation of an LLM like ChatGPT in a single, clean, minimal, hackable, dependency-lite codebase. nanochat is designed to run on a single 8XH100 node via scripts like speedrun.sh, that run the entire pipeline start to end. This includes tokenization, pretraining, finetuning, evaluation, inference, and web serving over a simple UI so that you can talk to your own LLM just like ChatGPT. nanochat will become the capstone project of the course LLM101n being developed by Eureka Labs.
Related contents:
This course is intended to provide you with a comprehensive step-by-step understanding of how to engineer optimal prompts within Claude.
Artificial Intelligence Infrastructure-as-Code Generator.
aiac is a library and command line tool to generate IaC (Infrastructure as Code) templates, configurations, utilities, queries and more via LLM providers such as OpenAI, Amazon Bedrock and Ollama.
The CLI allows you to ask a model to generate templates for different scenarios (e.g. "get terraform for AWS EC2"). It composes an appropriate request to the selected provider, and stores the resulting code to a file, and/or prints it to standard output.
Fabric is an open-source framework for augmenting humans using AI. It provides a modular system for solving specific problems using a crowdsourced set of AI prompts that can be used anywhere.
Give Claude perfect recall - Redis-powered persistent memory for LLM. Give Claude perfect recall with persistent memory that survives context limits and session restarts.
Recall is a brain extension for Claude that stores memories in Redis.
Your AI assistant can now remember important context, decisions, and patterns across all conversations—no more repeating yourself or losing critical information when the context window fills up.
NOVA: The Prompt Pattern Matching. NOVA is a Python-based tool for detecting adversarial prompts - prompt injections, exfiltration, evasion and TTPs.
NOVA is an open-source prompt pattern matching system combining keyword detection, semantic similarity, and LLM-based evaluation to analyze and detect prompt content.
Related contents:
ChatKit is a framework for building high-quality, AI-powered chat experiences. It’s designed for developers who want to add advanced conversational intelligence to their apps fast—with minimal setup and no reinventing the wheel. ChatKit delivers a complete, production-ready chat interface out of the box.
Jules is an experimental coding agent that helps you fix bugs, add documentation, and build new features. It integrates with GitHub, understands your codebase, and works asynchronously — so you can move on while it handles the task.
Related contents:
Claude Code + Obsidian Starter Kit.
Turn your Obsidian vault into an AI-powered second brain using Claude Code.
Related contents:
Symfony AI is a set of components that integrate AI capabilities into PHP applications.
Related contents:
AI-Assisted Authoring Environment for WordPress Blocks.
Related contents:
The AI Toolkit for TypeScript.
From the creators of Next.js, the AI SDK is a free open-source library that gives you the tools you need to build AI-powered products.
Related contents:
Laravel Boost accelerates AI-assisted development by providing the essential context and structure that AI needs to generate high-quality, Laravel-specific code.
Related contents:
💬 Typebot is a powerful chatbot builder that you can self-host.
Hack the bot game: Build faster, Chat smarter.
Typebot is a no-code platform that enables you to effortlessly create and integrate advanced chatbots into websites and chat platforms like WhatsApp.
The Privacy-First Alternative to Ollama.
⚡ Python-free Rust inference server — OpenAI-API compatible. GGUF + SafeTensors, hot model swap, auto-discovery, single binary. FREE now, FREE forever.
Shimmy is a 5.1MB single-binary that provides 100% OpenAI-compatible endpoints for GGUF models. Point your existing AI tools to Shimmy and they just work — locally, privately, and free.
Prism is a powerful Laravel package for integrating Large Language Models (LLMs) into your applications.
Related contents:
Turn ideas into detections your SIEM understands. Generate Sigma, KQL, and SPL rules with tests and playbooks in seconds.
DetectPack Forge turns plain-English behaviors or sample logs into production-ready detection packs — Sigma, KQL (Sentinel), SPL (Splunk) — plus tests and a response playbook, mapped to MITRE ATT&CK, fully powered by Gen AI.
Generate code from the terminal! Supercharge Your AI Coding. Simple. Fast. Powerful. Codebuff works in your terminal.
Codebuff is an open-source AI coding assistant that edits your codebase through natural language instructions. Instead of using one model for everything, it coordinates specialized agents that work together to understand your project and make precise changes.
A sane distributed file system.
9P is a network protocol developed at Bell Labs for the Plan 9 from Bell Labs distributed operating system as the means of accessing and manipulating resources and applications transparently in a distributed environment. 9P works both as a distributed file system and as a network transparent and language agnostic ‘API’.
The Filesystem That Makes S3 your Primary Storage. ZeroFS is 9P/NFS/NBD on top of S3.
ZeroFS makes S3 storage feel like a real filesystem. It provides file-level access via NFS and 9P and block-level access via NBD. Fast enough to compile code on, with clients already built into your OS. No FUSE drivers, no kernel modules, just mount and go.
Background Agents for Claude Code.
Terragon is a developer platform that lets you run AI coding agents in the cloud. Whether you're at your browser, in your terminal, or on your phone, Terragon works wherever you are. Offload coding tasks and come back to pull requests ready to review.
A visualization and walkthrough of the LLM algorithm that backs OpenAI's ChatGPT. Explore the algorithm down to every add & multiply, seeing the whole process in action.
This project displays a 3D model of a working implementation of a GPT-style network. That is, the network topology that's used in OpenAI's GPT-2, GPT-3, (and maybe GPT-4).
A tool to perform transformations on code repositories with AI.
Related contents:
Collection of extracted System Prompts from popular chatbots like ChatGPT, Claude & Gemini
A free and open documentation platform built with Laravel and Filament, enhanced by Ollama for local AI features, focused on clarity, structure, and self-hosted simplicity.
Servers and tools from the community that connect models to files, APIs, databases, and more.
Challenging Software Optimization Tasks for Evaluating SWE-Agents.
A benchmark for evaluating language models' capabilities in developing high-performance software.
GSO (Global Software Optimization) is a benchmark for evaluating language models' capabilities in developing high-performance software. We present 100+ challenging optimization tasks across 10 codebases spanning diverse domains and programming languages. Each task provides a codebase and performance test as a precise specification, with agents required to optmize the codebase and measured against expert developer commits.
System Optimization through Reflective Text Evolution. Optimize prompts, code, and more with AI-powered Reflective Text Evolution.
GEPA (Genetic-Pareto) is a framework for optimizing arbitrary systems composed of text components—like AI prompts, code snippets, or textual specs—against any evaluation metric. It employs LLMs to reflect on system behavior, using feedback from execution and evaluation traces to drive targeted improvements. Through iterative mutation, reflection, and Pareto-aware candidate selection, GEPA evolves robust, high-performing variants with minimal evaluations, co-evolving multiple components in modular systems for domain-specific gains.
Related contents:
The framework for programming—not prompting—language models
DSPy is a declarative framework for building modular AI software. It allows you to iterate fast on structured code, rather than brittle strings, and offers algorithms that compile AI programs into effective prompts and weights for your language models, whether you're building simple classifiers, sophisticated RAG pipelines, or Agent loops.
Related contents:
A framework for building AI agents and applications.
AutoGen is a framework for creating multi-agent AI applications that can act autonomously or work alongside humans.
Related contents:
A modern open source language for analyzing, transforming, and modeling data.
Malloy is a modern open source language for describing data relationships and transformations. It is both a semantic modeling language and a query language that uses an existing SQL engine to execute queries. Malloy currently can connect to BigQuery, Snowflake, PostgreSQL, MySQL, Trino, or Presto, and natively supports DuckDB. We've built a Visual Studio Code extension to facilitate building Malloy data models, querying and transforming data, and creating simple visualizations and dashboards.
Related contents:
A proposal to standardise on using an /llms.txt file to provide information to help LLMs use a website at inference time.
We propose adding a /llms.txt markdown file to websites to provide LLM-friendly content. This file offers brief background information, guidance, and links to detailed markdown files.
llms.txt markdown is human and LLM readable, but is also in a precise format allowing fixed processing methods (i.e. classical programming techniques such as parsers and regex).
Related contents:
AI Orchestration Platform
🌊 The leading agent orchestration platform for Claude. Deploy intelligent multi-agent swarms, coordinate autonomous workflows, and build conversational AI systems. Features enterprise-grade architecture, distributed swarm intelligence, RAG integration, and native Claude Code support via MCP protocol. Ranked #1 in agent-based frameworks.
An open-source framework for building AI-powered apps, built and used in production by Google
Open-source framework for building AI-powered apps in JavaScript, Go, and Python, built and used in production by Google
It offers a unified interface for integrating AI models from providers like Google, OpenAI, Anthropic, Ollama, and more. Rapidly build and deploy production-ready chatbots, automations, and recommendation systems using streamlined APIs for multimodal content, structured outputs, tool calling, and agentic workflows.
Related contents:
Chronon is a data platform for serving for AI/ML applications.
Chronon is a platform that abstracts away the complexity of data computation and serving for AI/ML applications. Users define features as transformation of raw data, then Chronon can perform batch and streaming computation, scalable backfills, low-latency serving, guaranteed correctness and consistency, as well as a host of observability and monitoring tools.
It allows you to utilize all of the data within your organization, from batch tables, event streams or services to power your AI/ML projects, without needing to worry about all the complex orchestration that this would usually entail.
Semantic Data Processing. Build data processing and data analysis pipelines that leverage the power of LLMs 🧠
Semlib is a Python library for building data processing and data analysis pipelines that leverage the power of large language models (LLMs). Semlib provides, as building blocks, familiar functional programming primitives like map, reduce, sort, and filter, but with a twist: Semlib's implementation of these operations are programmed with natural language descriptions rather than code. Under the hood, Semlib handles complexities such as prompting, parsing, concurrency control, caching, and cost tracking.
☁️ Convert any site to clean markdown & llms.txt. Boost your site's AI discoverability or generate LLM context for a project you're working with.
Mdream core is a highly optimized primitive for producing Markdown from HTML that is optimized for LLMs.
Chat Gipitty (Chat Get Information, Print Information TTY) is a command line client primarily intended for the official OpenAI Chat Completions API. It allows you to chat with language models in a terminal and even pipe output into it. While optimized for OpenAI's ChatGPT (with GPT-4 as the default model), it can also work with other providers that expose OpenAI-compatible endpoints.
Related contents:
Your personal AI productivity tool for a smarter brain.
🤯 Lobe Chat - an open-source, modern design AI chat framework. Supports multiple AI providers (OpenAI / Claude 4 / Gemini / DeepSeek / Ollama / Qwen), Knowledge Base (file upload / RAG ), one click install MCP Marketplace and Artifacts / Thinking. One-click FREE deployment of your private AI Agent application.
A powerful coding agent toolkit providing semantic retrieval and editing capabilities (MCP server & other integrations).
🚀 Serena is a powerful coding agent toolkit capable of turning an LLM into a fully-featured agent that works directly on your codebase. Unlike most other tools, it is not tied to an LLM, framework or an interface, making it easy to use it in a variety of ways.
Your best engineers shouldn't Q&A all day.
Dosu is an AI-powered Github maintainer agent. Dosu helps open source maintainers encourage contributors, organize issues, triage bugs, answer questions, and more.
Dosu turns your codebase into a living knowledge base that every team member can understand—from engineering to operations—so builders can build.
Related contents:
EPFL, ETH Zurich, and the Swiss National Supercomputing Centre (CSCS) has released Apertus, Switzerland’s first large-scale open, multilingual language model — a milestone in generative AI for transparency and diversity. Trained on 15 trillion tokens across more than 1,000 languages – 40% of the data is non-English – Apertus includes many languages that have so far been underrepresented in LLMs, such as Swiss German, Romansh, and many others. Apertus serves as a building block for developers and organizations for future applications such as chatbots, translation systems, or educational tools.
Related contents:
Qwen Code is a coding agent that lives in the digital world.
Qwen Code brings the capabilities of advanced code models to your terminal in an interactive Read-Eval-Print Loop (REPL) environment. Qwen Code consists of a client-side application (packages/cli) that communicates with a local server (packages/core). Qwen Code also contains a variety of tools for tasks such as performing file system operations, running shells, and web fetching, which are managed by packages/core.
Related contents:
Open-source LLMOps platform for hosting and scaling AI in your own infrastructure 🏓🦙 Paddler is an open-source LLMOps platform that lets teams run inference and deploy LLMs on their own infrastructure.
Tensors and Dynamic neural networks in Python with strong GPU acceleration.
PyTorch is a Python package that provides two high-level features:
- Tensor computation (like NumPy) with strong GPU acceleration
- Deep neural networks built on a tape-based autograd system
You can reuse your favorite Python packages such as NumPy, SciPy, and Cython to extend PyTorch when needed.
Related contents:
Call all LLM APIs using the OpenAI format [Bedrock, Huggingface, VertexAI, TogetherAI, Azure, OpenAI, Groq etc.]
Python SDK, Proxy Server (LLM Gateway) to call 100+ LLM APIs in OpenAI format - [Bedrock, Azure, OpenAI, VertexAI, Cohere, Anthropic, Sagemaker, HuggingFace, Replicate, Groq]
Related contents:
Single-Binary Kubernetes Dashboard with Multi-Cluster Management & AI Integration. (OpenAI / Claude 4 / Gemini / DeepSeek / OpenRouter / Ollama / Qwen / LMStudio).
It provides a simple and rich realtime interface to manage and investigate your clusters.
Advanced LLM-powered brute-force tool combining AI intelligence with automated login attacks.
Related contents:
Build compliant AI chat agents, in minutes. LLM agents built for control. Designed for real-world use. Deployed in minutes.
Parlant gives you all the structure you need to build customer-facing agents that behave exactly as your business requires
An honest CLI for openrouter.ai. Fast and tmux-aware. ort sends your prompts to AI models on openrouter.ai.
It is built the old fashioned way, in solid Rust. It doesn't slow you down with Python interpreters. This is a modest 2 MiB ELF binary.
Related contents:
The huggingface_hub library allows you to interact with the Hugging Face Hub, a machine learning platform for creators and collaborators. Discover pre-trained models and datasets for your projects or play with the hundreds of machine learning apps hosted on the Hub. You can also create and share your own models and datasets with the community. The huggingface_hub library provides a simple way to do all these things with Python.
Related contents:
Render any git repo into a single static HTML page for humans or LLMs.
Tired of clicking around complex file hierarchies of GitHub repos? Do you just want to see all of the code on a single page? Enter rendergit. Flatten any GitHub repository into a single, static HTML page with syntax highlighting, markdown rendering, and a clean sidebar navigation. Perfect for code review, exploration, and an instant Ctrl+F experience.
Build / Deploy Agent Workflows.
Launch agentic workflows with an open source, user-friendly environment for devs and agents.
Sim is an open-source AI agent workflow builder. Sim Studio's interface is a lightweight, intuitive way to quickly build and deploy LLMs that connect with your favorite tools.
A simple, open format for guiding coding agents, used by over 20k open-source projects.
Think of AGENTS.md as a README for agents: a dedicated, predictable place to provide the context and instructions to help AI coding agents work on your project.
Related contents:
The Elegant Desktop Companion for Claude Code.
A powerful GUI app and Toolkit for Claude Code - Create custom agents, manage interactive Claude Code sessions, run secure background agents, and more.
Related contents: