llm
AI-Powered Mobile App Builder.
Transform your ideas into polished mobile applications with our intuitive mobile app generator. No complex setup required.
A non-saturating, open-ended environment for evaluating LLMs in Factorio.
Large Language Models (LLMs) are rapidly saturating existing benchmarks, necessitating new open-ended evaluations. We introduce the Factorio Learning Environment (FLE), based on the game of Factorio, that tests agents in long-term planning, program synthesis, and resource optimization.
A native macOS app that allows users to chat with a local LLM that can respond with information from files, folders and websites on your Mac without installing any other software. Powered by llama.cpp.
Chat with an local LLM that can respond with information from your files, folders and websites on your Mac without installing any other software. All conversations happen offline, and your data is saved locally.
Affiche une alerte sur des sites générés tout ou partie par des IA. Liste gérée par Next.ink.
Related contents:
Evolving agents is a production-grade environment for orchestrating, evolving, and managing AI agents.
A production-grade framework for creating, managing, and evolving AI agents with intelligent agent-to-agent communication. The framework enables you to build collaborative agent ecosystems that can semantically understand requirements, evolve based on past experiences, and communicate effectively to solve complex tasks.
An in-depth book and reference on building agentic systems like Claude Code. A deep-dive guide into architecture patterns for building responsive, reliable AI coding agents.
There's been a lot of asking about how Claude Code works under the hood. Usually, people see the prompts, but they don't see how it all comes together. This is that book. All of the systems, tools, and commands that go into building one of these.
A practical deep dive and code review into how to build a self-driving coding agent, execution engine, tools and commands. Rather than the prompts and AI engineering, this is the systems and design decisions that go into making agents that are real-time, self-corrective, and useful for productive work.
Terminal-based AI coding tool that can use any model that supports the OpenAI-style API.
The Platform for Building Stateful Agents. Build agents with infinite context and human-like memory, that can learn from data and improve with experience. Letta (formerly MemGPT) is a framework for creating LLM services with memory.
👾 Letta is an open source framework for building stateful LLM applications. You can use Letta to build stateful agents with advanced reasoning capabilities and transparent long-term memory. The Letta framework is white box and model-agnostic.
Related contents:
Finding the Scaling Laws of Agents. The first and the best multi-agent framework.
🐫 CAMEL is an open-source community dedicated to finding the scaling laws of agents. We believe that studying these agents on a large scale offers valuable insights into their behaviors, capabilities, and potential risks. To facilitate research in this field, we implement and support various types of agents, tasks, prompts, models, and simulated environments.
The framework enables multi-agent systems to continuously evolve by generating data and interacting with environments. This evolution can be driven by reinforcement learning with verifiable rewards or supervised learning.
Ollama Automated Security Intelligence Scanner.
🛡️ An AI-powered security auditing tool that leverages Ollama models to detect and analyze potential security vulnerabilities in your code.
Advanced code security analysis through the power of AI
Related contents:
Go beyond nascent AI demos. The intelligent AI-native gateway for prompts and agentic apps.
Effortlessly build AI apps that can answer questions and help users get things done. Arch is the AI-native proxy that handles the pesky heavy-lifting so that you can move faster in building agentic apps, prevent harmful outcomes, and rapidly incorporate latest models.
AI-native (edge and LLM) proxy for agents. Move faster by letting Arch handle the pesky heavy lifting in building agentic apps -- ⚡️ query understanding and routing, seamless integration of prompts with tools, and unified access and observability of LLMs. Built by the contributors of Envoy proxy.
Prompt, run, edit, and deploy full-stack web applications using any LLM you want!
Nut is an open source fork of Bolt.new for helping you develop full stack apps using AI. AI developers frequently struggle with fixing even simple bugs when they don't know the cause, and get stuck making ineffective changes over and over. We want to crack these tough nuts, so to speak, so you can get back to building.
Opsy - Your AI-Powered SRE Colleague.
Opsy is an intelligent command-line assistant designed for Site Reliability Engineers (SREs), DevOps professionals, and platform engineers. It uses AI to help you navigate operational challenges, troubleshoot issues, and automate routine workflows. Opsy integrates with your existing tools and provides contextual assistance to make your daily operations more efficient.
The agents.json Specification is an open specification that formally describes contracts for API and agent interactions, built on top of the OpenAPI standard.
⚔️ Chatbot Arena (formerly LMSYS): Free AI Chat to Compare & Test Best AI Chatbots.
Related contents:
A command-line Ollama client for scripting.
The Ollama function caller, otherwise known as ofc, is a command-line tool for prompting Ollama models locally on your system. There are other programs out there that do similar things, but they either don't support streaming or don't give me access to important settings, like context length or temperature.
Plauna is a privacy focused service that helps you categorize your e-mails. All of its data is stored only on your computer.
You can parse your e-mails from mbox files or directly by connecting to your IMAP servers. Plauna helps you categorize your e-mails and automates the process so that incoming e-mails are moved to their respective folders on your IMAP server.
LLM playground to experiment with local models and build fine-tuning datasets and benchmarks.
A playground that gives you full control over the contents of a chat conversation: add, remove and edit messages (system, user and assistant) and shape the flow of the conversation to be exactly what you need.
Revolutionize Your Development with Windsurf AI IDE
As the first truly intelligent development environment, Codeium Windsurf AI combines powerful AI agents with intuitive co-pilots to create an unparalleled coding experience that adapts to your needs and enhances your productivity.
Related contents:
AI you can trust
Easily summarize emails, docs, articles, and videos across the web — without sacrificing your privacy
Related contents:
Generate a Markdown Document of Your Codebase Structure and Content.
CodeWeaver is a command-line tool designed to weave your codebase into a single, easy-to-navigate Markdown document. It recursively scans a directory, generating a structured representation of your project's file hierarchy and embedding the content of each file within code blocks. This tool simplifies codebase sharing, documentation, and integration with AI/ML code analysis tools by providing a consolidated and readable Markdown output.
Cost-efficient and pluggable Infrastructure components for GenAI inference. AIBrix is a open source cost-efficient, scalable and pluggable Infrastructure for GenAI applications.
AIBrix is an open-source initiative designed to provide essential building blocks to construct scalable GenAI inference infrastructure. AIBrix delivers a cloud-native solution optimized for deploying, managing, and scaling large language model (LLM) inference, tailored specifically to enterprise needs.
AntiSquat leverages AI techniques such as natural language processing (NLP), large language models (ChatGPT) and more to empower detection of typosquatting and phishing domains.
Related contents:
Turn any Git repository into a simple text digest of its codebase. Replace 'hub' with 'ingest' in any github url to get a prompt-friendly extract of a codebase.
This is useful for feeding a codebase into any LLM.
Related contents:
curated list of resources for AI Engineering.
Related contents:
Write Go, online, with LLMs.
sketch.dev is a new kind of Go playground that lets you write and run Go code in your browser, in a real editor. Build packages and modules, import dependencies, run tests, code with chat, and use a lightweight unix sandbox.
Related contents:
The Open-Source LLM Evaluation Framework.
DeepEval is a simple-to-use, open-source LLM evaluation framework, for evaluating and testing large-language model systems. It is similar to Pytest but specialized for unit testing LLM outputs. DeepEval incorporates the latest research to evaluate LLM outputs based on metrics such as G-Eval, hallucination, answer relevancy, RAGAS, etc., which uses LLMs and various other NLP models that runs locally on your machine for evaluation.
The TypeScript framework for agents & workflows with react-like components. Lightning fast dev loop. Easy to learn. Easy to extend.
Build complex AI applications with React-like components. GenSX is a simple typescript framework for building agents and workflows with reusable React-like components. GenSX takes a lot of inspiration from React, but the programming model is very different - it’s a Node.js framework designed for data flow.
Neon Core extends Mycroft core with more modular code, extended multi-user support, and more.
Neon AI is an open source voice assistant.
🌟 OpenVoiceOS is an open-source platform for smart speakers and other voice-centric devices.
OpenVoiceOS is a community-driven, open-source voice AI platform for creating custom voice-controlled interfaces across devices with NLP, a customizable UI, and a focus on privacy and security.
Connect home devices into a powerful cluster to accelerate LLM inference. More devices mean faster performance, leveraging tensor parallelism and high-speed synchronization over Ethernet.
Supports Linux, macOS, and Windows. Optimized for ARM and x86_64 AVX2 CPUs.
Related contents:
R1 1776 is a DeepSeek-R1 reasoning model that has been post-trained by Perplexity AI to remove Chinese Communist Party censorship. The model provides unbiased, accurate, and factual information while maintaining high reasoning capabilities.
Related content:
Local LLM-Powered Coding Assistant for Vim.
VimLM is a Vim plugin that provides an LLM-powered assistant for code editing by allowing users to interact with a local LLM model through Vim commands and automatically ingesting code context.
Scira (Formerly MiniPerplx) is a minimalistic AI-powered search engine that helps you find information on the internet. Powered by Vercel AI SDK! Search with models like Grok 2.0.
Data processing with ML, LLM and Vision LLM.
Sparrow is an innovative open-source solution for efficient data extraction and processing from various documents and images. It seamlessly handles forms, bank statements, invoices, receipts, and other unstructured data sources. Sparrow stands out with its modular architecture, offering independent services and pipelines all optimized for robust performance.
Related contents:
A real-time chat application integrated with N8N workflows for automated responses and processing.
Retrieve, aggregate, filter, evaluate, rewrite and serve RSS feeds using Large Language Models for fun, research and learning purposes.
UglyFeed is a simple application designed to retrieve, aggregate, filter, rewrite, evaluate and serve content (RSS feeds) written by a large language model.
🪄 Create rich visualizations with AI
Data Formulator is an application from Microsoft Research that uses large language models to transform data, expediting the practice of data visualization.
Data Formulator is an AI-powered tool for analysts to iteratively create rich visualizations. Unlike most chat-based AI tools where users need to describe everything in natural language, Data Formulator combines user interface interactions (UI) and natural language (NL) inputs for easier interaction. This blended approach makes it easier for users to describe their chart designs while delegating data transformation to AI.
LLM Frontend for Power Users.
SillyTavern provides a single unified interface for many LLM APIs (KoboldAI/CPP, Horde, NovelAI, Ooba, Tabby, OpenAI, OpenRouter, Claude, Mistral and more), a mobile-friendly layout, Visual Novel Mode, Automatic1111 & ComfyUI API image generation integration, TTS, WorldInfo (lorebooks), customizable UI, auto-translate, more prompt options than you'd ever want or need, and endless growth potential via third-party extensions.
Your Novel Writing Toolbox. Write, without restriction.
Novelcrafter's got your back, arming you with the tools you need to dream, draft, and deliver your story.
build ml models in natural language and minimal code.
Create machine learning models with minimal code by describing what you want them to do in plain words. You explain the task, and the library builds a model for you, including data generation, feature engineering, training, and packaging.
OpenAI just published Deep Research, a very nice assistant that can perform deep searches on the web to answer user questions.
However, their agent has a huge downside: it's not open. So we've started a 24-hour rush to replicate and open-source it. Our resulting open-Deep-Research agent took the #1 rank of any open submission on the GAIA leaderboard! ✨
Related contents:
For better or for worse, LLMs are here to stay. We all read content that they produce online, most of us interact with LLM chatbots, and many of us use them to produce content of our own.
In a series of five- to ten-minute lessons, we will explain what these machines are, how they work, and how to thrive in a world where they are everywhere.
You will learn when these systems can save you a lot of time and effort. You will learn when they are likely to steer you wrong. And you will discover how to see through the hype to tell the difference. ?
LLM Gateway to provide model access, logging and usage tracking across 100+ LLMs. All in the OpenAI format.
Related contents:
The World's Leading AI Platform for Enterprise. The all-in-one platform for private and secure AI.
Cohere brings you cutting-edge multilingual models, advanced retrieval, and an AI workspace tailored for the modern enterprise — all within a single, secure platform.
Related contents:
LLM inference in C/C++.
The main goal of llama.cpp is to enable LLM inference with minimal setup and state-of-the-art performance on a wide range of hardware - locally and in the cloud.
Related contents:
Never Forget Your Achievements. Get recognized for all your hard work.
Bragdoc helps you get recognized for your hard work by tracking your achievements and creating beautiful documents to share with your boss.
A Systems View of LLMs on TPUs.
This book aims to demystify the art of scaling LLMs on TPUs. We try to explain how TPUs work, how LLMs actually run at scale, and how to pick parallelism schemes during training and inference that avoid communication bottlenecks.
Simple, secure, and reproducible packaging for AI/ML projects.
KitOps is an open source DevOps tool that packages and versions your AI/ML model, datasets, code, and configuration into a reproducible artifact called a ModelKit. ModelKits are built on existing standards, ensuring compatibility with the tools your data scientists and developers already use.
The Cloud Built for AI.
Train, fine-tune and deploy AI models with RunPod.
Globally distributed GPU cloud for your AI workloads. Deploy any GPU workload seamlessly, so you can focus less on infrastructure and more on running ML models.
Rent GPUs.
Vast.ai is the market leader in low-cost cloud GPU rental. Use one simple interface to save 5-6X on GPU compute.
Secure & reliable LLMs. Test & secure your LLM apps. Open-source LLM testing used by 51,000+ developers.
Test your prompts, agents, and RAGs. Red teaming, pentesting, and vulnerability scanning for LLMs. Compare performance of GPT, Claude, Gemini, Llama, and more. Simple declarative configs with command line and CI/CD integration.
Related contents:
Open Universal Machine Intellingence. E2E Foundation Model Research Platform. Everything you need to build state-of-the-art foundation models, end-to-end.
Oumi is a fully open-source platform that streamlines the entire lifecycle of foundation models - from data preparation and training to evaluation and deployment. Whether you're developing on a laptop, launching large scale experiments on a cluster, or deploying models in production, Oumi provides the tools and workflows you need.
A short introduction to RLHF and post-training focused on language models.
llamafile lets you distribute and run LLMs with a single file.
Our goal is to make open LLMs much more accessible to both developers and end users. We're doing that by combining llama.cpp with Cosmopolitan Libc into one framework that collapses all the complexity of LLMs down to a single-file executable (called a "llamafile") that runs locally on most computers, with no installation.
Use your locally running AI models to assist you in your web browsing.
Page Assist is an open-source browser extension that provides a sidebar and web UI for your local AI model. It allows you to interact with your model from any webpage.
Your Open-Source Personal Assistant.
Leon is your open-source personal assistant who can live on your server. He does stuff when you ask him to.
Enable AI to control your browser. Make websites accessible for AI agents.
We make websites accessible for AI agents by extracting all interactive elements, so agents can focus on what makes their beer taste better.
Related contents: