Your all-knowing guide that unpacks every PDF into clear, actionable insights.
Auntie PDF is a web application that helps users extract information and insights from PDF documents. With a sassy, helpful personality, Auntie PDF makes understanding complex documents easier and more engaging.
LLM Frontend for Power Users.
SillyTavern provides a single unified interface for many LLM APIs (KoboldAI/CPP, Horde, NovelAI, Ooba, Tabby, OpenAI, OpenRouter, Claude, Mistral and more), a mobile-friendly layout, Visual Novel Mode, Automatic1111 & ComfyUI API image generation integration, TTS, WorldInfo (lorebooks), customizable UI, auto-translate, more prompt options than you'd ever want or need, and endless growth potential via third-party extensions.
Unlock your LLM's creativity. User-friendly WebUI for LLMs (Formerly Ollama WebUI)
Open WebUI is an extensible, self-hosted interface for AI that adapts to your workflow, all while operating entirely offline.
Open WebUI is an extensible, feature-rich, and user-friendly self-hosted WebUI designed to operate entirely offline. It supports various LLM runners, including Ollama and OpenAI-compatible APIs. For more information, be sure to check out our Open WebUI Documentation.
Related contents:
Every AI in One PlaceBuilt for Everyone
LibreChat is a free, open source AI chat platform. This Web UI offers vast customization, supporting numerous AI providers, services, and integrations. Serves all AI Conversations in one place with a familiar interface, innovative enhancements, for as many users as you need.
Enhanced ChatGPT Clone: Features OpenAI, Assistants API, Azure, Groq, GPT-4 Vision, Mistral, Bing, Anthropic, OpenRouter, Vertex AI, Gemini, AI model switching, message search, langchain, DALL-E-3, ChatGPT Plugins, OpenAI Functions, Secure Multi-User System, Presets, completely open-source for self-hosting. More features in development
Distribute and run LLMs with a single file.
Our goal is to make open LLMs much more accessible to both developers and end users. We're doing that by combining llama.cpp with Cosmopolitan Libc into one framework that collapses all the complexity of LLMs down to a single-file executable (called a "llamafile") that runs locally on most computers, with no installation.
deep learning at the speed of light.
Luminal is a deep learning library that uses composable compilers to achieve high performance.
Finetune AI & LLMs faster.
Unslow AI training & finetuning Get 30x faster with unsloth. 5X faster 60% less memory QLoRA finetuning. Finetune Mistral, Gemma, Llama 2-5x faster with 70% less memory!
Related contents:
Get up and running with large language models, locally.
Run Llama 2, Code Llama, Mistral, Gemma, and other models. Customize and create your own.
Related contents: