I created my Digital Palace as a public platform for self-expression, reflection, and sharing my memories. It is where I openly document my thoughts, experiences, milestones, creative works - my personal journey.
My Digital Palace allows me to write journal entries, upload multimedia, record audio, and categorize content by date and themes. The tools help me capture toughts to share publicly.
With my palace being public, anyone can read my entries, view my content, and leave comments to connect with me and others in my community. I enjoy fostering this open dialogue and discussion around my shared stories.
In summary, my Digital Palace provides an engaging platform for public self-expression, reflection, and sharing memories.
The goal of a "Today I Learned" (TIL) journal is to document and reflect upon something new or interesting that you have learned on a given day.
It serves as a personal knowledge base where I can record and reflect on my daily learning experiences.
TIL journals are often used by programmers, developers, and other professionals to share tips, tricks, and lessons learned in their respective fields.
They can also be a valuable resource for others who are looking to learn and expand their knowledge in a specific domain.
š Todays I Learned
The "For the Impatient" tutorial series is designed to provide a fast-paced and efficient way for learners to master programming languages and other technical subjects. The core idea behind this series is to cater to the needs of individuals who are eager to gain proficiency quickly, without getting bogged down in lengthy, comprehensive tutorials.
The typical learner in the "Impatient" target audience might be someone who:
- Has a strong motivation to learn a new skill quickly, perhaps for a time-sensitive project or job requirement
- Prefers a more streamlined, concise learning approach that focuses on the most essential concepts and practical applications
- Wants to get up and running with a new language or technology as rapidly as possible, without spending too much time on theoretical background or peripheral topics
š The impatient series
A mental model is a compression of how something works
- AI Courses
- Prompt engineering
- System, Database, Operating System
- š¤ GenAI Cookbook
- Generative AI for Beginners Microsoft
š Top Quality Blogs
š Top Quality Cloud Providers
š Top Quality Datasets
-
Beyond Prompt Engineering: Modular and Optimized LM Programs with DSPy
-
Mastering the Art of Training Large Language Models from Scratch
- LLMs from scratch Implementing a ChatGPT-like LLM from scratch, step by step by from Sebastian Raschka
- A list of prompt engineering techniques
- Mastering the art of prompt engineering in French
- Mastering the art of prompt engineering
What are the common use cases of LLM Applications ?
Bionic GPT BionicGPT is an on-premise replacement for ChatGPT, offering the advantages of Generative AI while maintaining strict data confidentialityĀ BionicGPT can run on your laptop or scale into the data center. Lobe Chat š¤Æ Lobe Chat - an open-source, modern-design AI chat framework. Supports Multi AI Providers( OpenAI / Claude 3 / Gemini / Ollama / Azure / DeepSeek), Knowledge Base (file upload / knowledge management / RAG ), Multi-Modals (Vision/TTS) and plugin system. One-click FREE deployment of your private ChatGPT/ Claude application.
-
QLLM QLLM: A powerful CLI for seamless interaction with multiple Large Language Models. Simplify AI workflows, streamline development, and unlock the potential of cutting-edge language models. by Quantalogic
-
ell A language model programming framework.
-
LangChain Framework to construct LLMs application documentation š¦
-
LLamaIndexTurn your enterprise data into production-ready LLM applications š¦
-
HuggingFace Model, Datasets, Inference Space, the GitHub of AI models. š¤
-
instructor Instructor makes it easy to reliably get structured data like JSON from Large Language Models (LLMs) like GPT-3.5, GPT-4, GPT-4-Vision, including open source models like Mistral/Mixtral fromĀ Together,Ā Anyscale,Ā Ollama, andĀ llama-cpp-python.
-
instructor_ex (Elixir version of Instructor)
-
marvin Marvin is a lightweight AI toolkit for building natural language interfaces that are reliable, scalable, and easy to trust.
-
Vercel AI SDK (Stream React Components from LLMs to deliver richer user experiences)
-
easyllm EasyLLM is an open source project that provides helpful tools and methods for working with large language models (LLMs), both open source and closed source.
-
Flowneum A toolkit for controllable, private AI on consumer hardware in rust
-
Kalosm Kalosm is an open source framework for private language, audio, and image models in Rust
-
PhiData Phidata is a toolkit for building AI Assistants using function calling.
-
Microsoft AutoDev š§āAutoDev: The AI-powered coding wizard with multilingual support
-
Cognee Memory management for the AI Applications and AI Agents
-
MelloTTS High-quality multi-lingual text-to-speech library by MyShell.ai. Support English, Spanish, French, Chinese, Japanese and Korean.
-
MemGPT Enable Next-Gen Large Language Model Applications
-
lumentis Generate beautiful docs from your transcripts and unstructured information with a single command.
-
Wantdb/OpenUI OpenUI let's you describe UI using your imagination, then see it rendered live.
-
FireCrawl Turn entire websites into LLM-ready markdown
-
Sammo A library for prompt engineering and optimization (SAMMO = Structure-aware Multi-Objective Metaprompt Optimization)
- Plandex An AI coding engine for complex tasks
- [Aider][https://github.com/paul-gauthier/aider] aider is AI pair programming in your terminal
- RagFlow RAGFlow is an open-source RAG (Retrieval-Augmented Generation) engine based on deep document understanding.
- RAG Techniques This repository showcases various advanced techniques for Retrieval-Augmented Generation (RAG) systems. RAG systems combine information retrieval with generative models to provide accurate and contextually rich responses.
- byaldi Use late-interaction multi-modal models such as ColPali in just a few lines of code.
-MindGraph proof of concept prototype for generating and querying against an ever-expanding knowledge graph with ai
- Transformer Head Toolkit for attaching, training, saving and loading of new heads for transformer models
- AICI AICI: Prompts as (Wasm) Programs. Controlling inference using Wasm programs.
- Aimet AIMET is a library that provides advanced quantization and compression techniques for trained neural network models from Qualcomm Innovation Center.
- hqq Official implementation of Half-Quadratic Quantization (HQQ). HQQĀ is a fast and accurate model quantizer that skips the need for calibration data. It's super simple to implement (just a few lines of code for the optimizer). It can crunch through quantizing the Llama2-70B model in only 4 minutes! š
- llms A CLI utility and Python library for interacting with Large Language Models, both via remote APIs and models that can be installed and run on your own machine. By Simon Willison
- Open Interpreter A natural language interface for computers
- Documentation Open Interpreter A new way to use computers
- whisper Whisper is a general-purpose speech recognition model. It is trained on a large dataset of diverse audio and is also a multitasking model that can perform multilingual speech recognition, speech translation, and language identification.
- plock From anywhere you can type, query and stream the output of an LLM or any other script
- Screen shot to code Drop in a screenshot and convert it to clean code (HTML/Tailwind/React/Vue)
- Code2prompt Convert a codebase to an AI prompt
- Claude Engineer Claude Engineer is an interactive command-line interface (CLI) that leverages the power of Anthropic's Claude-3.5-Sonnet model to assist with software development tasks. This tool combines the capabilities of a large language model with practical file system operations and web search functionality.
- Guidance A guidance language for controlling large language models.
- AICI AICI: Prompts as (Wasm) Programs
- Representation Ingeneering Representation Engineering Mistral-7B an Acid Trip š
- VoiceCraft Zero-Shot Speech Editing and Text-to-Speech in the Wild
- Enchanted Enchanted is iOS and macOS app for chatting with private self hosted language models such as Llama2, Mistral or Vicuna using Ollama.
- BoltAI BoltAI is a beautiful and powerful AI chat app for Mac. It integrates ChatGPT and other large language models (LLMs) deeply into your workflows so you donāt have to go back and forth with the ChatGPT web UI.
- llamaC++ LLM inference in C/C++
- vLLM Easy, fast, and cheap LLM serving for everyone, documentation
- nm-vllm A high-throughput and memory-efficient inference and serving engine for LLMs (sparse compressing)
- LLama Cpp Python Binding OpenAI compatible web server
- SkyPilot Run LLMs and AI on Any Cloud documentation
- Ollama Go program that encapsulate llamac++. documentation
- fastassert Dockerized LLM inference server with constrained output (JSON mode), built on top of vLLM and outlines.
- luminal Deep learning at the speed of light coded in Rust. The aim for 0.3 is to achieve SOTA performance on an M1 pro (50 tok/s), and near SOTA on single nvidia gpus (>100 tok/s)
- LLamafile Turning a LLM model into a Multiplatform executable
- candle Minimalist ML framework for Rust. Run and Serve Models in Rust.
- Mlx Server This Python library is the easist way to begin building on top of Apple's machine learning library MLX
- AICI AICI: Prompts as (Wasm) Programs
- List of tools that serves AI locally An awesome repository of local AI tools
- SkypilotSkyPilot is a framework for running LLMs, AI, and batch jobs on any cloud, offering maximum cost savings, highest GPU availability, and managed execution.
- Jan Jan is an open source alternative to ChatGPT that runs 100% offline on your computer
- LoraX Multi-LoRA inference server that scales to 1000s of fine-tuned LLMs (Documentation)
- Candle Minimalist ML framework for Rust from HuggingFace to run models in production.
- ZML High performance AI inference stack. Built for production.Ā @ziglangĀ /Ā @openxlaĀ / MLIR /Ā @bazelbuild
- MLX Omni Server MLX Omni Server is a local inference server powered by Apple's MLX framework, specifically designed for Apple Silicon (M-series) chips. It implements OpenAI-compatible API endpoints, enabling seamless integration with existing OpenAI SDK clients while leveraging the power of local ML inference.
- MegaGPTThe Multi-Agent Framework: Given one line Requirement, return PRD, Design, Tasks, Repo
- CrewAI Framework for orchestrating role-playing, autonomous AI agents. By fostering collaborative intelligence, CrewAI empowers agents to work together seamlessly, tackling complex tasks.
- Devika Devika is an Agentic AI Software Engineer that can understand high-level human instructions, break them down into steps, research relevant information, and write code to achieve the given objective. Devika aims to be a competitive open-source alternative to Devin by Cognition AI.
- AgentStudio AgentStudio is an open toolkit covering the entire lifespan of building virtual agents that can interact with everything on digital worlds
- AIOS AIOS: LLM Agent Operating System
- AutoGen Enable Next-Gen Large Language Model Applications
- Screen Agents ScreenAgent: A Computer Control Agent Driven by Visual Language Large Model
- Synapse Synapse: Trajectory-as-Exemplar Prompting with Memory for Computer Control
- Cradles Towards General Computer Control: A Multimodal Agent for Red Dead Redemption II as a Case Study
- kwaiagents A generalized information-seeking agent system with Large Language Models (LLMs).
- LLocalSearch LLocalSearch is a completely locally running search aggregator using LLM Agents. The user can ask a question and the system will use a chain of LLMs to find the answer. The user can see the progress of the agents and the final answer. No OpenAI or Google API keys are needed.
- SWE Agents SWE-agent: Agent Computer Interfaces Enable Software Engineering Language Models
- Presentation of Agentic Models from Angrew NG
- Zep Long term memory for Agent Zep: Long-Term Memory for āAI Assistants.
- OpenSSA OpenSSA: Small Specialist AgentsāEnabling Efficient, Domain-Specific Planning + Reasoning for AI
- Cognitive Architectures for Language Agents
- š® Awesome AI Agents
- Wikipedia Cognitive Architecture
- Baseline AI Agentic Architecture
- Summoning the Magic of Prompts: A Deep Dive into Prompt Engineering Patterns
- Claude Anthropic prompts Library Explore optimized prompts for a breadth of business and personal tasks
- Navigating the Prompt Engineering Landscape: A Comprehensive Survey for NLP Practitioners
- MLX Swift Examples Examples using MLX Swift
- Gorilla OpenFunction š¦ Gorilla: Large Language Model Connected with Massive APIs
- Models list Open Sources model list
- NousResearch/Nous-Hermes-llama-2-7b
- [Ollama version of Nous-Hermes2Pro][https://ollama.com/adrienbrault/nous-hermes2pro]
- LLavaNext LLaVA-NeXT is recently merged to transformers and it outperforms many of the proprietary models like Gemini on various benchmarks!
- Startling LM 7B Beta The best 7B Models - Finetuned from model:Ā Openchat-3.5-0106Ā (based onĀ Mistral-7B-v0.1) - From Berkeley https://starling.cs.berkeley.edu/
- pgvectorscale A complement to pgvector for high performance, cost efficient vector search on large workloads.
- Berkeley Function-Calling Leaderboard LeaderBoard for function calling
- OpenLLMs Leaderboard š The š¤ Open LLM Leaderboard aims to track, rank and evaluate open LLMs and chatbots.
- HumanEval Leaderboard # š EvalPlus Leaderboard š EvalPlus evaluates AI Coders with rigorous tests
- instructor structured outputs for llms
- fructose LLM calls as strongly-typed functions
- Bonito A lightweight library for generating synthetic instruction tuning datasets for your data without GPT.
- unsloth 5X faster 60% less memory QLoRA finetuning
- Documentation from Premai about Finetuning
- LLama-Factory Unify Efficient Fine-tuning of 100+ LLMs
- peft š¤ PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.
- Tochtune A Native-PyTorch Library for LLM Fine-tuning
- LLMTuner Tune LLM in few lines of code
- LMFlow An Extensible Toolkit for Finetuning and Inference of Large Foundation Models. Large Models for All. (Include Lisa Finetuning)
- # Efficient finetuning of Llama 3 with FSDP QDoRA A blog article that explains how to use the stat of the art QDoRA fine tuning method on LLAMA3
Lighting thunder Source to source compiler for PyTorch. It makes PyTorch programs faster on single accelerators and distributed.
How to document an AI/Machine Learning Model: Improving Transparency in AI through Model Cards
- Useful VSCode extensions
- Open source alternatives to popular products DiscoverĀ Open Source AlternativesĀ to Popular Software
- Fluvio Lean and mean distributed stream processing system written in rust and web assembly. Alternative to Kafka + Flink in one.
- Unstract No-code LLM Platform to launch APIs and ETL Pipelines to structure unstructured documents
- Shot Scraper A command-line utility for taking automated screenshots of websites from Simon Willison
- ttok Count and truncate text based on tokens from Simon Willison
- github2files Create a big files from a GitHub files
Forget the Cloud āļø ā A MacBook Pro with M1/M2 is the Only AI Brain You Need ā¦ š§
mindmap
root((AI))
Applications
LLM Applications
Serving LLMs
Engineering
AI Engineering
Framework
LangChain
LLamaIndex
Vector Databases
Prompt Engineering
Fine Tuning
RAG
Models
AI Models
Embeddings
Ethics
AI and Ethics