- Stanford
- https://twitter.com/karpathy
Highlights
- Pro
Block or Report
Block or report karpathy
Contact GitHub support about this user’s behavior. Learn more about reporting abuse.
Report abuseStars
Language
Sort by: Recently starred
A lightweight library for portable low-level GPU computation using WebGPU.
Simple Byte pair Encoding mechanism used for tokenization process . written purely in C
Official implementation of "Samba: Simple Hybrid State Space Models for Efficient Unlimited Context Language Modeling"
User-friendly WebUI for LLMs (Formerly Ollama WebUI)
Implementation of Diffusion Transformer (DiT) in JAX
Schedule-Free Optimization in PyTorch
SkyPilot: Run LLMs, AI, and Batch jobs on any cloud. Get maximum savings, highest GPU availability, and managed execution—all with a simple interface.
Perplexica is an AI-powered search engine. It is an Open source alternative to Perplexity AI
A minimal GPU design in Verilog to learn how GPUs work from the ground up
lightweight, standalone C++ inference engine for Google's Gemma models.
Distribute and run LLMs with a single file.
A subset of PyTorch's neural network modules, written in Python using OpenAI's Triton.
Fast bare-bones BPE for modern tokenizer training
The official PyTorch implementation of Google's Gemma models
A benchmark to evaluate language models on questions I've previously asked them to solve.
Official implementation for the paper: "Code Generation with AlphaCodium: From Prompt Engineering to Flow Engineering""
A high-throughput and memory-efficient inference and serving engine for LLMs
RuLES: a benchmark for evaluating rule-following in language models
Code and documentation to train Stanford's Alpaca models, and generate the data.
Fine-tune mistral-7B on 3090s, a100s, h100s
Simple and efficient pytorch-native transformer text generation in <1000 LOC of python.
Finetune Llama 3, Mistral, Phi & Gemma LLMs 2-5x faster with 80% less memory
A python script to help manage a Gmail inbox by filtering out promotional emails using GPT-3 or GPT-4.