[go: nahoru, domu]

Skip to content
View whuaxiom's full-sized avatar
Block or Report

Block or report whuaxiom

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Please don't include any personal information such as legal names or email addresses. Maximum 100 characters, markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
Beta Lists are currently in beta. Share feedback and report bugs.

Starred repositories

Showing results

Repository for Meta Chameleon, a mixed-modal early-fusion foundation model from FAIR.

Python 1,360 87 Updated Jun 21, 2024

YaFSDP: Yet another Fully Sharded Data Parallel

Python 737 34 Updated Jun 27, 2024

(ICML 2024) Alphazero-like Tree-Search can guide large language model decoding and training

Python 131 14 Updated May 26, 2024

[CVPR 2023 Best Paper] Planning-oriented Autonomous Driving

Python 3,061 325 Updated Mar 19, 2024

Repository hosting code used to reproduce results in "Actions Speak Louder than Words: Trillion-Parameter Sequential Transducers for Generative Recommendations" (https://arxiv.org/abs/2402.17152, I…

Python 474 78 Updated Jun 30, 2024

Ongoing research training transformer models at scale

Python 9,238 2,086 Updated Jun 27, 2024

An efficient, flexible and full-featured toolkit for fine-tuning LLM (InternLM2, Llama3, Phi3, Qwen, Mistral, ...)

Python 3,154 255 Updated Jun 27, 2024

This project aim to reproduce Sora (Open AI T2V model), we wish the open source community contribute to this project.

Python 10,848 969 Updated Jun 27, 2024
Python 225 37 Updated Jul 1, 2024

ring-attention experiments

Python 75 10 Updated Apr 10, 2024

Hackable and optimized Transformers building blocks, supporting a composable construction.

Python 7,961 564 Updated Jun 27, 2024

Fast and Easy Infinite Neural Networks in Python

Jupyter Notebook 2,250 228 Updated Mar 1, 2024

Implementation of 💍 Ring Attention, from Liu et al. at Berkeley AI, in Pytorch

Python 397 22 Updated Apr 20, 2024

RAFT contains fundamental widely-used algorithms and primitives for machine learning and information retrieval. The algorithms are CUDA-accelerated and form building blocks for more easily writing …

Cuda 666 178 Updated Jul 1, 2024

Transformers with Arbitrarily Large Context

Python 560 43 Updated Jun 22, 2024

📖A curated list of Awesome LLM Inference Paper with codes, TensorRT-LLM, vLLM, streaming-llm, AWQ, SmoothQuant, WINT8/4, Continuous Batching, FlashAttention, PagedAttention etc.

1,848 129 Updated Jun 30, 2024

Code examples and resources for DBRX, a large language model developed by Databricks

Python 2,473 231 Updated May 1, 2024

HPT - Open Multimodal LLMs from HyperGAI

Python 301 14 Updated Jun 6, 2024

High performance distributed framework for training deep learning recommendation models based on PyTorch.

Rust 388 51 Updated Feb 7, 2024

The official PyTorch implementation of Google's Gemma models

Python 5,124 487 Updated Jun 30, 2024

The TinyLlama project is an open endeavor to pretrain a 1.1B Llama model on 3 trillion tokens.

Python 7,272 421 Updated May 3, 2024

The AI-native database built for LLM applications, providing incredibly fast full-text and vector search

C++ 2,025 171 Updated Jul 1, 2024

Ongoing research training transformer language models at scale, including: BERT & GPT-2

Python 1,712 329 Updated Jun 28, 2024

A primitive library for neural network

C++ 1,239 209 Updated Jun 21, 2024

Universal LLM Deployment Engine with ML Compilation

Python 17,672 1,405 Updated Jun 27, 2024

FlashInfer: Kernel Library for LLM Serving

Cuda 747 61 Updated Jun 30, 2024

SGLang is a structured generation language designed for large language models (LLMs). It makes your interaction with models faster and more controllable.

Python 2,749 176 Updated Jul 1, 2024
Next