Stars
Model soups: averaging weights of multiple fine-tuned models improves accuracy without increasing inference time
Personal Project: MPP-Qwen14B & MPP-Qwen-Next(Multimodal Pipeline Parallel based on Qwen-LM). Support [video/image/multi-image] {sft/conversations}. Don't let the poverty limit your imagination! Tr…
4 bits quantization of LLaMA using GPTQ
A large-scale 7B pretraining language model developed by BaiChuan-Inc.
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
Implementation of the LLaMA language model based on nanoGPT. Supports flash attention, Int8 and GPTQ 4bit quantization, LoRA and LLaMA-Adapter fine-tuning, pre-training. Apache 2.0-licensed.
中文LLaMA&Alpaca大语言模型+本地CPU/GPU训练部署 (Chinese LLaMA & Alpaca LLMs)
Let ChatGPT teach your own chatbot in hours with a single GPU!
Making large AI models cheaper, faster and more accessible
CLIP (Contrastive Language-Image Pretraining), Predict the most relevant text snippet given an image
PyTorch code for EMNLP 2019 paper "LXMERT: Learning Cross-Modality Encoder Representations from Transformers".
Code for the ICML 2021 (long talk) paper: "ViLT: Vision-and-Language Transformer Without Convolution or Region Supervision"
本项目将《动手学深度学习》(Dive into Deep Learning)原书中的MXNet实现改为PyTorch实现。
Encapsulate server requests using react custom hooks
PlaidML is a framework for making deep learning work everywhere.
This project reproduces the book Dive Into Deep Learning (https://d2l.ai/), adapting the code from MXNet into PyTorch.
Awesome multilingual OCR toolkits based on PaddlePaddle (practical ultra lightweight OCR system, support 80+ languages recognition, provide data annotation and synthesis tools, support training and…
TensorFlow code and pre-trained models for BERT
中文命名实体识别,实体抽取,tensorflow,pytorch,BiLSTM+CRF
Pytorch-Named-Entity-Recognition-with-BERT
A fast and accurate POS and morphological tagging toolkit (EACL 2014)