๐ค HF Repo โข ๐ join our WeChat โข ๐ Demo
๐LLaMA2-Accessory is an open-source toolkit for pre-training, fine-tuning and deployment of Large Language Models (LLMs) and multimodal LLMs. This repo is mainly inherited from LLaMA-Adapter with more advanced features.๐ง
โจWithin this toolkit, we present SPHINX, a versatile multimodal large language model (MLLM) that combines a diverse array of training tasks, data domains, and visual embeddings.
- [2023-11-17] We release SPHINX-V2, featuring the same architecture but with enhanced and broader capabilities! ๐ฅ๐ฅ๐ฅ
- [2023.10.17] We release the demo, code, and model of SPHINX!๐ฅ๐ฅ๐ฅ
- [2023.09.15] We now support Falcon 180B!๐ฅ๐ฅ๐ฅ
- [2023.09.14] WeMix-LLaMA2-70B shows excellent performance on the OpenCompass benchmark!๐ฅ๐ฅ๐ฅ
- [2023.09.02] We now support InternLM๐ฅ๐ฅ๐ฅ
- [2023.08.28] We release quantized LLM with OmniQuant, which is an efficient, accurate, and omnibearing (even extremely low bit) quantization algorithm. Multimodal version is coming soon๐ฅ๐ฅ
- [2023.08.27] We now support CodeLLaMA and instruction fine-tuning on evol-code-alpaca๐ฅ๐ฅ
- [2023.08.27] We release our documentation in a webbook format ๐Check it out here
- [2023.08.21] We release the Quantization codes and Evaluation result๐ฅ
- [2023.08.05] We release the multimodel fine-tuning codes and checkpoints๐ฅ
- [2023.07.23] Initial release ๐
-
๐กSupport More Datasets and Tasks
- ๐ฏ Pre-training with RefinedWeb and StarCoder.
- ๐ Single-modal fine-tuning with Alpaca, ShareGPT, LIMA, WizardLM, Flacuna, Platypus, UltraChat and MOSS.
- ๐ Multi-modal fine-tuning with image-text pairs (LAION, COYO and more), interleaved image-text data (MMC4 and OBELISC) and visual instruction data (LLaVA, Shrika, Bard)
- ๐ง LLM for API Control (GPT4Tools and Gorilla).
-
โกEfficient Optimization and Deployment
- ๐ Parameter-efficient fine-tuning with Zero-init Attenion and Bias-norm Tuning.
- ๐ป Fully Sharded Data Parallel (FSDP), Flash Attention 2 and QLoRA.
-
๐๏ธโโ๏ธSupport More Visual Encoders and LLMs
โ๏ธ For environment installation, please refer to Environment Setup.
๐ค Instructions for model pre-training, fine-tuning, inference, and other related topics are all available in the document.
โ Encountering issues or have further questions? Find answers to common inquiries here. We're here to assist you!
- Instruction-tuned LLaMA2: alpaca & gorilla.
- Chatbot LLaMA2: dialog_sharegpt & dialog_lima & llama2-chat.
- Multimodal LLaMA2: in-context & alpacaLlava_llamaQformerv2_13b
- SPHINX: demo
๐ก Now, our model SPHINX supports generating high-quality bounding boxes and then present masks created by SAM for all objects within an image driven by input prompts. Give it a try here! ๐
Chris Liu, Ziyi Lin, Guian Fang, Jiaming Han, Yijiang Liu, Renrui Zhang
Peng Gao, Wenqi Shao, Shanghang Zhang
๐ฅ We are hiring interns, postdocs, and full-time researchers at the General Vision Group, Shanghai AI Lab, with a focus on multi-modality and vision foundation models. If you are interested, please contact gaopengcuhk@gmail.com.
If you find our code and paper useful, please kindly cite:
@article{zhang2023llamaadapter,
title = {LLaMA-Adapter: Efficient Fine-tuning of Language Models with Zero-init Attention},
author={Zhang, Renrui and Han, Jiaming and Liu, Chris and Gao, Peng and Zhou, Aojun and Hu, Xiangfei and Yan, Shilin and Lu, Pan and Li, Hongsheng and Qiao, Yu},
journal={arXiv preprint arXiv:2303.16199},
year={2023}
}
@article{gao2023llamaadapterv2,
title = {LLaMA-Adapter V2: Parameter-Efficient Visual Instruction Model},
author={Gao, Peng and Han, Jiaming and Zhang, Renrui and Lin, Ziyi and Geng, Shijie and Zhou, Aojun and Zhang, Wei and Lu, Pan and He, Conghui and Yue, Xiangyu and Li, Hongsheng and Qiao, Yu},
journal={arXiv preprint arXiv:2304.15010},
year={2023}
}
Show More
- @facebookresearch for ImageBind & LIMA & CodeLlama
- @Instruction-Tuning-with-GPT-4 for GPT-4-LLM
- @tatsu-lab for stanford_alpaca
- @tloen for alpaca-lora
- @lm-sys for FastChat
- @domeccleston for sharegpt
- @karpathy for nanoGPT
- @Dao-AILab for flash-attention
- @NVIDIA for apex & Megatron-LM
- @Vision-CAIR for MiniGPT-4
- @haotian-liu for LLaVA
- @huggingface for peft & OBELISC
- @Lightning-AI for lit-gpt & lit-llama
- @allenai for mmc4
- @StevenGrove for GPT4Tools
- @ShishirPatil for gorilla
- @OpenLMLab for MOSS
- @thunlp for UltraChat
- @LAION-AI for LAION-5B
- @shikras for shikra
- @kakaobrain for coyo-dataset
- @salesforce for LAVIS
- @openai for CLIP
- @bigcode-project for starcoder
- @tiiuae for falcon-refinedweb
- @microsoft for DeepSpeed
- @declare-lab for flacuna
- @nlpxucan for WizardLM
- @arielnlee for Platypus
- @InternLM for InternLM
- @Google for Bard
Llama 2 is licensed under the LLAMA 2 Community License, Copyright (c) Meta Platforms, Inc. All Rights Reserved.