ππ Welcome to the repo of SALMONN!
SALMONN is a large language model (LLM) enabling speech, audio events, and music inputs, which is developed by the Department of Electronic Engineering at Tsinghua University and ByteDance. Instead of speech-only input or audio-event-only input, SALMONN can perceive and understand all kinds of audio inputs and therefore obtain emerging capabilities such as multilingual speech recognition and translation and audio-speech co-reasoning. This can be regarded as giving the LLM "ears" and cognitive hearing abilities, which makes SALMONN a step towards hearing-enabled artificial general intelligence.
- [2024-09-04] We have released the model and inference code for video-SALMONN! See here!
- [2024-05-28] 𧳠We have released all the annotations (including 600k SQA/AQA data and 50k audio-based storytelling data) for the 3-stage training of SALMONN! Feel free to download them here!
- [2024-04-07] π€ We have released all the codes you need to train your own SALMONN! Try some cool things!
- [2024-01-16] π Our paper was accepted by ICLR 2024!
- [2023-11-13] π We have released a 7B version of SALMONN at tsinghua-ee/SALMONN-7B and built the 7B demo here!
- [2023-10-08] β¨ We have released the model checkpoint and the inference code for SALMONN-13B!
The model architecture of SALMONN is shown below. A window-level Q-Former is used as the connection module to fuse the outputs from a Whisper speech encoder and a BEATs audio encoder as augmented audio tokens, which are aligned with the LLM input space. The LoRA adaptor aligns the augmented LLM input space with its output space. The text prompt is used to instruct SALMONN to answer open-ended questions about the general audio inputs and the answers are in the LLM text responses.
Compared with traditional speech and audio processing tasks such as speech recognition and audio caption, SALMONN leverages the general knowledge and cognitive abilities of the LLM to achieve a cognitively oriented audio perception, which dramatically improves the versatility of the model and the richness of the task. In addition, SALMONN is able to follow textual commands and even spoken commands with a relatively high degree of accuracy. Since SALMONN only uses training data based on textual commands, listening to spoken commands is also a cross-modal emergent ability.
Here are some examples of SALMONN.
Audio | Response |
---|---|
gunshots.wav | |
duck.wav | |
music.wav |
For SALMONN-13B v1, you need to use the following dependencies:
- Our environment: The python version is 3.9.17, and other required packages can be installed with the following command:
pip install -r requirements.txt
. - Download whisper large v2 to
whisper_path
. - Download Fine-tuned BEATs_iter3+ (AS2M) (cpt2) to
beats_path
. - Download vicuna 13B v1.1 to
llama_path
. - Running with
python3 train.py --cfg-path configs/config.yaml
in A100-SXM-80GB.
- Same as How to train a model: 1-4.
- Download salmonn v1 to
ckpt
. - Running with
python3 cli_inference.py --cfg-path configs/decode_config.yaml
in A100-SXM-80GB. Now you can inputwav_path
andprompt
. Enjoy yourself !
- Same as How to train a model: 1-4.
- Download salmonn v1 to
ckpt
. - Running with
python3 web_demo.py --cfg-path configs/decode_config.yaml
in A100-SXM-80GB.
Team Tsinghua: Wenyi Yu, Changli Tang, Guangzhi Sun, Chao Zhang
Team ByteDance: Xianzhao Chen, Wei Li, Tian Tan, Lu Lu, Zejun Ma
If you find SALMONN / video-SALMONN useful, please cite the paper:
@inproceedings{
tang2024salmonn,
title={{SALMONN}: Towards Generic Hearing Abilities for Large Language Models},
author={Changli Tang and Wenyi Yu and Guangzhi Sun and Xianzhao Chen and Tian Tan and Wei Li and Lu Lu and Zejun MA and Chao Zhang},
booktitle={The Twelfth International Conference on Learning Representations},
year={2024},
url={https://openreview.net/forum?id=14rn7HpKVk}
}
@inproceedings{
sun2024videosalmonn,
title={video-{SALMONN}: Speech-Enhanced Audio-Visual Large Language Models},
author={Guangzhi Sun and Wenyi Yu and Changli Tang and Xianzhao Chen and Tian Tan and Wei Li and Lu Lu and Zejun MA and Yuxuan Wang and Chao Zhang},
booktitle={Forty-first International Conference on Machine Learning},
year={2024},
url={https://openreview.net/forum?id=nYsh5GFIqX}
}