LLaMA: Open and Efficient Foundation Language Models

root 提交于 周日, 01/21/2024 - 12:12
We introduce LLaMA, a collection of foundation language models ranging from 7B to 65B parameters. We train our models on trillions of tokens, and show that it is possible to train state-of-the-art models using publicly available datasets exclusively, without resorting to proprietary and inaccessible datasets. In particular, LLaMA-13B outperforms GPT-3 (175B) on most benchmarks, and LLaMA-65B is competitive with the best models, Chinchilla-70B and PaLM-540B. We release all our models to the research community.

相关内容

发布日期 08/04/2020 - 01:35
发布日期 08/04/2020 - 01:35
发布日期 10/17/2023 - 23:16
发布日期 02/18/2025 - 20:47
发布日期 01/18/2025 - 20:37
发布日期 01/31/2024 - 13:01
发布日期 02/29/2024 - 16:35