My Study Notes on Meta AI

(Illustration: Lac de Neuchatel, Switzerland. Image source: Ernest)

Briefing

History

From ancient to modern times. Get through all the context.


Products

Meta LLaMA

Meta Llama 2

Meta announced the release of Llama-2 on July 18, 2023, as the advanced iteration of LLaMA, available in model sizes of 7 billion, 13 billion, and 70 billion parameters. While maintaining a similar structure to the LLaMA-1 models, these newer versions were trained with 40% more data. A preprint document 1 also reveals plans for a 34 billion parameter model, pending the achievement of requisite safety criteria.

(Safety human evaluation results. Lower is safer.)

(Carbon Footprint of Pretraining.)

Meta Llama 3

Meta launched two configurations of the Llama-3 model on April 19, 2024, with sizes of 8 billion (8B) and 70 billion (70B) parameters, respectively. Pre-trained on about 15 trillion text tokens from sources available to the public, these models were refined further using instruction datasets also publicly accessible, complemented by more than 10 million human-annotated instances. Plans are in place to roll out multimodal capabilities, multilingual support, and larger contextual understanding in upcoming models. A model boasting over 400 billion parameters (400B+) is currently in training.


使用情境 Use Cases


參考資料 Reference

比較 Comparison


Loading comments…