1. Bommasani, R., Hudson, D. A., Adeli, E., et al. (2021). On the Opportunities and Risks of Foundation Models. Stanford University Center for Research on Foundation Models (CRFM), Stanford Institute for Human-Centered AI (HAI). In Section 1.1, "Definition," the authors state: "We define foundation models as models trained on broad data (generally self-supervised at scale) that can be adapted (e.g., fine-tuned) to a wide range of downstream tasks." (Page 5).
- Available at: https://arxiv.org/abs/2108.07258
2. NVIDIA Technical Blog. (2023, August 9). What Are Foundation Models? The article defines the term in its opening paragraph: "A foundation model is a large AI model pre-trained on a vast quantity of data that can be adapted to a wide range of downstream tasks."
- Available at: https://blogs.nvidia.com/blog/what-are-foundation-models/
3. Manning, C. (2023). Lecture 11: Pretraining and Transformers. Stanford University, CS224N: Natural Language Processing with Deep Learning, Winter 2023. The lecture describes the pre-training/fine-tuning paradigm, explaining that large models are pre-trained on massive text corpora and then fine-tuned for specific downstream tasks, which is the core concept of a foundation model.
- Lecture Slides available at: https://web.stanford.edu/class/cs224n/slides/cs224n-2023-lecture11-pretraining-transformers.pdf (Slides 5-12).