What is Fine-Tuning? | Stanford HAI
Stanford
University
  • Stanford Home
  • Maps & Directions
  • Search Stanford
  • Emergency Info
  • Terms of Use
  • Privacy
  • Copyright
  • Trademarks
  • Non-Discrimination
  • Accessibility
© Stanford University.  Stanford, California 94305.
Skip to content
  • About

    • About
    • People
    • Get Involved with HAI
    • Support HAI
    • Subscribe to Email
  • Research

    • Research
    • Fellowship Programs
    • Grants
    • Student Affinity Groups
    • Centers & Labs
    • Research Publications
    • Research Partners
  • Education

    • Education
    • Executive and Professional Education
    • Government and Policymakers
    • K-12
    • Stanford Students
  • Policy

    • Policy
    • Policy Publications
    • Policymaker Education
    • Student Opportunities
  • AI Index

    • AI Index
    • AI Index Report
    • Global Vibrancy Tool
    • People
  • News
  • Events
  • Industry
  • Centers & Labs

What is Fine-Tuning?

Fine-Tuning is the process of taking a pre-trained model and further training it on task or domain specific data. Instead of training a model from scratch (which requires massive amounts of data and computational resources), Fine-Tuning starts with a model that already understands general patterns and adjusts its parameters to perform well on your particular use case—like adapting a general language model to understand medical terminology or legal documents.

Navigate
  • About
  • Events
  • Careers
  • Search
Participate
  • Get Involved
  • Support HAI
  • Contact Us

Stay Up To Date

Get the latest news, advances in research, policy work, and education program updates from HAI in your inbox weekly.

Sign Up For Latest News


Fine-Tuning mentioned mentioned at Stanford Stanford HAI

Explore Similar Terms:

Transfer Learning | Foundation Model | Training Data

See Full List of Terms & Definitions

Enroll in a Human-Centered AI Course

This AI program covers technical fundamentals, business implications, and societal considerations.
Scalable Differential Privacy with Sparse Network Fine-Tuning
Zelun Luo, Daniel Wu, Ehsan Adeli
Nov 28
Research
Your browser does not support the video tag.

Scalable Differential Privacy with Sparse Network Fine-Tuning

Scalable Differential Privacy with Sparse Network Fine-Tuning

Zelun Luo, Daniel Wu, Ehsan Adeli
Nov 28

Scalable Differential Privacy with Sparse Network Fine-Tuning

Your browser does not support the video tag.
Research
Improving Equity and Access to Non-English Large Language Models
Prabha Kannan
Apr 22
news

The lessons learned from the fine-tuning and evaluation of Vietnamese LLMs could help broaden access to models beyond English speakers.

Improving Equity and Access to Non-English Large Language Models

Prabha Kannan
Apr 22

The lessons learned from the fine-tuning and evaluation of Vietnamese LLMs could help broaden access to models beyond English speakers.

Natural Language Processing
news
Safety Risks from Customizing Foundation Models via Fine-Tuning
Peter Henderson, Xiangyu Qi, Yi Zeng, Tinghao Xie, Pin-Yu Chen, Ruoxi Jia, Prateek Mittal
Quick ReadJan 08
policy brief

This brief underscores the safety risks inherent in custom fine-tuning of large language models.

Safety Risks from Customizing Foundation Models via Fine-Tuning

Peter Henderson, Xiangyu Qi, Yi Zeng, Tinghao Xie, Pin-Yu Chen, Ruoxi Jia, Prateek Mittal
Quick ReadJan 08

This brief underscores the safety risks inherent in custom fine-tuning of large language models.

Foundation Models
Privacy, Safety, Security
policy brief
Covert Racism in AI: How Language Models Are Reinforcing Outdated Stereotypes
Katharine Miller
Sep 03
news

Despite advancements in AI, new research reveals that large language models continue to perpetuate harmful racial biases, particularly against speakers of African American English. 

Covert Racism in AI: How Language Models Are Reinforcing Outdated Stereotypes

Katharine Miller
Sep 03

Despite advancements in AI, new research reveals that large language models continue to perpetuate harmful racial biases, particularly against speakers of African American English. 

news
LABOR-LLM: Language-Based Occupational Representations with Large Language Models
Susan Athey, Herman Brunborg, Tianyu Du, Ayush Kanodia, Keyon Vafa
Dec 11
Research
Your browser does not support the video tag.

Vafa et al. (2024) introduced a transformer-based econometric model, CAREER, that predicts a worker’s next job as a function of career history (an “occupation model”). CAREER was initially estimated (“pre-trained”) using a large, unrepresentative resume dataset, which served as a “foundation model,” and parameter estimation was continued (“fine-tuned”) using data from a representative survey. CAREER had better predictive performance than benchmarks. This paper considers an alternative where the resume-based foundation model is replaced by a large language model (LLM). We convert tabular data from the survey into text files that resemble resumes and fine-tune the LLMs using these text files with the objective to predict the next token (word). The resulting fine-tuned LLM is used as an input to an occupation model. Its predictive performance surpasses all prior models. We demonstrate the value of fine-tuning and further show that by adding more career data from a different population, fine-tuning smaller LLMs surpasses the performance of fine-tuning larger models.

LABOR-LLM: Language-Based Occupational Representations with Large Language Models

Susan Athey, Herman Brunborg, Tianyu Du, Ayush Kanodia, Keyon Vafa
Dec 11

Vafa et al. (2024) introduced a transformer-based econometric model, CAREER, that predicts a worker’s next job as a function of career history (an “occupation model”). CAREER was initially estimated (“pre-trained”) using a large, unrepresentative resume dataset, which served as a “foundation model,” and parameter estimation was continued (“fine-tuned”) using data from a representative survey. CAREER had better predictive performance than benchmarks. This paper considers an alternative where the resume-based foundation model is replaced by a large language model (LLM). We convert tabular data from the survey into text files that resemble resumes and fine-tune the LLMs using these text files with the objective to predict the next token (word). The resulting fine-tuned LLM is used as an input to an occupation model. Its predictive performance surpasses all prior models. We demonstrate the value of fine-tuning and further show that by adding more career data from a different population, fine-tuning smaller LLMs surpasses the performance of fine-tuning larger models.

Foundation Models
Natural Language Processing
Your browser does not support the video tag.
Research
Using AI to Streamline Speech and Language Services for Children
Katharine Miller
Oct 27
news
A child works with a speech pathologist to sound out words

Stanford researchers show that although top language models cannot yet accurately diagnose children’s speech disorders, fine-tuning and other approaches could well change the game.

Using AI to Streamline Speech and Language Services for Children

Katharine Miller
Oct 27

Stanford researchers show that although top language models cannot yet accurately diagnose children’s speech disorders, fine-tuning and other approaches could well change the game.

Generative AI
Healthcare
A child works with a speech pathologist to sound out words
news
Could Stable Diffusion Solve a Gap in Medical Imaging Data?
Nikki Goth Itoi
Nov 29
news

Stanford AIMI scholars found a way to generate synthetic chest X-rays by fine-tuning the open-source Stable Diffusion foundation model.

Could Stable Diffusion Solve a Gap in Medical Imaging Data?

Nikki Goth Itoi
Nov 29

Stanford AIMI scholars found a way to generate synthetic chest X-rays by fine-tuning the open-source Stable Diffusion foundation model.

news