This site is a work in progress and has not been widely shared. Content may contain errors. Feedback is welcome.
This site is undergoing review. Some annotations were human-generated, some AI-generated — all are being verified.
Back to papers

The Impact of Large Language Models on Open-Source Innovation: Evidence from GitHub Copilot

Yeverechyahu, Mayya, Oestreicher-Singer

2024Working paper7 citations
Observational labor marketInterdisciplinaryCausal
LLM / Generative AISoftware / codingAugmentation vs. substitutionHuman-AI collaboration
Summary

Yeverechyahu, Mayya, and Oestreicher-Singer exploit GitHub Copilot's selective programming language support at launch using difference-in-differences to study how LLMs affect open-source innovation volume and type across Python/Rust (treated) versus R/Haskell (control) packages from October 2019 to December 2022.

Main Finding

GitHub Copilot increases overall open-source contributions by 37% for Python (vs. R) and 54% for Rust (vs. Haskell), with disproportionately larger effects on iterative innovation (maintenance commits) compared to capability innovation (new feature development), particularly in high-activity projects with rich contextual information.

Primary Datasets

GitHub API data on commits and repository activity for Python, R, Rust, and Haskell packages; PyPI, CRAN, Hackage, and Crates.io for version release data

Secondary Datasets

None

Key Methods
Difference-in-differences with propensity score matching, comparing programming languages supported by GitHub Copilot (Python, Rust) versus unsupported languages (R, Haskell); synthetic difference-in-differences as robustness check; LLM-based classification of commit types
Sample Period
2019-2022
Geographic Coverage
Global (GitHub open-source projects)
Sample Size
Over 1.1 million commits across 1,187 matched Python/R packages and 1,373 matched Rust/Haskell packages from October 2019 to December 2022
Level of Analysis
Task, Firm
Occupation Classification
None
Industry Classification
None
Notes
arXiv:2409.08379 [Claude classification]: Uses LLMs (GPT-4o) as methodological tool to classify commit types; natural experiment exploits selective language support of GitHub Copilot at launch (October 2021); distinguishes between capability innovation (new features) and iterative innovation (maintenance/refinement); study period ends December 2022 before ChatGPT affected control group [Claude classification]: Uses LLMs (GPT-4o) as methodological tool to classify commit types; natural experiment exploits selective language support of GitHub Copilot at launch (October 2021); distinguishes between capability innovation (new features) and iterative innovation (maintenance/refinement); study period ends December 2022 before ChatGPT affected control group [Claude classification]: Uses LLMs (GPT-4o) as methodological tool to classify commit types; natural experiment exploits selective language support of GitHub Copilot at launch (October 2021); distinguishes between capability innovation (new features) and iterative innovation (maintenance/refinement); study period ends December 2022 before ChatGPT affected control group [Claude classification]: Uses LLMs (GPT-4o) as methodological tool to classify commit types; natural experiment exploits selective language support of GitHub Copilot at launch (October 2021); distinguishes between capability innovation (new features) and iterative innovation (maintenance/refinement); study period ends December 2022 before ChatGPT affected control group [Claude classification]: Uses LLMs (GPT-4o) as methodological tool to classify commit types; natural experiment exploits selective language support of GitHub Copilot at launch (October 2021); distinguishes between capability innovation (new features) and iterative innovation (maintenance/refinement); study period ends December 2022 before ChatGPT affected control group [Claude classification]: Uses LLMs (GPT-4o) as methodological tool to classify commit types; natural experiment exploits selective language support of GitHub Copilot at launch (October 2021); distinguishes between capability innovation (new features) and iterative innovation (maintenance/refinement); study period ends December 2022 before ChatGPT affected control group [Claude classification]: Uses LLMs (GPT-4o) as methodological tool to classify commit types; natural experiment exploits selective language support of GitHub Copilot at launch (October 2021); distinguishes between capability innovation (new features) and iterative innovation (maintenance/refinement); study period ends December 2022 before ChatGPT affected control group [Claude classification]: Uses LLMs (GPT-4o) as methodological tool to classify commit types; natural experiment exploits selective language support of GitHub Copilot at launch (October 2021); distinguishes between capability innovation (new features) and iterative innovation (maintenance/refinement); study period ends December 2022 before ChatGPT affected control group [Claude classification]: Uses LLMs (GPT-4o) as methodological tool to classify commit types; natural experiment exploits selective language support of GitHub Copilot at launch (October 2021); distinguishes between capability innovation (new features) and iterative innovation (maintenance/refinement); study period ends December 2022 before ChatGPT affected control group [Claude classification]: Uses LLMs (GPT-4o) as methodological tool to classify commit types; natural experiment exploits selective language support of GitHub Copilot at launch (October 2021); distinguishes between capability innovation (new features) and iterative innovation (maintenance/refinement); study period ends December 2022 before ChatGPT affected control group [Claude classification]: Uses LLMs (GPT-4o) as methodological tool to classify commit types; natural experiment exploits selective language support of GitHub Copilot at launch (October 2021); distinguishes between capability innovation (new features) and iterative innovation (maintenance/refinement); study period ends December 2022 before ChatGPT affected control group [Claude classification]: Uses LLMs (GPT-4o) as methodological tool to classify commit types; natural experiment exploits selective language support of GitHub Copilot at launch (October 2021); distinguishes between capability innovation (new features) and iterative innovation (maintenance/refinement); study period ends December 2022 before ChatGPT affected control group [Claude classification]: Uses LLMs (GPT-4o) as methodological tool to classify commit types; natural experiment exploits selective language support of GitHub Copilot at launch (October 2021); distinguishes between capability innovation (new features) and iterative innovation (maintenance/refinement); study period ends December 2022 before ChatGPT affected control group [Claude classification]: Uses LLMs (GPT-4o) as methodological tool to classify commit types; natural experiment exploits selective language support of GitHub Copilot at launch (October 2021); distinguishes between capability innovation (new features) and iterative innovation (maintenance/refinement); study period ends December 2022 before ChatGPT affected control group [Claude classification]: Uses LLMs (GPT-4o) as methodological tool to classify commit types; natural experiment exploits selective language support of GitHub Copilot at launch (October 2021); distinguishes between capability innovation (new features) and iterative innovation (maintenance/refinement); study period ends December 2022 before ChatGPT affected control group [Claude classification]: Uses LLMs (GPT-4o) as methodological tool to classify commit types; natural experiment exploits selective language support of GitHub Copilot at launch (October 2021); distinguishes between capability innovation (new features) and iterative innovation (maintenance/refinement); study period ends December 2022 before ChatGPT affected control group [Claude classification]: Uses LLMs (GPT-4o) as methodological tool to classify commit types; natural experiment exploits selective language support of GitHub Copilot at launch (October 2021); distinguishes between capability innovation (new features) and iterative innovation (maintenance/refinement); study period ends December 2022 before ChatGPT affected control group [Claude classification]: Uses LLMs (GPT-4o) as methodological tool to classify commit types; natural experiment exploits selective language support of GitHub Copilot at launch (October 2021); distinguishes between capability innovation (new features) and iterative innovation (maintenance/refinement); study period ends December 2022 before ChatGPT affected control group [Claude classification]: Uses LLMs (GPT-4o) as methodological tool to classify commit types; natural experiment exploits selective language support of GitHub Copilot at launch (October 2021); distinguishes between capability innovation (new features) and iterative innovation (maintenance/refinement); study period ends December 2022 before ChatGPT affected control group [Claude classification]: Uses LLMs (GPT-4o) as methodological tool to classify commit types; natural experiment exploits selective language support of GitHub Copilot at launch (October 2021); distinguishes between capability innovation (new features) and iterative innovation (maintenance/refinement); study period ends December 2022 before ChatGPT affected control group [Claude classification]: Uses LLMs (GPT-4o) as methodological tool to classify commit types; natural experiment exploits selective language support of GitHub Copilot at launch (October 2021); distinguishes between capability innovation (new features) and iterative innovation (maintenance/refinement); study period ends December 2022 before ChatGPT affected control group [Claude classification]: Uses LLMs (GPT-4o) as methodological tool to classify commit types; natural experiment exploits selective language support of GitHub Copilot at launch (October 2021); distinguishes between capability innovation (new features) and iterative innovation (maintenance/refinement); study period ends December 2022 before ChatGPT affected control group [Claude classification]: Uses LLMs (GPT-4o) as methodological tool to classify commit types; natural experiment exploits selective language support of GitHub Copilot at launch (October 2021); distinguishes between capability innovation (new features) and iterative innovation (maintenance/refinement); study period ends December 2022 before ChatGPT affected control group [Claude classification]: Uses LLMs (GPT-4o) as methodological tool to classify commit types; natural experiment exploits selective language support of GitHub Copilot at launch (October 2021); distinguishes between capability innovation (new features) and iterative innovation (maintenance/refinement); study period ends December 2022 before ChatGPT affected control group [Claude classification]: Uses LLMs (GPT-4o) as methodological tool to classify commit types; natural experiment exploits selective language support of GitHub Copilot at launch (October 2021); distinguishes between capability innovation (new features) and iterative innovation (maintenance/refinement); study period ends December 2022 before ChatGPT affected control group