NotesPublished in Science (2024); originally arXiv:2303.10130
[Claude classification]: Published in Science (2024) as doi:10.1126/science.adj0817; originally arXiv:2303.10130 (Aug 2023). Foundational LLM exposure measure widely cited. Three variants: α (direct LLM only), β (E1 + 0.5*E2, with complementary software), ζ (E1 + E2, upper bound). GPT-4 and human annotators show ~80.8% agreement on α. R² of 60-73% vs. prior exposure measures (Webb, Felten, SML, Frey-Osborne), with 28-40% unexplained variance unique to LLM exposure. Authors from OpenAI and University of Pennsylvania. Paper argues LLMs meet criteria for general-purpose technology status. Does not make predictions about adoption timeline or actual labor market outcomes, only technical feasibility of task exposure. Image capabilities (E3) coded separately but combined with E2 for analysis. Includes comparison with productivity growth data showing weak correlation between recent productivity gains and LLM exposure, suggesting potential to avoid exacerbating cost disease.
[Claude classification]: Published in Science (2024) as doi:10.1126/science.adj0817; originally arXiv:2303.10130 (Aug 2023). Foundational LLM exposure measure widely cited. Three variants: α (direct LLM only), β (E1 + 0.5*E2, with complementary software), ζ (E1 + E2, upper bound). GPT-4 and human annotators show ~80.8% agreement on α. R² of 60-73% vs. prior exposure measures (Webb, Felten, SML, Frey-Osborne), with 28-40% unexplained variance unique to LLM exposure. Authors from OpenAI and University of Pennsylvania. Paper argues LLMs meet criteria for general-purpose technology status. Does not make predictions about adoption timeline or actual labor market outcomes, only technical feasibility of task exposure. Image capabilities (E3) coded separately but combined with E2 for analysis. Includes comparison with productivity growth data showing weak correlation between recent productivity gains and LLM exposure, suggesting potential to avoid exacerbating cost disease.
[Claude classification]: Published in Science (2024) as doi:10.1126/science.adj0817; originally arXiv:2303.10130 (Aug 2023). Foundational LLM exposure measure widely cited. Three variants: α (direct LLM only), β (E1 + 0.5*E2, with complementary software), ζ (E1 + E2, upper bound). GPT-4 and human annotators show ~80.8% agreement on α. R² of 60-73% vs. prior exposure measures (Webb, Felten, SML, Frey-Osborne), with 28-40% unexplained variance unique to LLM exposure. Authors from OpenAI and University of Pennsylvania. Paper argues LLMs meet criteria for general-purpose technology status. Does not make predictions about adoption timeline or actual labor market outcomes, only technical feasibility of task exposure. Image capabilities (E3) coded separately but combined with E2 for analysis. Includes comparison with productivity growth data showing weak correlation between recent productivity gains and LLM exposure, suggesting potential to avoid exacerbating cost disease.
[Claude classification]: Published in Science (2024) as doi:10.1126/science.adj0817; originally arXiv:2303.10130 (Aug 2023). Foundational LLM exposure measure widely cited. Three variants: α (direct LLM only), β (E1 + 0.5*E2, with complementary software), ζ (E1 + E2, upper bound). GPT-4 and human annotators show ~80.8% agreement on α. R² of 60-73% vs. prior exposure measures (Webb, Felten, SML, Frey-Osborne), with 28-40% unexplained variance unique to LLM exposure. Authors from OpenAI and University of Pennsylvania. Paper argues LLMs meet criteria for general-purpose technology status. Does not make predictions about adoption timeline or actual labor market outcomes, only technical feasibility of task exposure. Image capabilities (E3) coded separately but combined with E2 for analysis. Includes comparison with productivity growth data showing weak correlation between recent productivity gains and LLM exposure, suggesting potential to avoid exacerbating cost disease.
[Claude classification]: Published in Science (2024) as doi:10.1126/science.adj0817; originally arXiv:2303.10130 (Aug 2023). Foundational LLM exposure measure widely cited. Three variants: α (direct LLM only), β (E1 + 0.5*E2, with complementary software), ζ (E1 + E2, upper bound). GPT-4 and human annotators show ~80.8% agreement on α. R² of 60-73% vs. prior exposure measures (Webb, Felten, SML, Frey-Osborne), with 28-40% unexplained variance unique to LLM exposure. Authors from OpenAI and University of Pennsylvania. Paper argues LLMs meet criteria for general-purpose technology status. Does not make predictions about adoption timeline or actual labor market outcomes, only technical feasibility of task exposure. Image capabilities (E3) coded separately but combined with E2 for analysis. Includes comparison with productivity growth data showing weak correlation between recent productivity gains and LLM exposure, suggesting potential to avoid exacerbating cost disease.
[Claude classification]: Published in Science (2024) as doi:10.1126/science.adj0817; originally arXiv:2303.10130 (Aug 2023). Foundational LLM exposure measure widely cited. Three variants: α (direct LLM only), β (E1 + 0.5*E2, with complementary software), ζ (E1 + E2, upper bound). GPT-4 and human annotators show ~80.8% agreement on α. R² of 60-73% vs. prior exposure measures (Webb, Felten, SML, Frey-Osborne), with 28-40% unexplained variance unique to LLM exposure. Authors from OpenAI and University of Pennsylvania. Paper argues LLMs meet criteria for general-purpose technology status. Does not make predictions about adoption timeline or actual labor market outcomes, only technical feasibility of task exposure. Image capabilities (E3) coded separately but combined with E2 for analysis. Includes comparison with productivity growth data showing weak correlation between recent productivity gains and LLM exposure, suggesting potential to avoid exacerbating cost disease.
[Claude classification]: Published in Science (2024) as doi:10.1126/science.adj0817; originally arXiv:2303.10130 (Aug 2023). Foundational LLM exposure measure widely cited. Three variants: α (direct LLM only), β (E1 + 0.5*E2, with complementary software), ζ (E1 + E2, upper bound). GPT-4 and human annotators show ~80.8% agreement on α. R² of 60-73% vs. prior exposure measures (Webb, Felten, SML, Frey-Osborne), with 28-40% unexplained variance unique to LLM exposure. Authors from OpenAI and University of Pennsylvania. Paper argues LLMs meet criteria for general-purpose technology status. Does not make predictions about adoption timeline or actual labor market outcomes, only technical feasibility of task exposure. Image capabilities (E3) coded separately but combined with E2 for analysis. Includes comparison with productivity growth data showing weak correlation between recent productivity gains and LLM exposure, suggesting potential to avoid exacerbating cost disease.
[Claude classification]: Published in Science (2024) as doi:10.1126/science.adj0817; originally arXiv:2303.10130 (Aug 2023). Foundational LLM exposure measure widely cited. Three variants: α (direct LLM only), β (E1 + 0.5*E2, with complementary software), ζ (E1 + E2, upper bound). GPT-4 and human annotators show ~80.8% agreement on α. R² of 60-73% vs. prior exposure measures (Webb, Felten, SML, Frey-Osborne), with 28-40% unexplained variance unique to LLM exposure. Authors from OpenAI and University of Pennsylvania. Paper argues LLMs meet criteria for general-purpose technology status. Does not make predictions about adoption timeline or actual labor market outcomes, only technical feasibility of task exposure. Image capabilities (E3) coded separately but combined with E2 for analysis. Includes comparison with productivity growth data showing weak correlation between recent productivity gains and LLM exposure, suggesting potential to avoid exacerbating cost disease.
[Claude classification]: Published in Science (2024) as doi:10.1126/science.adj0817; originally arXiv:2303.10130 (Aug 2023). Foundational LLM exposure measure widely cited. Three variants: α (direct LLM only), β (E1 + 0.5*E2, with complementary software), ζ (E1 + E2, upper bound). GPT-4 and human annotators show ~80.8% agreement on α. R² of 60-73% vs. prior exposure measures (Webb, Felten, SML, Frey-Osborne), with 28-40% unexplained variance unique to LLM exposure. Authors from OpenAI and University of Pennsylvania. Paper argues LLMs meet criteria for general-purpose technology status. Does not make predictions about adoption timeline or actual labor market outcomes, only technical feasibility of task exposure. Image capabilities (E3) coded separately but combined with E2 for analysis. Includes comparison with productivity growth data showing weak correlation between recent productivity gains and LLM exposure, suggesting potential to avoid exacerbating cost disease.
[Claude classification]: Published in Science (2024) as doi:10.1126/science.adj0817; originally arXiv:2303.10130 (Aug 2023). Foundational LLM exposure measure widely cited. Three variants: α (direct LLM only), β (E1 + 0.5*E2, with complementary software), ζ (E1 + E2, upper bound). GPT-4 and human annotators show ~80.8% agreement on α. R² of 60-73% vs. prior exposure measures (Webb, Felten, SML, Frey-Osborne), with 28-40% unexplained variance unique to LLM exposure. Authors from OpenAI and University of Pennsylvania. Paper argues LLMs meet criteria for general-purpose technology status. Does not make predictions about adoption timeline or actual labor market outcomes, only technical feasibility of task exposure. Image capabilities (E3) coded separately but combined with E2 for analysis. Includes comparison with productivity growth data showing weak correlation between recent productivity gains and LLM exposure, suggesting potential to avoid exacerbating cost disease.
[Claude classification]: Published in Science (2024) as doi:10.1126/science.adj0817; originally arXiv:2303.10130 (Aug 2023). Foundational LLM exposure measure widely cited. Three variants: α (direct LLM only), β (E1 + 0.5*E2, with complementary software), ζ (E1 + E2, upper bound). GPT-4 and human annotators show ~80.8% agreement on α. R² of 60-73% vs. prior exposure measures (Webb, Felten, SML, Frey-Osborne), with 28-40% unexplained variance unique to LLM exposure. Authors from OpenAI and University of Pennsylvania. Paper argues LLMs meet criteria for general-purpose technology status. Does not make predictions about adoption timeline or actual labor market outcomes, only technical feasibility of task exposure. Image capabilities (E3) coded separately but combined with E2 for analysis. Includes comparison with productivity growth data showing weak correlation between recent productivity gains and LLM exposure, suggesting potential to avoid exacerbating cost disease.
[Claude classification]: Published in Science (2024) as doi:10.1126/science.adj0817; originally arXiv:2303.10130 (Aug 2023). Foundational LLM exposure measure widely cited. Three variants: α (direct LLM only), β (E1 + 0.5*E2, with complementary software), ζ (E1 + E2, upper bound). GPT-4 and human annotators show ~80.8% agreement on α. R² of 60-73% vs. prior exposure measures (Webb, Felten, SML, Frey-Osborne), with 28-40% unexplained variance unique to LLM exposure. Authors from OpenAI and University of Pennsylvania. Paper argues LLMs meet criteria for general-purpose technology status. Does not make predictions about adoption timeline or actual labor market outcomes, only technical feasibility of task exposure. Image capabilities (E3) coded separately but combined with E2 for analysis. Includes comparison with productivity growth data showing weak correlation between recent productivity gains and LLM exposure, suggesting potential to avoid exacerbating cost disease.
[Claude classification]: Published in Science (2024) as doi:10.1126/science.adj0817; originally arXiv:2303.10130 (Aug 2023). Foundational LLM exposure measure widely cited. Three variants: α (direct LLM only), β (E1 + 0.5*E2, with complementary software), ζ (E1 + E2, upper bound). GPT-4 and human annotators show ~80.8% agreement on α. R² of 60-73% vs. prior exposure measures (Webb, Felten, SML, Frey-Osborne), with 28-40% unexplained variance unique to LLM exposure. Authors from OpenAI and University of Pennsylvania. Paper argues LLMs meet criteria for general-purpose technology status. Does not make predictions about adoption timeline or actual labor market outcomes, only technical feasibility of task exposure. Image capabilities (E3) coded separately but combined with E2 for analysis. Includes comparison with productivity growth data showing weak correlation between recent productivity gains and LLM exposure, suggesting potential to avoid exacerbating cost disease.
[Claude classification]: Published in Science (2024) as doi:10.1126/science.adj0817; originally arXiv:2303.10130 (Aug 2023). Foundational LLM exposure measure widely cited. Three variants: α (direct LLM only), β (E1 + 0.5*E2, with complementary software), ζ (E1 + E2, upper bound). GPT-4 and human annotators show ~80.8% agreement on α. R² of 60-73% vs. prior exposure measures (Webb, Felten, SML, Frey-Osborne), with 28-40% unexplained variance unique to LLM exposure. Authors from OpenAI and University of Pennsylvania. Paper argues LLMs meet criteria for general-purpose technology status. Does not make predictions about adoption timeline or actual labor market outcomes, only technical feasibility of task exposure. Image capabilities (E3) coded separately but combined with E2 for analysis. Includes comparison with productivity growth data showing weak correlation between recent productivity gains and LLM exposure, suggesting potential to avoid exacerbating cost disease.
[Claude classification]: Published in Science (2024) as doi:10.1126/science.adj0817; originally arXiv:2303.10130 (Aug 2023). Foundational LLM exposure measure widely cited. Three variants: α (direct LLM only), β (E1 + 0.5*E2, with complementary software), ζ (E1 + E2, upper bound). GPT-4 and human annotators show ~80.8% agreement on α. R² of 60-73% vs. prior exposure measures (Webb, Felten, SML, Frey-Osborne), with 28-40% unexplained variance unique to LLM exposure. Authors from OpenAI and University of Pennsylvania. Paper argues LLMs meet criteria for general-purpose technology status. Does not make predictions about adoption timeline or actual labor market outcomes, only technical feasibility of task exposure. Image capabilities (E3) coded separately but combined with E2 for analysis. Includes comparison with productivity growth data showing weak correlation between recent productivity gains and LLM exposure, suggesting potential to avoid exacerbating cost disease.
[Claude classification]: Published in Science (2024) as doi:10.1126/science.adj0817; originally arXiv:2303.10130 (Aug 2023). Foundational LLM exposure measure widely cited. Three variants: α (direct LLM only), β (E1 + 0.5*E2, with complementary software), ζ (E1 + E2, upper bound). GPT-4 and human annotators show ~80.8% agreement on α. R² of 60-73% vs. prior exposure measures (Webb, Felten, SML, Frey-Osborne), with 28-40% unexplained variance unique to LLM exposure. Authors from OpenAI and University of Pennsylvania. Paper argues LLMs meet criteria for general-purpose technology status. Does not make predictions about adoption timeline or actual labor market outcomes, only technical feasibility of task exposure. Image capabilities (E3) coded separately but combined with E2 for analysis. Includes comparison with productivity growth data showing weak correlation between recent productivity gains and LLM exposure, suggesting potential to avoid exacerbating cost disease.
[Claude classification]: Published in Science (2024) as doi:10.1126/science.adj0817; originally arXiv:2303.10130 (Aug 2023). Foundational LLM exposure measure widely cited. Three variants: α (direct LLM only), β (E1 + 0.5*E2, with complementary software), ζ (E1 + E2, upper bound). GPT-4 and human annotators show ~80.8% agreement on α. R² of 60-73% vs. prior exposure measures (Webb, Felten, SML, Frey-Osborne), with 28-40% unexplained variance unique to LLM exposure. Authors from OpenAI and University of Pennsylvania. Paper argues LLMs meet criteria for general-purpose technology status. Does not make predictions about adoption timeline or actual labor market outcomes, only technical feasibility of task exposure. Image capabilities (E3) coded separately but combined with E2 for analysis. Includes comparison with productivity growth data showing weak correlation between recent productivity gains and LLM exposure, suggesting potential to avoid exacerbating cost disease.
[Claude classification]: Published in Science (2024) as doi:10.1126/science.adj0817; originally arXiv:2303.10130 (Aug 2023). Foundational LLM exposure measure widely cited. Three variants: α (direct LLM only), β (E1 + 0.5*E2, with complementary software), ζ (E1 + E2, upper bound). GPT-4 and human annotators show ~80.8% agreement on α. R² of 60-73% vs. prior exposure measures (Webb, Felten, SML, Frey-Osborne), with 28-40% unexplained variance unique to LLM exposure. Authors from OpenAI and University of Pennsylvania. Paper argues LLMs meet criteria for general-purpose technology status. Does not make predictions about adoption timeline or actual labor market outcomes, only technical feasibility of task exposure. Image capabilities (E3) coded separately but combined with E2 for analysis. Includes comparison with productivity growth data showing weak correlation between recent productivity gains and LLM exposure, suggesting potential to avoid exacerbating cost disease.
[Claude classification]: Published in Science (2024) as doi:10.1126/science.adj0817; originally arXiv:2303.10130 (Aug 2023). Foundational LLM exposure measure widely cited. Three variants: α (direct LLM only), β (E1 + 0.5*E2, with complementary software), ζ (E1 + E2, upper bound). GPT-4 and human annotators show ~80.8% agreement on α. R² of 60-73% vs. prior exposure measures (Webb, Felten, SML, Frey-Osborne), with 28-40% unexplained variance unique to LLM exposure. Authors from OpenAI and University of Pennsylvania. Paper argues LLMs meet criteria for general-purpose technology status. Does not make predictions about adoption timeline or actual labor market outcomes, only technical feasibility of task exposure. Image capabilities (E3) coded separately but combined with E2 for analysis. Includes comparison with productivity growth data showing weak correlation between recent productivity gains and LLM exposure, suggesting potential to avoid exacerbating cost disease.
[Claude classification]: Published in Science (2024) as doi:10.1126/science.adj0817; originally arXiv:2303.10130 (Aug 2023). Foundational LLM exposure measure widely cited. Three variants: α (direct LLM only), β (E1 + 0.5*E2, with complementary software), ζ (E1 + E2, upper bound). GPT-4 and human annotators show ~80.8% agreement on α. R² of 60-73% vs. prior exposure measures (Webb, Felten, SML, Frey-Osborne), with 28-40% unexplained variance unique to LLM exposure. Authors from OpenAI and University of Pennsylvania. Paper argues LLMs meet criteria for general-purpose technology status. Does not make predictions about adoption timeline or actual labor market outcomes, only technical feasibility of task exposure. Image capabilities (E3) coded separately but combined with E2 for analysis. Includes comparison with productivity growth data showing weak correlation between recent productivity gains and LLM exposure, suggesting potential to avoid exacerbating cost disease.
[Claude classification]: Published in Science (2024) as doi:10.1126/science.adj0817; originally arXiv:2303.10130 (Aug 2023). Foundational LLM exposure measure widely cited. Three variants: α (direct LLM only), β (E1 + 0.5*E2, with complementary software), ζ (E1 + E2, upper bound). GPT-4 and human annotators show ~80.8% agreement on α. R² of 60-73% vs. prior exposure measures (Webb, Felten, SML, Frey-Osborne), with 28-40% unexplained variance unique to LLM exposure. Authors from OpenAI and University of Pennsylvania. Paper argues LLMs meet criteria for general-purpose technology status. Does not make predictions about adoption timeline or actual labor market outcomes, only technical feasibility of task exposure. Image capabilities (E3) coded separately but combined with E2 for analysis. Includes comparison with productivity growth data showing weak correlation between recent productivity gains and LLM exposure, suggesting potential to avoid exacerbating cost disease.
[Claude classification]: Published in Science (2024) as doi:10.1126/science.adj0817; originally arXiv:2303.10130 (Aug 2023). Foundational LLM exposure measure widely cited. Three variants: α (direct LLM only), β (E1 + 0.5*E2, with complementary software), ζ (E1 + E2, upper bound). GPT-4 and human annotators show ~80.8% agreement on α. R² of 60-73% vs. prior exposure measures (Webb, Felten, SML, Frey-Osborne), with 28-40% unexplained variance unique to LLM exposure. Authors from OpenAI and University of Pennsylvania. Paper argues LLMs meet criteria for general-purpose technology status. Does not make predictions about adoption timeline or actual labor market outcomes, only technical feasibility of task exposure. Image capabilities (E3) coded separately but combined with E2 for analysis. Includes comparison with productivity growth data showing weak correlation between recent productivity gains and LLM exposure, suggesting potential to avoid exacerbating cost disease.
[Claude classification]: Published in Science (2024) as doi:10.1126/science.adj0817; originally arXiv:2303.10130 (Aug 2023). Foundational LLM exposure measure widely cited. Three variants: α (direct LLM only), β (E1 + 0.5*E2, with complementary software), ζ (E1 + E2, upper bound). GPT-4 and human annotators show ~80.8% agreement on α. R² of 60-73% vs. prior exposure measures (Webb, Felten, SML, Frey-Osborne), with 28-40% unexplained variance unique to LLM exposure. Authors from OpenAI and University of Pennsylvania. Paper argues LLMs meet criteria for general-purpose technology status. Does not make predictions about adoption timeline or actual labor market outcomes, only technical feasibility of task exposure. Image capabilities (E3) coded separately but combined with E2 for analysis. Includes comparison with productivity growth data showing weak correlation between recent productivity gains and LLM exposure, suggesting potential to avoid exacerbating cost disease.
[Claude classification]: Published in Science (2024) as doi:10.1126/science.adj0817; originally arXiv:2303.10130 (Aug 2023). Foundational LLM exposure measure widely cited. Three variants: α (direct LLM only), β (E1 + 0.5*E2, with complementary software), ζ (E1 + E2, upper bound). GPT-4 and human annotators show ~80.8% agreement on α. R² of 60-73% vs. prior exposure measures (Webb, Felten, SML, Frey-Osborne), with 28-40% unexplained variance unique to LLM exposure. Authors from OpenAI and University of Pennsylvania. Paper argues LLMs meet criteria for general-purpose technology status. Does not make predictions about adoption timeline or actual labor market outcomes, only technical feasibility of task exposure. Image capabilities (E3) coded separately but combined with E2 for analysis. Includes comparison with productivity growth data showing weak correlation between recent productivity gains and LLM exposure, suggesting potential to avoid exacerbating cost disease.
[Claude classification]: Published in Science (2024) as doi:10.1126/science.adj0817; originally arXiv:2303.10130 (Aug 2023). Foundational LLM exposure measure widely cited. Three variants: α (direct LLM only), β (E1 + 0.5*E2, with complementary software), ζ (E1 + E2, upper bound). GPT-4 and human annotators show ~80.8% agreement on α. R² of 60-73% vs. prior exposure measures (Webb, Felten, SML, Frey-Osborne), with 28-40% unexplained variance unique to LLM exposure. Authors from OpenAI and University of Pennsylvania. Paper argues LLMs meet criteria for general-purpose technology status. Does not make predictions about adoption timeline or actual labor market outcomes, only technical feasibility of task exposure. Image capabilities (E3) coded separately but combined with E2 for analysis. Includes comparison with productivity growth data showing weak correlation between recent productivity gains and LLM exposure, suggesting potential to avoid exacerbating cost disease.
[Claude classification]: Published in Science (2024) as doi:10.1126/science.adj0817; originally arXiv:2303.10130 (Aug 2023). Foundational LLM exposure measure widely cited. Three variants: α (direct LLM only), β (E1 + 0.5*E2, with complementary software), ζ (E1 + E2, upper bound). GPT-4 and human annotators show ~80.8% agreement on α. R² of 60-73% vs. prior exposure measures (Webb, Felten, SML, Frey-Osborne), with 28-40% unexplained variance unique to LLM exposure. Authors from OpenAI and University of Pennsylvania. Paper argues LLMs meet criteria for general-purpose technology status. Does not make predictions about adoption timeline or actual labor market outcomes, only technical feasibility of task exposure. Image capabilities (E3) coded separately but combined with E2 for analysis. Includes comparison with productivity growth data showing weak correlation between recent productivity gains and LLM exposure, suggesting potential to avoid exacerbating cost disease.
[Claude classification]: Published in Science (2024) as doi:10.1126/science.adj0817; originally arXiv:2303.10130 (Aug 2023). Foundational LLM exposure measure widely cited. Three variants: α (direct LLM only), β (E1 + 0.5*E2, with complementary software), ζ (E1 + E2, upper bound). GPT-4 and human annotators show ~80.8% agreement on α. R² of 60-73% vs. prior exposure measures (Webb, Felten, SML, Frey-Osborne), with 28-40% unexplained variance unique to LLM exposure. Authors from OpenAI and University of Pennsylvania. Paper argues LLMs meet criteria for general-purpose technology status. Does not make predictions about adoption timeline or actual labor market outcomes, only technical feasibility of task exposure. Image capabilities (E3) coded separately but combined with E2 for analysis. Includes comparison with productivity growth data showing weak correlation between recent productivity gains and LLM exposure, suggesting potential to avoid exacerbating cost disease.
[Claude classification]: Published in Science (2024) as doi:10.1126/science.adj0817; originally arXiv:2303.10130 (Aug 2023). Foundational LLM exposure measure widely cited. Three variants: α (direct LLM only), β (E1 + 0.5*E2, with complementary software), ζ (E1 + E2, upper bound). GPT-4 and human annotators show ~80.8% agreement on α. R² of 60-73% vs. prior exposure measures (Webb, Felten, SML, Frey-Osborne), with 28-40% unexplained variance unique to LLM exposure. Authors from OpenAI and University of Pennsylvania. Paper argues LLMs meet criteria for general-purpose technology status. Does not make predictions about adoption timeline or actual labor market outcomes, only technical feasibility of task exposure. Image capabilities (E3) coded separately but combined with E2 for analysis. Includes comparison with productivity growth data showing weak correlation between recent productivity gains and LLM exposure, suggesting potential to avoid exacerbating cost disease.
[Claude classification]: Published in Science (2024) as doi:10.1126/science.adj0817; originally arXiv:2303.10130 (Aug 2023). Foundational LLM exposure measure widely cited. Three variants: α (direct LLM only), β (E1 + 0.5*E2, with complementary software), ζ (E1 + E2, upper bound). GPT-4 and human annotators show ~80.8% agreement on α. R² of 60-73% vs. prior exposure measures (Webb, Felten, SML, Frey-Osborne), with 28-40% unexplained variance unique to LLM exposure. Authors from OpenAI and University of Pennsylvania. Paper argues LLMs meet criteria for general-purpose technology status. Does not make predictions about adoption timeline or actual labor market outcomes, only technical feasibility of task exposure. Image capabilities (E3) coded separately but combined with E2 for analysis. Includes comparison with productivity growth data showing weak correlation between recent productivity gains and LLM exposure, suggesting potential to avoid exacerbating cost disease.
[Claude classification]: Published in Science (2024) as doi:10.1126/science.adj0817; originally arXiv:2303.10130 (Aug 2023). Foundational LLM exposure measure widely cited. Three variants: α (direct LLM only), β (E1 + 0.5*E2, with complementary software), ζ (E1 + E2, upper bound). GPT-4 and human annotators show ~80.8% agreement on α. R² of 60-73% vs. prior exposure measures (Webb, Felten, SML, Frey-Osborne), with 28-40% unexplained variance unique to LLM exposure. Authors from OpenAI and University of Pennsylvania. Paper argues LLMs meet criteria for general-purpose technology status. Does not make predictions about adoption timeline or actual labor market outcomes, only technical feasibility of task exposure. Image capabilities (E3) coded separately but combined with E2 for analysis. Includes comparison with productivity growth data showing weak correlation between recent productivity gains and LLM exposure, suggesting potential to avoid exacerbating cost disease.
[Claude classification]: Published in Science (2024) as doi:10.1126/science.adj0817; originally arXiv:2303.10130 (Aug 2023). Foundational LLM exposure measure widely cited. Three variants: α (direct LLM only), β (E1 + 0.5*E2, with complementary software), ζ (E1 + E2, upper bound). GPT-4 and human annotators show ~80.8% agreement on α. R² of 60-73% vs. prior exposure measures (Webb, Felten, SML, Frey-Osborne), with 28-40% unexplained variance unique to LLM exposure. Authors from OpenAI and University of Pennsylvania. Paper argues LLMs meet criteria for general-purpose technology status. Does not make predictions about adoption timeline or actual labor market outcomes, only technical feasibility of task exposure. Image capabilities (E3) coded separately but combined with E2 for analysis. Includes comparison with productivity growth data showing weak correlation between recent productivity gains and LLM exposure, suggesting potential to avoid exacerbating cost disease.
[Claude classification]: Published in Science (2024) as doi:10.1126/science.adj0817; originally arXiv:2303.10130 (Aug 2023). Foundational LLM exposure measure widely cited. Three variants: α (direct LLM only), β (E1 + 0.5*E2, with complementary software), ζ (E1 + E2, upper bound). GPT-4 and human annotators show ~80.8% agreement on α. R² of 60-73% vs. prior exposure measures (Webb, Felten, SML, Frey-Osborne), with 28-40% unexplained variance unique to LLM exposure. Authors from OpenAI and University of Pennsylvania. Paper argues LLMs meet criteria for general-purpose technology status. Does not make predictions about adoption timeline or actual labor market outcomes, only technical feasibility of task exposure. Image capabilities (E3) coded separately but combined with E2 for analysis. Includes comparison with productivity growth data showing weak correlation between recent productivity gains and LLM exposure, suggesting potential to avoid exacerbating cost disease.
[Claude classification]: Published in Science (2024) as doi:10.1126/science.adj0817; originally arXiv:2303.10130 (Aug 2023). Foundational LLM exposure measure widely cited. Three variants: α (direct LLM only), β (E1 + 0.5*E2, with complementary software), ζ (E1 + E2, upper bound). GPT-4 and human annotators show ~80.8% agreement on α. R² of 60-73% vs. prior exposure measures (Webb, Felten, SML, Frey-Osborne), with 28-40% unexplained variance unique to LLM exposure. Authors from OpenAI and University of Pennsylvania. Paper argues LLMs meet criteria for general-purpose technology status. Does not make predictions about adoption timeline or actual labor market outcomes, only technical feasibility of task exposure. Image capabilities (E3) coded separately but combined with E2 for analysis. Includes comparison with productivity growth data showing weak correlation between recent productivity gains and LLM exposure, suggesting potential to avoid exacerbating cost disease.
[Claude classification]: Published in Science (2024) as doi:10.1126/science.adj0817; originally arXiv:2303.10130 (Aug 2023). Foundational LLM exposure measure widely cited. Three variants: α (direct LLM only), β (E1 + 0.5*E2, with complementary software), ζ (E1 + E2, upper bound). GPT-4 and human annotators show ~80.8% agreement on α. R² of 60-73% vs. prior exposure measures (Webb, Felten, SML, Frey-Osborne), with 28-40% unexplained variance unique to LLM exposure. Authors from OpenAI and University of Pennsylvania. Paper argues LLMs meet criteria for general-purpose technology status. Does not make predictions about adoption timeline or actual labor market outcomes, only technical feasibility of task exposure. Image capabilities (E3) coded separately but combined with E2 for analysis. Includes comparison with productivity growth data showing weak correlation between recent productivity gains and LLM exposure, suggesting potential to avoid exacerbating cost disease.
[Claude classification]: Published in Science (2024) as doi:10.1126/science.adj0817; originally arXiv:2303.10130 (Aug 2023). Foundational LLM exposure measure widely cited. Three variants: α (direct LLM only), β (E1 + 0.5*E2, with complementary software), ζ (E1 + E2, upper bound). GPT-4 and human annotators show ~80.8% agreement on α. R² of 60-73% vs. prior exposure measures (Webb, Felten, SML, Frey-Osborne), with 28-40% unexplained variance unique to LLM exposure. Authors from OpenAI and University of Pennsylvania. Paper argues LLMs meet criteria for general-purpose technology status. Does not make predictions about adoption timeline or actual labor market outcomes, only technical feasibility of task exposure. Image capabilities (E3) coded separately but combined with E2 for analysis. Includes comparison with productivity growth data showing weak correlation between recent productivity gains and LLM exposure, suggesting potential to avoid exacerbating cost disease.