0:00

0:00


Future of Work Report 2025: AI Collective Productivity

📌 Key Takeaways

  • Collective Productivity Frontier: AI delivers strong individual gains (40-60 minutes saved daily) but hasn’t yet translated to team or organizational-level productivity improvements.
  • Workslop Is Spreading: 40% of employees receive AI-generated content that looks polished but lacks substance, accounting for an estimated 15% of workplace content and potentially undermining group performance.
  • Junior Workers Hit Hardest: Employment for workers aged 22-25 in AI-exposed jobs fell by about 13%, while writing jobs declined by 30% and software jobs by 21% post-ChatGPT.
  • Cognitive Deskilling Is Real: Clinicians relying on AI for diagnostics showed significant decline in independent ability after just three months—a warning for all knowledge workers.
  • $33.9B Investment: Global private investment in generative AI reached $33.9 billion in 2024, up 18.7% from 2023, with ChatGPT surpassing 700 million weekly active users by mid-2025.

The Fifth Edition: From Individual Gains to Collective Productivity

The Microsoft New Future of Work Report 2025 marks the fifth annual edition of what has become one of the most comprehensive analyses of how artificial intelligence is transforming the workplace. Edited by a team of 11 senior Microsoft researchers and drawing on contributions from over 50 authors—including external voices from Harvard Business School, Stanford University, MIT, and Northwestern University—the 2025 edition shifts focus from a question largely answered (does AI help individuals?) to one that remains wide open: how can teams, organizations, and communities achieve collective productivity gains?

As Jaime Teevan, Microsoft’s Chief Scientist and Technical Fellow, frames it in the introduction: “Last year’s report highlighted research showing that AI delivers substantial gains in individual productivity. The next frontier is collective productivity: how teams, organizations, and communities can get better together.” This framing is significant because it acknowledges a gap that many organizations are beginning to feel—despite widespread AI tool adoption, the expected organizational-level transformation hasn’t materialized at the scale many predicted. The report explores why, and what can be done about it.

The historical progression of the series tells its own story about AI’s evolving role. The 2021 edition focused on remote work, 2022 on hybrid return-to-office dynamics, 2023 on how LLMs might reshape everyday work, and 2024 on real-world impact data. The 2025 edition’s pivot to collective intelligence signals a maturation in how researchers and practitioners think about AI at work—moving beyond “does this tool help me?” to “does this tool help us?” Understanding this shift is critical for organizations examining how generative AI impacts critical thinking and decision-making at scale.

AI Adoption and Investment Reach Record Levels

The scale of AI adoption documented in the 2025 report is staggering. Generative AI attracted $33.9 billion in global private investment in 2024—an 18.7% increase from 2023. Enterprise ChatGPT messages increased 8x in the past year. By June 2025, ChatGPT had surpassed 700 million weekly active users globally. Perhaps most notably, the consumer gender gap in ChatGPT usage has completely disappeared—a dramatic shift from early 2023 when more than 80% of users were male.

Workplace adoption data reveals interesting patterns across geographies and demographics. A German survey found 38% of employed respondents used AI for work. In the U.S., men were more likely than women to use generative AI professionally (29.1% vs. 23.5%). The three most common ChatGPT use topics—Practical Guidance, Seeking Information, and Writing—account for approximately 80% of all messages, suggesting that despite the technology’s broad capabilities, users have converged on a relatively narrow set of high-value applications.

However, adoption is not uniform across organizational levels. According to Anthropic’s data, 37% of Claude usage was for tasks associated with Computer and Mathematical occupations, indicating heavy concentration among technical workers. A 2025 IBM survey of 2,000 CEOs across 33 countries found that while most expect AI to transform their businesses, adoption barriers remain significant: organizational inflexibility, risk intolerance, and structural separation between exploration and exploitation. The report notes that many of the best AI applications “come from the edge, not the center”—suggesting that top-down AI mandates may be less effective than empowering individual teams to experiment. Global attitudes vary widely: in China, 83% agree AI has more benefits than drawbacks, compared to just 39% in the United States.

How AI Is Reshaping Productivity at Work

The productivity evidence compiled in the Microsoft New Future of Work Report is extensive and nuanced. Surveyed ChatGPT Enterprise users attribute 40 to 60 minutes saved per day to AI assistance. However, these savings are far from uniform across tasks: legal and management tasks see time savings of 80-85%, while checking diagnostic images yields only about 20% improvement. A privacy-preserving analysis of 50,000 Copilot-enabled Word users showed an average difference of 7 minutes per accepted Copilot output, with 10.7 minutes saved in content editing and 0.6 minutes in applying themes and styles.

These individual-level gains are meaningful. An hour saved per day across an organization represents enormous aggregate value. But the report repeatedly emphasizes a critical disconnect: individual productivity improvements are not translating proportionally to organizational outcomes. Several factors may explain this gap, including the workslop phenomenon discussed in the next section, coordination costs that offset individual time savings, and the challenge of integrating AI-generated outputs across team workflows.

The report also documents how AI capabilities continue to advance at a remarkable pace. METR’s benchmarks show that frontier agents’ reliable task-completion horizons have been rising exponentially with an approximately 7-month doubling time. Multi-turn reinforcement learning for tool-using agents has achieved breakthrough results, with a 14-billion-parameter RL-trained agent scoring 85% versus 78% against frontier-class models on legal document search tasks. These capability gains suggest that the productivity ceiling of AI tools will continue rising, making it even more important for organizations to solve the collective productivity challenge highlighted by the PwC Global CEO Survey 2025.

Transform complex workplace research into interactive experiences your team will actually engage with.

Try It Free →

The Workslop Problem: When AI Output Lacks Substance

One of the report’s most provocative contributions is its analysis of “workslop”—AI-generated work content that appears useful but lacks substance. In a survey of 1,150 U.S. employees, 40% reported receiving workslop in the past month. Researchers estimate that workslop now accounts for approximately 15% of workplace content. The flow patterns are concerning: most slop moves between peers (40%), but it also flows upward to managers (18%) and downward to direct reports (16%), contaminating decision-making at every organizational level.

Workslop may be the hidden explanation for why individual AI productivity gains haven’t aggregated into organizational improvements. If workers save 40 minutes per day producing content faster but some portion of that content requires additional review, verification, or rework by colleagues, the net organizational benefit is reduced. The problem is compounded by the difficulty of identifying workslop—it often looks professional and well-structured on the surface, making detection a cognitive burden for recipients who must now evaluate not just what their colleagues produce but whether AI was involved and whether the output was adequately reviewed.

The workslop phenomenon connects to broader concerns about AI’s impact on workplace trust and norms. The report finds that people who use AI assistance are evaluated as “lazier,” less competent, less diligent, less trustworthy, and less moral. In software engineering, developers who used AI received lower competency ratings on identical work—and this penalty was doubled for women, who received competency scores 13% lower for identical AI-assisted code compared to 6% lower for men. Paradoxically, disclosing AI use can erode trust, creating a perverse incentive to hide it, which further undermines organizational transparency and collective productivity.

Labor Market Impact: Jobs, Skills, and the Junior Worker Squeeze

The labor market data in the Microsoft New Future of Work Report 2025 paints a complex picture that defies simplistic narratives of mass displacement or universal benefit. Large-scale studies conducted in Denmark and the United States found no significant aggregate effect on unemployment, working hours, or job openings. At the macro level, the feared AI-driven unemployment crisis has not materialized—at least not yet.

However, beneath the aggregate stability, significant redistributive effects are underway. Employment for workers aged 22-25 in highly AI-exposed jobs fell by about 13% compared to less-exposed roles. Resume and job posting evidence shows hiring for junior and entry-level positions slowing in AI-exposed occupations after firms adopt AI tools. This “junior worker squeeze” raises profound questions about the future pipeline of experienced professionals—if fewer young workers gain entry-level experience, who will fill senior roles in a decade?

The shift in specific job categories is stark. Writing jobs declined by approximately 30% after ChatGPT’s release. Software, app, and web development jobs fell by about 21%. Engineering positions dropped by 10%. Image-generating AI led to roughly 17% fewer posts in graphic design and 3D modeling. However, the remaining openings tend to be more complex and slightly higher paying, suggesting a flight to quality. Job postings requiring AI skills are growing over 70% year-over-year, extending well beyond technical roles. Roles requiring AI skills are nearly twice as likely to also request analytical thinking, resilience, ethics, or digital literacy. Meanwhile, hiring of AI-specific talent has increased by more than 300% over the past eight years. For context on how AI is reshaping firm-level dynamics, the NBER research on AI, firm productivity, and employment provides complementary evidence.

Human-AI Collaboration: Building Common Ground

The report dedicates significant attention to the mechanics of human-AI collaboration, identifying it as the critical capability that will determine whether organizations unlock collective AI benefits. Central to this analysis is the concept of common ground—the shared knowledge and assumptions essential to effective communication. Current LLMs, the researchers find, generate language with less conversational grounding than humans, instead producing text that presumes shared understanding has already been established. The report identifies 12 specific challenges to establishing common ground in human-agent interaction.

Content creation is evolving from one-shot AI outputs to iterative, multi-turn collaborative refinement. Users rarely accept AI’s initial outputs passively—they engage in complex dialogues characterized by what researchers call Prototypical Human-AI Collaboration Behaviors (PATHs). Shared editing spaces foster greater user control, accuracy, and efficiency than chat-only designs, suggesting that the future of human-AI collaboration lies in environments where both parties can see and modify the same artifact simultaneously.

The expertise dimension adds further complexity. Domain experts prefer delegating routine, low-level tasks to AI while retaining control over high-level analysis, synthesis, and interpretation. The report identifies three distinct types of expertise that matter: expertise in a work domain, expertise in working with AI, and expertise in managing AI agents. This third category—agent management—is emerging as a new professional competency that didn’t exist two years ago. Researchers also found that when clinicians used an LLM through a collaborative workflow that compared AI and clinician perspectives, performance was significantly better than conventional resources and on par with the LLM working alone—suggesting that the right collaboration design can elevate human performance to AI levels.

Help your team explore cutting-edge AI research through engaging, interactive document experiences.

Get Started →

AI and Teams: Why Group Productivity Lags Behind

The team-level findings represent perhaps the report’s most important contribution to the ongoing AI productivity debate. The central finding is clear: AI has been more successful at improving productivity at the individual scale than at the team scale. Research by Dell’Acqua et al. (2025) found that an individual working with AI performed just as well as a pair of humans on laboratory tasks—a striking result that suggests AI can effectively replace a teammate for certain bounded activities but doesn’t necessarily make existing teams better.

Several hypotheses for this gap emerge from the research. AI facilitators boosted a measure of information sharing by 22% in group settings, and proactive information gathering improved user satisfaction by 42%. But neither human nor AI facilitators changed actual group decisions, suggesting that the bottleneck isn’t information flow but rather the human dynamics of consensus, persuasion, and collective sense-making. Standard RLHF alignment techniques fail to maintain reliability in extended multi-turn and multi-party dialogues, meaning current AI models are fundamentally optimized for one-on-one interaction, not group dynamics.

The report explores two competing visions for AI’s impact on team structure. One possibility is much larger, more ephemeral teams—possibly challenging traditional organizational boundaries as AI reduces coordination costs. The alternative is that teams get much smaller, embodied in the “one-person unicorn” hypothesis where AI enables individuals to accomplish what previously required entire teams. The researchers note that LLMs themselves are “amazing new forms of collective intelligence,” suggesting that “collective intelligence” may be a more accurate term than “artificial intelligence.” Four prototypical AI team roles emerge from the research: Coordinator, Creator, Perfectionist, and Doer—and different collaboration settings benefit from different AI behaviors, with no single AI persona fitting all teams. For deeper analysis of how AI systems interact with research methodologies, see this guide on deep research systems and survey methodologies.

Cognitive Deskilling and the Thinking Deficit

Among the report’s most alarming findings is the evidence for cognitive deskilling—the gradual erosion of human expertise through over-reliance on AI systems. The mechanism is straightforward: AI shifts effort from “thinking by doing” to “choosing from outputs,” potentially reducing the judgment needed to build and maintain expertise. The real-world evidence is compelling and concerning.

In a clinical setting, healthcare professionals who relied on AI for polyp detection during colonoscopies showed significant decline in their independent diagnostic ability after just three months of AI-assisted practice. This finding carries implications far beyond medicine. Any profession where expertise develops through repeated practice and pattern recognition—law, engineering, financial analysis, software architecture—faces a similar risk. If AI handles the routine cognitive work that builds expertise, the pipeline of skilled professionals may be undermined even as productivity appears to increase in the short term.

The creativity data tells a parallel story. Using LLMs before independent ideation reduced original ideas and lowered creative self-efficacy, compared to using AI after initial brainstorming. Those who performed unassisted ideation with no LLM exposure at all had the best creative outcomes. LLM-generated strategies reduced idea diversity, and these effects continued even after AI use stopped, suggesting a form of creative contamination. LLM support also significantly reduced group elaboration of ideas—the collective building process where teams refine and extend each other’s thinking.

Not all findings are pessimistic. The report identifies contrastive explanations—AI outputs that explain why one answer is preferred over alternatives—as a technique that can enhance users’ independent decision-making skills without sacrificing accuracy. Cognitively engaging AI designs that encourage reflection rather than passive acceptance reduced overreliance on incorrect AI outputs. These findings point toward a design philosophy where AI tools deliberately introduce productive friction rather than optimizing purely for speed and convenience. Organizations tracking workforce capabilities should also consider how these dynamics interact with broader trends documented in comprehensive LLM survey guides.

AI in Education: Promise, Peril, and the Equity Gap

The education findings in the Microsoft New Future of Work Report reveal both extraordinary adoption rates and troubling equity dynamics. An astonishing 80% of K-12 teachers and 95% of higher education educators have used AI at least once. Among students, the numbers are even higher: 90% of K-12 students and 95% of higher education students have tried AI tools. These adoption rates dwarf those in most professional settings and suggest that the next generation of workers will arrive at their jobs with significant AI experience—for better or worse.

The promise is substantial. An AI tutor using pedagogical best practices helped students learn more, faster, than an active-learning classroom—one of the gold standards of educational research. This finding suggests that personalized AI tutoring could transform education at scale, particularly in resource-constrained environments where skilled human tutors are scarce. The report also notes that two-thirds of employees report trusting AI agents, indicating growing workplace comfort with AI-driven learning and development.

The peril, however, is equally significant. Learning benefits from AI thus far favor students of higher socioeconomic status, meaning AI in education risks widening rather than narrowing achievement gaps. Only about half of surveyed school districts reported providing training to teachers on AI use, leaving educators to navigate AI integration largely on their own. The concern about AI use in education mirrors the OECD’s analysis of AI in education, which emphasizes the need for deliberate equity-focused policies. Additionally, 60% of employees skip accuracy checks on AI output—a habit that, if formed during education, could persist throughout careers.

What Leaders Should Do Now

The Microsoft New Future of Work Report 2025 implicitly but powerfully argues that the next phase of AI value creation requires a fundamentally different approach than the first. Individual productivity tools have been deployed and are delivering measurable returns. The collective productivity frontier—making teams, departments, and entire organizations more effective through AI—demands new strategies, new metrics, and new organizational designs.

For executives, the most urgent action is addressing the workslop problem before it erodes the trust and quality standards that make organizations functional. This means establishing clear norms around AI disclosure, creating review processes calibrated for AI-assisted content, and measuring output quality alongside output quantity. The Harvard Business Review’s ongoing AI research provides complementary frameworks for organizational AI governance.

For talent leaders, the junior worker squeeze demands immediate attention. If entry-level hiring slows in AI-exposed roles, organizations must create alternative pathways for developing expertise—apprenticeships, rotation programs, and deliberate “AI-free zones” where young professionals build foundational skills through hands-on practice. The cognitive deskilling evidence suggests that AI tools should be designed (or configured) to support learning, not just task completion.

For technology leaders, the team collaboration gap represents the largest unaddressed opportunity. Current AI tools are optimized for individual interaction. The next generation of workplace AI must be designed for multi-party, multi-turn collaboration—supporting the messy, iterative, politically charged process of real teamwork. The report’s findings on AI team roles, facilitation, and common ground provide a research foundation for this next wave of product development. Organizations that solve collective AI productivity first will gain a decisive advantage over those still focused on individual tool optimization.

Share AI workplace research with your entire organization through interactive, engaging document experiences.

Start Now →

Frequently Asked Questions

What are the key findings of the Microsoft New Future of Work Report 2025?

The Microsoft New Future of Work Report 2025 focuses on collective productivity—how teams, organizations, and communities can benefit from AI together. Key findings include: ChatGPT Enterprise users save 40-60 minutes daily, 40% of employees receive AI-generated “workslop,” AI improves individual but not yet team-level productivity, the gender gap in ChatGPT usage has disappeared, and $33.9 billion was invested globally in generative AI in 2024.

What is AI workslop and why does it matter?

AI workslop refers to AI-generated work content that appears useful but lacks substance. A survey of 1,150 U.S. employees found that 40% received workslop in the past month, estimating it accounts for 15% of workplace content. Most flows between peers (40%), but also moves upward (18%) and downward (16%) in hierarchies. Workslop may explain why individual AI productivity gains haven’t translated to organizational-level improvements.

How is AI affecting the job market in 2025?

The labor market impact is nuanced. Large-scale studies in Denmark and the U.S. found no significant aggregate effect on unemployment or working hours. However, employment for workers aged 22-25 in AI-exposed jobs fell by about 13%, writing jobs declined by 30%, and software jobs by 21% after ChatGPT’s launch. Meanwhile, hiring of AI talent increased by over 300% in eight years, and job postings requiring AI skills grew over 70% year-over-year.

What is cognitive deskilling from AI use?

Cognitive deskilling occurs when AI shifts effort from “thinking by doing” to “choosing from outputs,” potentially reducing the judgment needed to build expertise. A real-world example shows clinicians who relied on AI for polyp detection experienced significant decline in independent diagnostic ability after just three months. The report recommends contrastive explanations and cognitively engaging AI designs that encourage reflection to mitigate this risk.

How does AI impact team collaboration versus individual productivity?

AI has been more successful at improving individual productivity than team-level performance. Research shows an individual with AI performed as well as a pair of humans on laboratory tasks. However, AI facilitators boosted information sharing by 22% in group settings. The report identifies challenges including standard alignment techniques failing in multi-turn dialogues, reduced social interaction in AI-mediated teams, and the need for different AI personas for different team configurations.

What percentage of teachers and students use AI in education?

AI adoption in education is remarkably high: 80% of K-12 teachers and 95% of higher education educators have used AI at least once, while 90% of K-12 students and 95% of higher education students have tried AI tools. However, the report warns that learning benefits from AI currently favor students of higher socioeconomic status, and only about half of surveyed school districts provide teacher training on AI use.

Your documents deserve to be read.

PDFs get ignored. Presentations get skipped. Reports gather dust.

Libertify transforms them into interactive experiences people actually engage with.

No credit card required · 30-second setup