As generative AI spreads across offices, many organizations report little measurable return on investment, raising questions about how the tools are being used and managed. Research cited by workplace analysts points to a growing problem: shiny outputs that mask shallow thinking and shift work to colleagues.
The concern is emerging now because adoption is high and expectations are higher. Leaders want faster content creation and decision support. Instead, teams often report more rework, more confusion, and less trust in shared outputs.
The Rise of “Workslop” and Its Costs
Researchers describe a pattern of content that looks presentable but lacks substance. The term now used inside many teams is workslop.
“Workslop—content that appears polished but lacks real substance, offloading cognitive labor onto coworkers.”
According to research from BetterUp Labs and Stanford, 41% of workers have faced this kind of AI-generated output. Each instance takes nearly two hours to fix. That time loss multiplies across projects, pushing deadlines and sapping momentum.
“Research from BetterUp Labs and Stanford found that 41% of workers have encountered such AI-generated output, costing nearly two hours of rework per instance.”
Workers also report damage to trust and collaboration when AI drafts circulate without careful review.
“Creating downstream productivity, trust, and collaboration issues.”
Why ROI Remains Elusive
Many companies set broad AI mandates but give little guidance on quality. That puts pressure on teams to produce more, not better. Analysts say this can push tools into roles they are not suited to fill, such as expert analysis without human oversight.
“Leaders need to consider how they may be encouraging indiscriminate organizational mandates and offering too little guidance on quality standards.”
The mismatch between expectation and execution helps explain why some organizations see limited gains despite heavy usage. Without clear standards, teams chase speed and volume. The result is more editing, more friction, and less confidence in shared work.
What Leaders Can Do Now
Researchers argue that leaders play a central role in reversing these trends. Practical steps focus on modeling good use, setting norms, and clarifying when AI should support, not replace, human judgment.
“To counteract workslop, leaders should model purposeful AI use, establish clear norms, and encourage a ‘pilot mindset’ that combines high agency with optimism—promoting AI as a collaborative tool, not a shortcut.”
- Set quality bars for AI outputs and require human review.
- Define use cases where AI drafting helps, and where it does not.
- Track rework time to spot problem areas and improve workflows.
- Share examples of effective AI prompts and edits across teams.
The shift is cultural as much as technical. Early pilots with clear goals, lightweight metrics, and open feedback help teams learn fast without overpromising.
Balancing Speed With Substance
Many workers welcome AI for first drafts and idea generation. The trouble comes when drafts are treated as final. Teams that gain the most treat AI output as raw material, then apply domain knowledge, source checks, and editing.
That approach protects trust within teams. It also reduces the two-hour rework spiral reported in the research. Over time, better prompts and clearer standards raise the baseline for quality.
What to Watch Next
The central question is not whether AI adds value, but under what conditions. Organizations that define quality, measure rework, and train for thoughtful use are most likely to see measurable gains. Those that rely on volume alone risk more workslop and weaker outcomes.
For now, the message from the research is clear: speed without judgment does not pay. Leaders can restore value by pairing tools with standards, modeling best practices, and making AI a true partner in skilled work.
The next phase will test whether these practices reduce rework and rebuild trust. If they do, ROI may follow.