Professionals are using AI to write detailed reports, while their managers use AI to summarize them. This creates a feedback loop where AI generates content for other AIs to consume, with humans acting merely as conduits. This "AI slop" replaces deep thought with inefficient, automated communication.
The problem with bad AI-generated work ('slop') isn't just poor writing. It's that subtle inaccuracies or context loss can derail meetings and create long, energy-wasting debates. This cognitive overload makes it difficult for teams to sense-make and ultimately costs more in human time than it saves.
While AI tools once gave creators an edge, they now risk producing democratized, undifferentiated output. IBM's AI VP, who grew to 200k followers, now uses AI less. The new edge is spending more time on unique human thinking and using AI only for initial ideation, not final writing.
Using AI to generate content without adding human context simply transfers the intellectual effort to the recipient. This creates rework, confusion, and can damage professional relationships, explaining the low ROI seen in many AI initiatives.
Users who treat AI as a collaborator—debating with it, challenging its outputs, and engaging in back-and-forth dialogue—see superior outcomes. This mindset shift produces not just efficiency gains, but also higher quality, more innovative results compared to simply delegating discrete tasks to the AI.
The true danger of LLMs in the workplace isn't just sloppy output, but the erosion of deep thinking. The arduous process of writing forces structured, first-principles reasoning. By making it easy to generate plausible text from bullet points, LLMs allow users to bypass this critical thinking process, leading to shallower insights.
While AI can accelerate tasks like writing, the real learning happens during the creative process itself. By outsourcing the 'doing' to AI, we risk losing the ability to think critically and synthesize information. Research shows our brains are physically remapping, reducing our ability to think on our feet.
Research highlights "work slop": AI output that appears polished but lacks human context. This forces coworkers to spend significant time fixing it, effectively offloading cognitive labor and damaging perceptions of the sender's capability and trustworthiness.
GSB professors warn that professionals who merely use AI as a black box—passing queries and returning outputs—risk minimizing their own role. To remain valuable, leaders must understand the underlying models and assumptions to properly evaluate AI-generated solutions and maintain control of the decision-making process.
Companies racing to add AI features while ignoring core product principles—like solving a real problem for a defined market—are creating a wave of failed products, dubbed "AI slop" by product coach Teresa Torres.
The ease of generating AI summaries is creating low-quality 'slop.' This imposes a hidden productivity cost, as collaborators must waste time clarifying ambiguous or incorrect AI-generated points, derailing work and leading to lengthy, unnecessary corrections.