Get your free personalized podcast brief

We scan new podcasts and send you the top 5 insights daily.

Intense pressure to hit goals corrupts data-driven cultures. Teams may block improvements to A/B testing tools if accurate results threaten a 'win'. This pathology extends to shipping features solely to meet a deadline, with a plan to delete the code immediately after the performance review cycle ends.

Related Insights

The proliferation of AI leaderboards incentivizes companies to optimize models for specific benchmarks. This creates a risk of "acing the SATs" where models excel on tests but don't necessarily make progress on solving real-world problems. This focus on gaming metrics could diverge from creating genuine user value.

By ranking engineers on AI token consumption, Meta is experiencing Goodhart's Law: "When a measure becomes a target, it ceases to be a good measure." Employees reportedly build bots to needlessly burn tokens for status, demonstrating how gamifying a proxy metric can backfire and disconnect from actual business impact.

In large companies, a culture of A/B testing every decision can become a crutch that stifles innovation and speed. It leads to risk aversion and organizational lethargy, as teams lose the muscle for making convicted, gut-based decisions informed by qualitative customer feedback.

Foster a culture of experimentation by reframing failure. A test where the hypothesis is disproven is just as valuable as a 'win' because it provides crucial user insights. The program's success should be measured by the quantity of quality tests run, not the percentage of successful hypotheses.

The speaker observed a pattern at Meta where leadership sets ambitious, often unrealistic deadlines. When these are consistently missed without consequence, the pressure becomes artificial. This erodes motivation, causing engineers to disengage and treat the deadlines as noise rather than serious goals.

Measuring engineering success with metrics like velocity and deployment frequency (DORA) incentivizes shipping code quickly, not creating customer value. This focus on output can actively discourage the deep product thinking required for true innovation.

According to Goodhart's Law, when a measure becomes a target, it ceases to be a good measure. If you incentivize employees on AI-driven metrics like 'emails sent,' they will optimize for the number, not quality, corrupting the data and giving false signals of productivity.

Setting rigid targets incentivizes employees to present favorable numbers, even subconsciously. This "performance theater" discourages them from investigating negative results, which are often the source of valuable learning. The muscle for detective work atrophies, and real problems remain hidden beneath good-looking metrics.

The "hamster wheel of execution" persists because performance reviews and incentives overwhelmingly focus on shipped features. Until companies tangibly reward strategic vision and planning, PMs will continue to prioritize execution, regardless of time saved by tools like AI.

Teams often self-limit output because they know overperformance will simply raise future targets to unsustainable levels. This "prison of expectations" incentivizes predictable mediocrity over breakthrough results, as employees actively manage goals to avoid future failure.

Performance Goals Drive Teams to Game A/B Tests and Ship Throwaway Code | RiffOn