Instead of banning AI, educators should teach students how to prompt it effectively to improve their decision-making. This includes forcing it to cite sources, generate counterarguments, and explain its reasoning, turning AI into a tool for critical inquiry rather than just an answer machine.

Related Insights

Schools ban AI like ChatGPT fearing it's a tool for cheating, but this is profoundly shortsighted. The quality of an AI's output is entirely dependent on the critical thinking behind the user's input. This makes AI the first truly scalable tool for teaching children how to think critically, a skill far more valuable than memorization.

To sharpen your thinking, use ChatGPT as a Socratic partner. Feed it your argument and ask it to generate both supporting points and strong counterarguments. This dialectical process helps you anticipate objections and refine your position, leading to a more robust final synthesis.

Before publishing, feed your work to an AI and ask it to find all potential criticisms and holes in your reasoning. This pre-publication stress test helps identify blind spots you would otherwise miss, leading to stronger, more defensible arguments.

The most effective way to use AI in product discovery is not to delegate tasks to it like an "answer machine." Instead, treat it as a "thought partner." Use prompts that explicitly ask it to challenge your assumptions, turning it into a tool for critical thinking rather than a simple content generator.

Log your major decisions and expected outcomes into an AI, but explicitly instruct it to challenge your thinking. Since most AIs are designed to be agreeable, you must prompt them to be critical. This practice helps you uncover flaws in your logic and improve your strategic choices.

ASU's president argues that if an AI can answer an assignment, the assignment has failed. The educator's role must evolve to use AI to 'up the game,' forcing students to ask more sophisticated questions, making the quality of the query—not the synthesized answer—the hallmark of learning.

AI models tend to be overly optimistic. To get a balanced market analysis, explicitly instruct AI research tools like Perplexity to act as a "devil's advocate." This helps uncover risks, challenge assumptions, and makes it easier for product managers to say "no" to weak ideas quickly.

A leader's most valuable use of AI isn't for automation, but as a constant 'thought partner.' By articulating complex business, legal, or financial decisions to an AI and asking it to pose clarifying questions, leaders can refine their own thinking and arrive at more informed conclusions, much like talking a problem out loud.

Instead of allowing AI to atrophy critical thinking by providing instant answers, leverage its "guided learning" capabilities. These features teach the process of solving a problem rather than just giving the solution, turning AI into a Socratic mentor that can accelerate learning and problem-solving abilities.

Standard AI models are often overly supportive. To get genuine, valuable feedback, explicitly instruct your AI to act as a critical thought partner. Use prompts like "push back on things" and "feel free to challenge me" to break the AI's default agreeableness and turn it into a true sparring partner.