Description:AI agents from OpenAI, Google, and Anthropic promise to act on your behalf—booking flights, handling tasks, making decisions. What kind of agency do these systems actually have? And whose interests are they serving?Enterprise AI agents are already deployed in customer support, code generation, and task automation. Consumer agents—ChatGPT Agent Mode, personal task assistants—face a wider gap between marketing promises and actual capabilities.The alignment problem: agents need access to your calendar, email, and personal preferences to help you effectively. But the agent that knows you well enough to serve you is also positioned to steer you. When you delegate decisions to an agent, who decides what success looks like?To stay in touch, sign up for our newsletter at https://www.superprompt.fm
--------
15:50
--------
15:50
AI Safety: Constitutional AI vs Human Feedback
With great power comes great responsibility. How do leading AI companies implement safety and ethics as language models scale? OpenAI uses Model Spec combined with RLHF (Reinforcement Learning from Human Feedback). Anthropic uses Constitutional AI. The technical approaches to maximizing usefulness while minimizing harm. Solo episode on AI alignment.REFERENCEOpenAI Model Spechttps://cdn.openai.com/spec/model-spec-2024-05-08.html#overviewAnthropic Constitutional AIhttps://www.anthropic.com/news/claudes-constitutionTo stay in touch, sign up for our newsletter at https://www.superprompt.fm
--------
16:38
--------
16:38
Open Source LLMs: How Open Is "Open"?
Notable open source large language models from Meta, French AI company Mistral (valued at $2B), Microsoft, and Apple. Not all open source models are equally open—the restrictions and licensing constraints you need to know before deploying one. Solo episode.To stay in touch, sign up for our newsletter at https://www.superprompt.fm
--------
13:28
--------
13:28
Open Source AI: The Safety Debate
Why enterprises and entrepreneurs choose open source LLMs like Meta's Llama—cost-effectiveness, control, privacy, and security. The safety and ethics debate: which poses greater risk to humanity, open source or proprietary AI models? Both? Neither? Solo episode on open source LLMs.To stay in touch, sign up for our newsletter at https://www.superprompt.fm
--------
16:29
--------
16:29
LLM Benchmarks: How to Know Which AI Is Better
Beyond ChatGPT and Gemini: Anthropic's Claude and the $4 billion Amazon investment. How AI industry benchmarks work, including LMSYS Arena Elo and MMLU (Measuring Massive Multitask Language Understanding). How benchmarks are constructed, what they measure, and how to use them to evaluate LLMs. Solo episode.Anthropic's Claude https://claude.ai [Note: I am not sponsored by Anthropic]LMSYS Leaderboardhttps://chat.lmsys.org/?leaderboardTo stay in touch, sign up for our newsletter at https://www.superprompt.fm
Examining generative AI—not to hype breakthroughs or warn of apocalypse, but to understand how things actually work. Mental models over hot takes. Technology specifics over marketing fog.Welcome to Super Prompt. Hosted by Tony Wan, ex-Silicon Valley insider.For The Independents—people who think for themselves, refuse narrative capture, and value depth over certainty.Independent analysis. Unsponsored. Weekly. The future belongs to better questions.