DataFramed

DataCamp
DataFramed
Latest episode

353 episodes

  • DataFramed

    #354 Beyond BI: Decision Intelligence with Graphs with Jamie Hutton, CTO at Quantexa

    04/06/2026 | 46 mins.
    Decision intelligence is showing up across data and AI teams as companies move beyond dashboards to decisions made with context. Graphs, entity resolution, and better data products are becoming core tools as messy, siloed data meets stricter risk and compliance needs. In day-to-day work, this means linking “James,” “Jim,” and “Jamie” across systems, enriching records with third‑party sources, and pushing models where the data already lives in your lakehouse. How do you trust your customer counts? Which links in a graph matter, and which are noise? Can graph-based context reduce LLM hallucinations enough for regulated decisions with humans still in-loop.
    Jamie Hutton is the Co-founder and Chief Technology Officer of Quantexa, where he leads the company’s global research and development organization in advancing its market-leading Decision Intelligence Platform. With over two decades of experience pioneering data-driven technologies, Jamie has been at the forefront of innovations that connect and unify data at scale to solve complex real-world challenges. He is the creator of dynamic Entity Resolution, a pioneering capability that has redefined how the world’s leading organizations transform raw data into trusted, decision-ready intelligence. This innovation enables enterprises to prepare their data for AI, uncover new revenue streams, and expose hidden connections in even the most sophisticated criminal networks. By providing the foundation for accurate, explainable, and actionable insights, Jamie’s work has empowered governments, financial institutions, and global enterprises to make faster, smarter, and more confident decisions.
    Prior to co-founding Quantexa, Jamie held senior technology and analytics leadership roles at SAS and Detica, where he delivered mission-critical solutions for organizations operating in some of the most complex and high-stakes environments in the world. Jamie holds a First-Class master’s degree in computer engineering and is recognized as a leading authority in contextual analytics, data integration, and applied AI for mission-critical decision-making.
    In the episode, Richie and Jamie explore decision intelligence beyond BI, entity resolution across siloed data, building context graphs for fraud, AML, credit risk, and growth, how graph analytics separates meaningful links from noise, graph-RAG for LLMs to cut hallucinations, human-in-the-loop workflows, and ways to start today, and much more.
    Links Mentioned in the Show:
    Quantexa
    Dun & Bradstreet Data Enrichment
    Connect with Jamie
    AI-Native Course: Intro to AI for Work
    Related Episode: How Optimization Powers Decision Intelligence with Duke Perrucci & Ed Klotz, CEO and Senior Mathematical Optimization Specialist at Gurobi Optimization
    Explore AI-Native Learning on DataCamp

    New to DataCamp?
    Learn on the go using the DataCamp mobile app

    Empower your business with world-class data and AI skills with DataCamp for business
  • DataFramed

    #353 The Data Team's Agentic Future with Ketan Karkhanis, CEO at ThoughtSpot

    03/30/2026 | 49 mins.
    Data and AI platforms are racing toward agentic and even autonomous analytics. But the bottleneck is rarely the model—it’s data readiness: governed metrics, clear metadata, and a semantic layer machines can read. For data engineers and analysts, this shifts work from hand-built SQL and dashboard tweaks to designing meaning and trust. If an agent can draft column descriptions, propose a model for a new business question, and build the first dashboard layout, where do you add the most value? What do you measure to prove ROI in 30 days? How do you prevent “shiny demos” from driving strategy too early.
    Ketan Karkhanis is the CEO of ThoughtSpot. Prior to joining the company in September 2024, Ketan was the Executive Vice President and General Manager of Sales Cloud at Salesforce. He returned to Salesforce in March 2022 after his time as the COO of Turvo, an emerging supply-chain collaboration platform. Before that, Ketan spent nearly a decade at Salesforce, where he led product areas in Sales, Service Cloud, Lightning Platform, and finally Analytics, wherein as the Senior Vice President & GM of Einstein Analytics, he pioneered incredible innovation, customer success, and business acceleration from launch to over $300M and a 30,000 strong user community. Prior to Salesforce, Ketan was at Cisco Systems where he led various technology initiatives and initiatives spanning Customer Advocacy, Cisco Certifications & eLearning.
    In the episode, Richie and Ketan explore AI agents for analytics, why “self‑service BI” often fails, using agents to answer questions, build dashboards, and automate data modeling, how analyst and engineer roles shift toward governance and agent design, how transparency, culture, and ROI drive safe adoption, and much more.
    Links Mentioned in the Show:
    Thoughtspot
    Thoughspot’s Spotter Agents
    Connect with Ketan
    AI-Native Course: Intro to AI for Work
    Related Episode: AI Agents at Work: What Actually Breaks (and How to Fix It) with Danielle Crop, EVP Digital Strategy & Alliances at WNS
    Explore AI-Native Learning on DataCamp

    New to DataCamp?
    Learn on the go using the DataCamp mobile app
    Empower your business with world-class data and AI skills with DataCamp for business
  • DataFramed

    #352 AI Agents at Work: What Actually Breaks (and How to Fix It) with Danielle Crop, EVP Digital Strategy & Alliances at WNS

    03/23/2026 | 56 mins.
    AI agents are spreading across the data and AI industry, promising to automate everything from research to outreach. At the same time, teams are learning that these tools can hallucinate, leak data, or act in surprising ways. In day-to-day work, the challenge is deciding which tasks to hand off, what data to share, and how to keep the output trustworthy. Do your agents actually add value, or just add noise? Are they running in a secured, ring-fenced environment? How do you balance playful experimentation with critical checking when an agent confidently gets a key fact wrong?
    Danielle leads go-to-market strategy at WNS, Capgemini's AI transformation services arm. Previously, Danielle was Chief Data Officer at American Express and Albertsons. She also write The Remix substack on technology trends, and is an Editorial Board Member for CDO Magazine.
    In the episode, Richie and Danielle explore AI agents at work, experimentation with guardrails, data privacy, access, tone controls, OpenClaw automation wins and failures, token costs, tying AI plans to P&L strategy, shifts in careers and hiring, how data teams handle unstructured data governance, and much more.
    Links Mentioned in the Show:
    WNS
    Connect with Danielle
    AI-Native Course: Intro to AI for Work
    Catch Danielle speaking at RADAR—April 1
    Related Episode: AI Agents Are the New Shadow IT (And Your Governance Isn’t Ready) with Stijn Christiaens, CEO at Collibra
    Explore AI-Native Learning on DataCamp

    New to DataCamp?
    Learn on the go using the DataCamp mobile app

    Empower your business with world-class data and AI skills with DataCamp for business
  • DataFramed

    #351 Will World Models Bring us AGI? with Eric Xing, President & Professor at MBZUAI

    03/16/2026 | 1h 3 mins.
    World models are emerging as the next step after large language models, pushing AI from book knowledge toward systems that can simulate the physical and social world. Instead of just generating text or short videos, the goal is steerable simulation with long-horizon consistency and planning. For practitioners, this raises practical choices: what data and representations do you need, and when do you mix symbolic reasoning with generative models? How do you test whether a model can follow actions over minutes, not seconds? And where do you start—robotics, driving safety, or synthetic data generation?
    Professor Eric Xing is President of Mohamed bin Zayed University of Artificial Intelligence (MBZUAI) and a world-leading computer scientist whose work spans statistical machine learning, distributed systems, computational biology, and healthcare AI. A fellow of AAAI, IEEE, and the American Statistical Association, he has authored over 400 research papers cited more than 44,000 times.Before MBZUAI, Eric was a Professor of Computer Science at Carnegie Mellon University, where he also founded the Center for Machine Learning and Health. He is the founder and chief scientist of Petuum Inc., recognized as a World Economic Forum Technology Pioneer, and has held visiting roles at Stanford and Facebook. He holds PhDs in both Molecular Biology and Computer Science.
    In the episode, Richie and Eric explore world models as simulators for action, the jump from book intelligence to physical and social skills, why long-horizon planning is still hard, architectures, robots, data generation, open K2 Think LLMs, virtual-cell biology, and much more.
    Links Mentioned in the Show:
    MBZUAI
    Pan World Model
    Connect with Eric
    AI-Native Course: Intro to AI for Work
    Related Episode: Developing Better Predictive Models with Graph Transformers with Jure Leskovec, Pioneer of Graph Transformers, Professor at Stanford
    Explore AI-Native Learning on DataCamp

    New to DataCamp?
    Learn on the go using the DataCamp mobile app

    Empower your business with world-class data and AI skills with DataCamp for business
  • DataFramed

    #350 How to Make Hard Choices in AI with Atay Kozlovski, Researcher at the University of Zurich

    03/09/2026 | 1h 10 mins.
    Across the AI industry, high-stakes tools are being deployed in places where errors can harm people: sepsis alerts in hospitals, identity checks, welfare fraud detection, immigration enforcement, and recommendation systems that shape life outcomes. The pattern is familiar: scale and speed go up, while human review becomes rushed, shallow, or punished for disagreeing. In daily work, that can look like a nurse forced to act on false alarms, or a team using an LLM summary in ways the designers never planned. When should you slow down deployment? How do you detect new “wild” use cases early? And what does responsible tracking and oversight look like under real pressure?
    Atay Kozlovski is a Postdoctoral Researcher at the University of Zurich’s Center for Ethics. He holds a PhD in Philosophy from the University of Zurich, an MA in PPE from the University of Bern, and a BA from Tel Aviv University. His current research focuses on normative ethics, hard choices, and the ethics of AI.
    In the episode, Richie and Atay explore why AI failures keep happening, from automation bias to opaque targeting and hiring models. They unpack “meaningful human control,” accountability, and design in healthcare, government, and warfare. You’ll also hear about deepfakes, consent, digital twins, and AI-driven civic engagement, and much more.
    Links Mentioned in the Show:
    “Lavender” IDF recommendation system
    Amnesty International reports on AI/automation in welfare systems
    “Meaningful Human Control” (MHC) framework
    Connect with Atay
    AI-Native Course: Intro to AI for Work
    Related Episode: Harnessing AI to Help Humanity with Sandy Pentland, HAI Fellow at Stanford
    Explore AI-Native Learning on DataCamp

    New to DataCamp?
    Learn on the go using the DataCamp mobile app

    Empower your business with world-class data and AI skills with DataCamp for business

More Business podcasts

About DataFramed

Welcome to DataFramed, a weekly podcast exploring how artificial intelligence and data are changing the world around us. On this show, we invite data & AI leaders at the forefront of the data revolution to share their insights and experiences into how they lead the charge in this era of AI. Whether you're a beginner looking to gain insights into a career in data & AI, a practitioner needing to stay up-to-date on the latest tools and trends, or a leader looking to transform how your organization uses data & AI, there's something here for everyone. Join co-hosts Adel Nehme and Richie Cotton as they delve into the stories and ideas that are shaping the future of data. Subscribe to the show and tune in to the latest episode on the feed below.
Podcast website

Listen to DataFramed, The Learning Leader Show With Ryan Hawk and many other podcasts from around the world with the radio.net app

Get the free radio.net app

  • Stations and podcasts to bookmark
  • Stream via Wi-Fi or Bluetooth
  • Supports Carplay & Android Auto
  • Many other app features