ออฟไลน์ด้วยแอป Player FM !
S3E2 - Intentional AI: Maximizing AI for research & analysis
Manage episode 513584392 series 2422944
In Episode 2 of the Intentional AI series, Cole and Virgil dive into the first real stage of the content lifecycle: research and analysis.
From brainstorming ideas to verifying data sources, AI is being used everywhere in the early stages of content creation. But how much of that information can you actually trust? In this episode, the team unpacks where AI helps, where it hurts, and why you still need to be the researcher of the research.
In this episode, they explore:
- How AI fits into the research and analysis stage of the content lifecycle
- The major risks of using AI for research, including accuracy, bias, and misinformation
- Why trust, verification, and validation are now part of your job
- Security and legal concerns around AI scraping and data usage
- How different tools handle citations, transparency, and usability
- Why you can’t skip the human role in confirming, editing, and contextualizing AI outputs
This episode also features the first step in a real experiment: researching a blog topic on digital accessibility using the tools Perplexity, ChatGPT, and Copilot. The results of that research will directly fuel the next episode on content creation.
A downloadable Episode Companion Guide is available below. It includes key episode takeaways, tool comparisons, and practical guidance on how to use AI responsibly during the research stage.
DS-S3-E2-CompanionDoc.pdf
Upcoming episodes in the Intentional AI series:
- Oct 28, 2025 — Content Creation
- Nov 11, 2025 — Content Management
- Dec 2, 2025 — Accessibility
- Dec 16, 2025 — SEO / AEO / GEO
- Jan 6, 2026 — Content Personalization
- Jan 20, 2026 — Front End Development & Wireframing
- Feb 3, 2026 — Design & Media
- Feb 17, 2026 — Back End Development
- Mar 3, 2026 — Conversational Search (with special guest!)
- Mar 17, 2026 — Chatbots & Agentic AI
- Mar 31, 2026 — Series Finale & Tool Review
Whether you’re a marketer, strategist, or developer, this conversation is about making AI adoption intentional and keeping your critical thinking sharp.
New episodes every other Tuesday.
For more conversations about AI, digital strategy, and all the ways we get it wrong (and how to get it right), visit www.discussingstupid.com and subscribe on your favorite podcast platform.
(0:00) - Intro
(1:44) - Better research with AI
(3:46) - Risk: Trust & reliability
(5:29) - Risk: Security/legal concerns
(7:04) - Risk: Hallucinations
(9:17) - We tested 3 tools for AI research
(11:03) - Testing Perplexity
(14:38) - Testing ChatGPT
(17:45) - Testing Copilot
(19:54) - Comparing the tools and key takeaways
(20:52) - Outro
Subscribe for email updates on our website:
https://www.discussingstupid.com/
Watch us on YouTube:
https://www.youtube.com/@discussingstupid
Listen on Apple Podcasts, Spotify, or Soundcloud:
https://open.spotify.com/show/0c47grVFmXk1cco63QioHp?si=87dbb37a4ca441c0
https://soundcloud.com/discussing-stupid
Check Us Out on Socials:
https://www.linkedin.com/company/discussing-stupid
https://www.instagram.com/discussingstupid/
31 ตอน
Manage episode 513584392 series 2422944
In Episode 2 of the Intentional AI series, Cole and Virgil dive into the first real stage of the content lifecycle: research and analysis.
From brainstorming ideas to verifying data sources, AI is being used everywhere in the early stages of content creation. But how much of that information can you actually trust? In this episode, the team unpacks where AI helps, where it hurts, and why you still need to be the researcher of the research.
In this episode, they explore:
- How AI fits into the research and analysis stage of the content lifecycle
- The major risks of using AI for research, including accuracy, bias, and misinformation
- Why trust, verification, and validation are now part of your job
- Security and legal concerns around AI scraping and data usage
- How different tools handle citations, transparency, and usability
- Why you can’t skip the human role in confirming, editing, and contextualizing AI outputs
This episode also features the first step in a real experiment: researching a blog topic on digital accessibility using the tools Perplexity, ChatGPT, and Copilot. The results of that research will directly fuel the next episode on content creation.
A downloadable Episode Companion Guide is available below. It includes key episode takeaways, tool comparisons, and practical guidance on how to use AI responsibly during the research stage.
DS-S3-E2-CompanionDoc.pdf
Upcoming episodes in the Intentional AI series:
- Oct 28, 2025 — Content Creation
- Nov 11, 2025 — Content Management
- Dec 2, 2025 — Accessibility
- Dec 16, 2025 — SEO / AEO / GEO
- Jan 6, 2026 — Content Personalization
- Jan 20, 2026 — Front End Development & Wireframing
- Feb 3, 2026 — Design & Media
- Feb 17, 2026 — Back End Development
- Mar 3, 2026 — Conversational Search (with special guest!)
- Mar 17, 2026 — Chatbots & Agentic AI
- Mar 31, 2026 — Series Finale & Tool Review
Whether you’re a marketer, strategist, or developer, this conversation is about making AI adoption intentional and keeping your critical thinking sharp.
New episodes every other Tuesday.
For more conversations about AI, digital strategy, and all the ways we get it wrong (and how to get it right), visit www.discussingstupid.com and subscribe on your favorite podcast platform.
(0:00) - Intro
(1:44) - Better research with AI
(3:46) - Risk: Trust & reliability
(5:29) - Risk: Security/legal concerns
(7:04) - Risk: Hallucinations
(9:17) - We tested 3 tools for AI research
(11:03) - Testing Perplexity
(14:38) - Testing ChatGPT
(17:45) - Testing Copilot
(19:54) - Comparing the tools and key takeaways
(20:52) - Outro
Subscribe for email updates on our website:
https://www.discussingstupid.com/
Watch us on YouTube:
https://www.youtube.com/@discussingstupid
Listen on Apple Podcasts, Spotify, or Soundcloud:
https://open.spotify.com/show/0c47grVFmXk1cco63QioHp?si=87dbb37a4ca441c0
https://soundcloud.com/discussing-stupid
Check Us Out on Socials:
https://www.linkedin.com/company/discussing-stupid
https://www.instagram.com/discussingstupid/
31 ตอน
ทุกตอน
×ขอต้อนรับสู่ Player FM!
Player FM กำลังหาเว็บ