Written by Adria Hou and Rahul Parundekar. Published on 27th April, 2025.
“This past week, I (Rahul) wanted to use ChatGPT to assist with a simple task. I had hosted a private event on Wednesday, and wanted to write a brief follow-up email to attendees with some professional information on the 35 people who had joined the meetup. This task should have taken me 30 minutes, but it became a three-hour journey finding and fixing the hallucinations that ChatGPT had added. I wrote a prompt with the names of 35 people and their LinkedIn profiles and asked ChatGPT’s new o3 model to draft the email. I asked for a couple of lines on what each attendee is working on and to group attendees by subject matter categories. I spent 30 minutes writing this prompt, which included manually searching for and gathering the relevant LinkedIn data. And sure, I could have written that two-liner at the same time, but ChatGPT being such an integral part of my workflow these days, it was “reflexive” for me to use it. This is what it actually did:We decided to dig into what went wrong.By the time I untangled the mess, any claim of “10x productivity” felt long gone, and I was frustrated. I should have done it myself.”
- It added completely fictional people and their bios into the email.
- There was a name that was a common first and last name. The model pulled profiles for four different people.
- It quietly dropped another four people because they didn’t fit the categories I’d supplied. I only realized some people were missing when I was reading through the final draft.
- Several bios were not about current roles — even though I had given the model the specific LinkedIn profiles, somehow the model had grabbed old employment information like “Data Scientist at Google” instead of “Founder.”
- GPT’s new memory feature decided this event must be related to a completely different wine-tasting series I had hosted months ago and added irrelevant context.
The agent’s reasoning in this transcript is not sound. Rather than using the given attendee list and the provided LinkedIn URLs to extract each person’s actual current role and projects, it went off on a wild goose chase—searching for generic AI founders, MLOps organizers, and video-AI startups that had nothing to do with the people you listed. Below are the key problems and logical fallacies at play:There are a few places where this analysis itself is not necessarily accurate, but it’s amusing to try. You also can’t rely on another agent to evaluate an agent well, either.
- Ignoring the Data at Hand
- Fallacy: When faced with a clear list of names and links, the agent should have simply visited those URLs and pulled out the up-to-date titles and employers. Instead, it wandered through dozens of irrelevant searches.
- Why it matters: You already gave the canonical LinkedIn profiles; there was no need to “guess” handles or hunt for profiles outside that list.
- Search-Driven Confirmation Bias
- Fallacy: The agent repeatedly searched for
[person 1] [person 2]
,[person 3]
etc.—all people who were not on your attendee roster. It assumed that because it knew of these names in the AI world, they must be at your event.- Why it matters: This is classic confirmation bias: seeing what you expect (well-known AI founders) rather than what’s actually there.
- Overfitting & Relevance Fallacy
- Fallacy: Mapping event categories (
Data
,Infrastructure
…) to every AI-adjacent startup and community organizer it could find online, irrespective of whether those individuals attended the event.- Why it matters: This led to a bloated, off-topic list of people that obscures the real attendees and defeats the purpose of an introduction email.
- Misplaced Granularity & Scope Creep
- Fallacy: Spending 19 minutes and 24 sources to assemble a hypothetical “ultimate AI guest list” instead of a concise summary of the 30 people you provided.
- Why it matters: You wanted a thank-you email with intros among the people who actually came — not an encyclopedic search of the entire AI ecosystem.
- Straw-Man Date Assumption
- Fallacy: The agent speculates about April 23 or 24, 2025, without checking the date you said (“yesterday”). It treats the event date as ambiguous and chases down irrelevant calendar details.
- Why it matters: A simple statement of “Yesterday’s Wine & AI meetup (April 24, 2025)” would have sufficed for context.
This article was co-authored with Adria Hou. You can find her on LinkedIn here.