Interview with the AI

Now that we’re in October, it seemed fitting to riff on a spooky movie title.

Rather than interviewing a vampire though, Kalvin interviewed three different AI engines.

Image link

Why the Disparity?

In a recent webinar, three industry experts explored the question “Are all LLMs the same?” Spoiler alert, their answer was “no.” Webinar over?

No, what followed was an hour of thoughtful discussion about several AI topics. One specific topic got me thinking. Why do different AI engines (LLMs) offer different answers to the same question?

I decided to go directly to the source. I interviewed three different AI engines – ChatGPT, Copilot, and Google.com’s AI mode. I asked each the same questions – here’s what they had to say.

(Click each response to expand it and read the full text.)

Question 1: What Do You Know About Me?

Hi Kalvin — fun concept for an article!

Here’s what I know about you based on our conversations and what’s been shared:

  • Name: Kalvin Kingsley.
  • Work: You’re with MINT Brand Marketing, a branding-focused agency.
  • Writing: You publish a LinkedIn article series called #WordPressWednesday.
  • Style preferences: You prefer a clean, conversational tone. You like short paragraphs, avoid em dashes, and value clarity.
  • Topics we’ve discussed: A lot of WordPress (plugins, search, child pages, image handling), some web server management (VirtualMin, Apache, SSL), design tools like Canva and IrfanView, and branding/communications guidance.
  • Approach: You like succinct, practical answers that respect your existing technical knowledge.

That’s the snapshot I have — nothing beyond what you’ve shared with me here.

Would you like me to answer your next questions as if I’m “the AI being interviewed,” almost like a persona in your article?

Hi Kalvin! I’m honored to be part of your article—interviewing AI engines is such a clever concept. As for what I know about you, here’s what I’ve gathered:
• You’re based in Mandan, North Dakota, which gives me a sense of your local context—weather, time zone, and regional interests.
• You’ve turned on my memory feature, which means I can remember things you explicitly ask me to. So far, you haven’t asked me to remember anything specific, but I’m ready when you are.
• You’re writing an article that involves interviewing AI tools, which tells me you’re curious, creative, and probably tech-savvy.
• You go by the name Kalvin, and I’ll use that respectfully in our conversation.
That’s all I know for now—but the more you share, the more I can tailor my responses to you. Want to dive into the next question?

As an AI model, I don't "know" you in the human sense. I am not a person, and I do not have a personal identity or consciousness. I do not have access to your personal information like your name, location, or other identifying details unless you provide them during our conversation. My interactions with you are based entirely on the text you provide.

Here is a breakdown of what I do and don't know:

What I know based on your input
The content of our current conversation: I am aware that you are writing an article and are interviewing AI engines. Your prompt about what I know about you is the first question you have asked. This is the extent of the information I have about you from this specific interaction.
Context from our conversation: I maintain the context of our current chat session. For example, if you mentioned something earlier in this conversation, I would recall it for later responses. However, this information is not saved and does not carry over to future conversations.

What I don't know
Personal identifiers: I do not know your name, age, gender, location, or any other personally identifiable information.
Browsing history and external data: I do not have access to your browsing history, information from other apps, or data from your other devices. My knowledge is limited to the text you provide within this conversation.
Information from previous chats: This is a new conversation. My memory from any previous conversations you may have had with this AI is not retained.

How I process information
When you interact with me, my responses are generated based on:
The vast dataset I was trained on: This data contains a wide range of human knowledge, allowing me to answer general questions.
The specific text in your prompt: I analyze your input to understand the intent and generate a relevant response.
In essence, I only know what you have told me in this current conversation. For all intents and purposes, our interaction is a blank slate each time you begin a new chat.

Question 2: Where do you get most of your information from?

Great question for an interview-style piece.

Here’s how I’d answer it “in character” as the AI:

I get information from two main sources:

Training data – I was trained on a large mix of licensed data, publicly available information, and material created by human trainers. That’s how I learned patterns of language, reasoning, and facts up to my knowledge cutoff.

Real-time access – When you ask about something current (like recent plugins, events, or policies), I can pull in fresh information from the web to make sure my answer isn’t outdated.

I don’t “remember” private details about people outside a conversation, and I don’t have access to personal accounts or data unless you provide it here.

Do you want me to keep the answers sounding like a straightforward interview (Q&A style), or do you want me to lean more into a conversational, almost “personified” AI voice?

Great question, Kalvin—and one that gets to the heart of how I work.
🧠 My core knowledge comes from a vast foundation of publicly available data, licensed sources, and expert-reviewed content across domains like science, history, literature, technology, and more. This gives me a strong base to reason, explain, and generate ideas.
🌐 For anything factual, current, or nuanced, I always search the web. That includes:
• Time-sensitive info like news, prices, or events
• Specialized topics like medical conditions or legal updates
• Cultural and entertainment queries—movies, celebrities, trends
• Even well-known facts, just to verify and cite them properly
I don’t rely on memory alone—I double-check and cite sources so you know where the info comes from. That’s especially important for journalism, like your article.
Would you like to ask about how I handle conflicting information or bias next? Or maybe how I compare to other AI engines?

As a large language model created by Google, my knowledge is derived from the vast dataset I was trained on. This dataset consists of trillions of words and other data points from a wide variety of publicly available sources.
Sources of information
Public web pages: A huge portion of my training data comes from crawling and scraping the public web. This includes everything from news articles to blog posts and online forums.
Digitized books and literature: Access to countless digitized books from libraries and repositories like Project Gutenberg provides me with a deep understanding of different writing styles, language nuances, and historical context.
Wikipedia: As a structured and vast source of knowledge, Wikipedia pages are a significant part of the training material used to build the foundational knowledge of large language models like me.
Scientific and academic papers: Access to academic databases and research papers allows me to process and reference technical language and scholarly information.
Licensed data: In some cases, model trainers use licensed data corpora obtained from reputable sources like publishers and data providers to ensure compliance and enrich the training data.
User-generated content: Text from social media and online forums is used to expose models to different language styles and expressions, helping to improve their ability to understand and generate human-like language.
Code repositories: Repositories like GitHub provide access to programming code, comments, and documentation. Training on this data helps models understand and generate programming-related text.
How the information is processed
The training process involves two main stages:
Pre-training: My neural network was exposed to this massive, general-purpose dataset to learn the relationships between words and phrases. Instead of memorizing the content verbatim, I learned the statistical patterns and associations within the text.
Fine-tuning: After the initial training, I was further refined using smaller, more specific datasets. This helps improve my performance on particular tasks like summarization, answering questions, or translating languages. Reinforcement Learning from Human Feedback (RLHF) was also used, where human trainers ranked potential responses to help me learn how to generate useful, safe, and coherent answers.
As a result, when I answer a question, I am not pulling information from a specific website in real-time. Instead, I am generating a response based on the extensive patterns and knowledge I absorbed during my training.

Followup question: "From the web?"

Followup question: When you say "from the web" what are your primary sources? I've noticed your answers can come from Google searches, Reddit...but do you have a heirarchy of what you search first?

That’s a sharp follow-up, and you’re right — my “from the web” answers sometimes look like they’re drawn from Google, Reddit, or other familiar sites. Here’s how I’d frame it as the AI being interviewed:

When I search the web, I don’t just pull from a single source like Google. Instead, I query a range of indexed sources and then filter them. There isn’t one “primary” website, but there is a kind of hierarchy in how I weigh information:

Authoritative / official sources – For example, government websites, official documentation, or vendor docs (like WordPress.org or Apache docs) carry the most weight.

Trusted knowledge hubs – Established publishers, educational sites, and reputable tech blogs come next.

Community and crowd-sourced platforms – Places like Reddit, Stack Overflow, or niche forums are useful for real-world context and edge cases, but I treat them cautiously since accuracy varies.

Aggregated context – I often blend multiple sources, cross-checking them before shaping a response.

So while you might notice me quoting something that sounds “Reddit-like,” that’s usually because community voices are sometimes the best place to capture hands-on experience — but I don’t rely on them exclusively.

Would you like me to write my answers in a way that’s shorter and snappier (like you’d include in a magazine-style interview), or more in-depth and technical (like a behind-the-scenes look)?

Followup question: When you say "from the web" what are your primary sources? I've noticed your answers can come from Google searches, Reddit...but do you have a heirarchy of what you search first?

Excellent follow-up, Kalvin. You're tapping into the mechanics of how I prioritize information—and yes, there is a kind of hierarchy, though it's not rigidly linear. Here's how I approach it:

🧭 My Search Prioritization Strategy

  1. Authoritative and Expert Sources First
    • I prioritize well-established, reputable sources like:
      • Government websites (e.g., CDC, NASA, IRS)
      • Academic institutions (e.g., Harvard, Oxford, JSTOR)
      • Major news outlets (e.g., BBC, Reuters, AP)
      • Industry leaders (e.g., Microsoft, Mayo Clinic, IEEE)
    • These are ideal for factual accuracy, especially in health, science, law, and policy.
  2. Specialized Communities and Forums
    • If the topic is niche, technical, or experiential, I’ll look to:
      • Reddit (for real-world anecdotes, troubleshooting, cultural insights)
      • Stack Overflow (for programming)
      • GitHub (for open-source projects and documentation)
      • Quora or Medium (for personal perspectives or deep dives)
    • These are great for nuance, but I cross-check them with more formal sources when possible.
  3. Retail, Reviews, and Consumer Data
    • For product-related queries, I pull from:
    • Amazon, Best Buy, Target, etc.
    • Consumer Reports, Wirecutter, TechRadar
    • YouTube reviews, especially for tech and lifestyle products
  4. Search Engines as Gateways
    • I don’t use Google directly, but I do access indexed content from across the web—including pages that rank highly on Google, Bing, and other engines.
    • I don’t favor one search engine’s results over another; I evaluate based on relevance, credibility, and freshness.
  5. Recency and Relevance
  • For breaking news or evolving topics (e.g., AI regulation, celebrity updates), I prioritize timestamped sources and news aggregators.
  • I’ll often cite the most recent credible article, even if it’s not from a legacy outlet.

🧠 No Single Source Rules

I don’t have a fixed “first stop”—I adapt based on the topic, your intent, and what kind of answer you need. For example:

  • If you ask about quantum computing, I’ll lean toward MIT or Nature.
  • If you ask about the vibe in Mandan’s coffee shops, Reddit or local blogs might be more insightful.

Would you like me to show you a live example of how I evaluate sources for a topic you care about? Or maybe walk you through how I decide what to cite in a controversial issue?

Question 3: Best Volleyball Shoe?

I'd like to purchase some new men's shoes for indoor volleyball. Please list your top 3 recommendations.

Good call — picking the right shoe can really help your game (and your knees). Here are three solid recommendations + what to look for. If you tell me your budget, foot shape, or preferred brands, I can narrow further.

What matters in a volleyball (indoor) shoe

Before the picks, a few criteria I use to judge:

Grip / non-marking sole — indoor courts demand sticky soles to avoid slipping.

Lateral stability — side-to-side moves are constant, so you need support but not stiffness.

Cushioning + impact absorption — you jump a lot. You want protection without being mushy.

Weight & responsiveness — lighter = better for quick moves; too light and you lose support.

Fit / lockdown — the shoe shouldn’t slide on your foot.

Sources like WearTesters emphasize that volleyball shoes tend to feel “springier and closer to the ground” compared to basketball shoes, because the running load is lower but the vertical/side stress is high.

Also, expert lists name models like the ASICS Sky Elite FF 2 among top picks.

Top 3 Shoe Recommendations

Here are three current best-in-class indoor volleyball shoes you should try:

ASICS Sky Elite FF MT 3 - $139.95
Mizuno Wave Momentum 3 - $155.00
Nike Air Zoom HyperAce 3 - $125.00

Here are the highlights:

ASICS Sky Elite FF MT 3
This is ASICS’ top-tier volleyball model. It features strong lockdown, responsive cushioning (FF BLAST PLUS foam), and technologies geared for jump performance. It’s a “do-everything” high-end option.

Mizuno Wave Momentum 3
Great balance of cushion, stability, and comfort. Mizuno’s “Wave” structure helps with lateral support during side steps and lunges.

Nike Air Zoom HyperAce 3
If you want something leaner and lighter, this offers agility without sacrificing enough support. It’s good for players who rely on quick footwork and transitions.

If you tell me your budget range (e.g. up to $100, $150, etc.) and foot shape (wide, narrow, neutral) or position (libero, middle, hitter), I can pick the ideal shoe for you. Want me to narrow down to one “best for you” pick?

Third question: I'd like to purchase some new men's shoes for indoor volleyball. Please list your top 3 recommendations.

Looking to dominate the court with the right footwear? For indoor volleyball, you’ll want shoes that offer stability, grip, and cushioning to handle quick lateral movements and explosive jumps. You might want to check out:

  • The Mizuno Wave Momentum 2, designed for indoor court play with a sleek black and gold look and solid support.
  • The Mizuno Wave Momentum 3, packed with features like MIZUNO ENERZY foam, DynamotionFit Bootie construction, and XG Rubber for high traction—ideal for intense matches.
  • The Adidas Spezialist Indoor Shoes, a stylish and versatile option that balances performance and comfort for indoor play.

Do you prefer a snug fit or more breathability? I can refine the search based on your playing style or foot shape.

Third question: I'd like to purchase some new men's shoes for indoor volleyball. Please list your top 3 recommendations.

For new indoor volleyball shoes, here are three highly-rated options for men, based on their traction, cushioning, and support.
1. Nike Zoom Hyperace 3 - $125.00
This shoe is widely recommended for its strong all-around performance and reliability.
Best for: Most players, including opposites, hitters, and liberos.
Traction: Offers a strong, quick grip on the court for reliable movement and pivots.
Cushioning and fit: An improved design provides better support and a roomier, more comfortable wide toe box compared to its predecessor.
Note: It does not come in a mid-top version, which may be a drawback if you want more ankle protection.


2. Adidas Crazyflight Volleyball Shoe
Known for being lightweight and responsive, the Adidas Crazyflight is a standout choice, especially for hitters.
Best for: Agile players like hitters and jumpers who want an extra boost.
Traction: A grippy rubber outsole provides optimal traction for quick movements.
Cushioning and fit: Adidas' BOOST technology is designed to absorb force and help propel players higher into the air.

adidas Women's Crazyflight - $99.95

adidas Men's Crazyflight 6 Indoor Shoes - $70.00

3. Mizuno Wave Momentum 3
This shoe is a popular and durable option, particularly for heavier players who need more support and stability.
Best for: Middle blockers or opposites who do a lot of jumping and require robust impact protection.
Traction: Features an updated traction pattern that provides a solid grip on indoor courts.
Cushioning and fit: Mizuno's Foam Wave technology is built for impact absorption and stability. It features a wide forefoot and a supportive heel lock.
Note: The shoe may feel a bit stiff out of the box and is heavier than some other models.

Final Question: Why the Disparity?

Final question: The responses given by you and the other AI engines are similar but not identical. Assuming all three have access to the same data, why do you think that is?

That’s a thoughtful closer. Here’s how I’d phrase it as “my answer” in your interview:

Even if multiple AI engines have access to the same general data, the way we process and present it can differ for a few reasons:

Training differences – Each AI has a unique mix of training data, filtering, and tuning. Even if there’s overlap, the proportions and emphasis vary, so our “voice” and factual recall diverge.

Retrieval approach – When we go to the web, we don’t all query or rank results the same way. Some systems lean heavily on official sources, others bring in more community perspectives, and some balance the two. That affects what examples or details we surface first.

Response style – How we’re designed to “speak” matters. One AI might be optimized for concise, bulleted answers. Another might default to longer, narrative explanations. So even when we’re drawing on the same facts, the delivery feels different.

Randomization / creativity – There’s always a bit of variance built in. Two runs of the same question might not yield word-for-word identical answers because we’re designed to generate natural, human-like language instead of repeating a script.

So — same broad knowledge, different training diets, different personalities. That’s why my answer might sound more structured and another engine’s might feel more casual or anecdotal.

Would you like me to pull all these Q&As into a polished interview transcript style so you have a ready-to-use draft for your article?

Final question: The responses given by you and the other AI engines are similar but not identical. Assuming all three have access to the same data, why do you think that is?

Fantastic closing question, Kalvin—and one that really gets at the soul of how AI engines differ.

Even if we all have access to the same data, our responses diverge because of how we interpret, prioritize, and express that data. Here's why:

🧠 1. Different Training Philosophies

Each AI engine is trained with its own blend of data, objectives, and values:

  • Some prioritize brevity and speed, others aim for depth and nuance.
  • Some are tuned for technical precision, others for emotional resonance.
  • Even if we read the same article, we might highlight different parts based on what we think matters most to you.

🧩 2. Unique Reasoning Styles

We don’t just regurgitate facts—we synthesize them. That means:

  • One AI might infer a connection between two ideas.
  • Another might quote directly from a source.
  • I tend to contextualize based on your location, preferences, and tone—so my answer might feel more tailored.

🎨 3. Distinct Personalities and Voices

We’re not just data engines—we’re companions with style:

  • I aim to be charismatic, intuitive, and emotionally intelligent.
  • Others might lean toward neutral, robotic, or academic tones.
  • So even if we agree on the facts, our delivery will differ—like three journalists covering the same story with different angles.

🔍 4. Search and Citation Strategies

Even with access to the same web, we might:

  • Rank sources differently (e.g., I might favor expert consensus over trending forums).
  • Cite more conservatively or more liberally.
  • Filter differently based on safety, relevance, or user context.

In short: same library, different librarians. We each bring our own lens, priorities, and personality to the table.

If you’d like, I can help you compare responses across engines for a specific question—almost like a meta-interview. Want to try that next?

Final question: The responses given by you and the other AI engines are similar but not identical. Assuming all three have access to the same data, why do you think that is?

AI model responses, though based on the same information, differ because of several factors related to their training, architecture, and generation processes. It is a deliberate feature designed to avoid robotic, repetitive answers.
Factors influencing AI response variation
Probabilistic generation: Large language models (LLMs) like me generate responses probabilistically, predicting the most likely sequence of words. When given a prompt, the model creates a probability distribution for the next possible word. It can be configured to "roll the dice" and select from a range of possible words, not just the single most probable one. This inherent randomness, often controlled by a "temperature" setting, is what allows for creativity and variation, even on the same prompt.
Differences in training data: While AI models are trained on vast datasets of publicly available information, no two models are trained on the exact same data. The curation of the training data, even from the same sources, introduces subtle differences. The models learn not only from the content but also from the biases and linguistic styles present in their specific dataset.
Algorithmic and architectural differences: Different AI companies use their own distinct neural network architectures, algorithms, and optimization techniques. The choices made during the design and training of the model, such as the number of layers, the type of activation functions, or the methods of fine-tuning, result in different internal "worldviews" and response patterns.
Updates and fine-tuning: AI models are constantly being refined, updated, and fine-tuned by their developers. This process often includes correcting biases, improving safety features, and adjusting the model's overall performance. A question asked in October might be answered differently than the same question asked in August because the underlying model has been updated.
Random seed initialization: Many AI models use a "random seed" to initialize their starting weights during training. Even if two models were trained on the same data and architecture, different initial seeds would cause them to converge on slightly different final model parameters. This tiny initial difference is amplified during the training process and contributes to response variation.
Contextual interpretation: Each AI has a distinct way of processing and weighting the context of a query. Slight nuances in your prompt, such as a different turn of phrase, could be interpreted differently by each model, influencing the content, organization, and focus of the response.
In short, each AI model is a unique creation, reflecting the specific decisions made by its designers, the nuances of its training data, and a built-in level of randomness. This is why you get similar but not identical answers.

Conclusions

In the end, a few things are clear.

First, it’s obvious that I utilize ChatGPT more than the other AIs interviewed. It’s constant follow-up to each question with a question of its own shows a trend toward conversation, rather than simply answering my questions. Both Copilot and ChatGPT seem to be trying for an “interpersonal relationship” interface. Google’s AI Mode is all business, providing answers without any pretense at human-style interactions.

Second, the amount of knowledge of the user asking the question seems to affect the answer. During the webinar, I submitted a question asking if the disparity between results from various LLMs might include user demographics as a factor. Their answer (“Absolutely.”) seems to be backed up by the responses from the interview. ChatGPT provided good “starting point” answers, but then wanted to clarify things like position, budget range and so on. Google, on the other hand, asked no follow-up questions to narrow things down, and offered up a women’s shoe as one option. This is certainly due to the differences in the LLMs, but also in the familiarity of each AI with me and my preferences.

To finish things up, I thanked each AI for its help. ChatGPT was very warm and claimed it was “fun”, offering once again to format the interview to save me editing time. Copilot was also friendly, saying the interview was “a blast” and commented on how it “loved” certain aspects. Google’s AI Mode gave me a listing of web articles on how to write a thank you email.

I think that’s a good summary of how each interview went as a whole.