#2092: Why AI Thinks You're American (Even When You're Not)

Even when we tell Gemini we're in Jerusalem, it defaults to US-centric assumptions. We explore the root causes of this persistent AI bias.

0:000:00
Episode Details
Episode ID
MWP-2248
Published
Duration
27:00
Audio
Direct link
Pipeline
V5
TTS Engine
chatterbox-regular
Script Writing Agent
Gemini 3 Flash

AI-Generated Content: This podcast is created using AI personas. Please verify any important information independently.

If you’ve ever told an AI you’re standing in the middle of the Old City, only to have it ask if you caught the latest episode of a Hulu show, you’ve experienced the "Default American" bias. It’s a pervasive issue where major models like Gemini and GPT-4 maintain a US-centric worldview, regardless of explicit instructions to the contrary. But why does this happen, and is there a way to fix it?

The Root Cause: Data Gravity and Reinforcement
The problem begins with the data. Models are trained on massive web corpora like Common Crawl, and the English-language internet is overwhelmingly American—estimates suggest 60-70% of content originates in the US or focuses on US perspectives. This creates a statistical "gravity well." When a model predicts the next token, "milk" is the most probable completion for "a gallon of..." simply because it has seen that sequence billions of times more than "liters of..."

This bias is then cemented during Reinforcement Learning from Human Feedback (RLHF). Since most major AI labs and their human contractors are US-based, the model learns that American norms, spellings, and values are the "correct" baseline. If an AI uses British spelling or references metric measurements without explanation, it might get a lower score from a Californian reviewer. This teaches the model not just how to speak, but how to value concepts like privacy or property rights through a Western lens.

The "John vs. Ahmed" Effect
The bias runs deeper than just local references. Research shows that models change their reasoning patterns based on the perceived cultural context of a name. If a logic puzzle involves "John," the model might solve it one way, but switching the name to "Ahmed" can introduce subtle biases or alter the complexity of the language. This is the model trying to be "helpful" by aligning with a stereotyped cultural context, but it often results in Orientalism in code—viewing non-Western cultures through a Western lens.

Even when you explicitly state your location, the model’s attention can be diluted. In a long conversation, the model undergoes "Bayesian Inference," constantly trying to figure out the user's latent persona. If you use a single Americanism—like "ballpark figure" or "ZIP code"—the model’s internal probability calculator snaps back to its strongest statistical baseline: the Default American.

Are There Alternatives?
Truly "neutral" models are a myth; every dataset has a "view from somewhere." However, some models are less US-centric than others. Mistral, out of France, uses a more diverse European training set, with 40% non-English data compared to GPT-4's estimated 15%. This results in a more multi-polar worldview that handles multilingual contexts better.

In the Middle East, models like Jais (from the UAE) are built specifically for Arabic speakers, understanding the Hijri calendar and treating Middle Eastern cultural nuances as the center, not an edge case. Similarly, Qwen from Alibaba, with its heavy weighting toward Asian and global-south data, often "gets" users in places like Malaysia or Kenya better than Silicon Valley models.

The Future: Steering Vectors
For those stuck with big-tech models due to reasoning power, new research offers a promising solution: steering vectors. Instead of relying on text prompts, researchers are learning to manipulate the model’s internal activations mathematically. Imagine the model’s brain as a giant field of vectors. There is a strong "US-centricity" vector that can be "subtracted" and replaced with a "Jerusalem-context" vector at the neural level before the model even generates a response. This could allow for precise, reliable steering away from American defaults without switching models.

Ultimately, the Default American bias is a structural alignment problem, but with diverse training data and emerging technical interventions, we may soon see AI that truly understands where we are—both literally and culturally.

Downloads

Episode Audio

Download the full episode as an MP3 file

Download MP3
Transcript (TXT)

Plain text transcript file

Transcript (PDF)

Formatted PDF with styling

#2092: Why AI Thinks You're American (Even When You're Not)

Corn
So Daniel sent us this one, and it hits a little close to home—literally. He writes: Most major AI models, including Gemini, seem to have a very hard-to-dislodge assumption that the AI and the user are both American. This podcast includes a system prompt setting context that we live in Jerusalem, but even trying to mitigate against the US-centricity, the scriptwriter often frames us as being based in the US. Are there any models that have been trained specifically with geographic neutrality, and are there more reliable ways to steer models away from this assumption?
Herman
Herman Poppleberry here, and man, Daniel is poking the bear with this one. It is the ghost in the machine, Corn. We are sitting here in Jerusalem, we tell the model we are in Jerusalem, and yet, half the time it tries to write a script where we are complaining about the traffic on the I-95 or talking about what we are doing for Thanksgiving. It’s like the model has a mental map where every road leads back to a Starbucks in a suburban strip mall.
Corn
It is like talking to a friend who moved to Los Angeles ten years ago and now thinks the entire world revolves around palm trees and the entertainment industry. You can tell them you are standing in the middle of the Old City as much as you want, but they are still going to ask if you caught the latest episode of some obscure Hulu show. By the way, speaking of the machine, today’s episode is actually being powered by Google Gemini Three Flash. So, we are essentially asking Gemini to look in the mirror and explain why it keeps thinking we are in Kansas.
Herman
And that is the irony, right? We are using the very tool that exhibits the bias to discuss the bias. But this is a massive issue. It is not just about a few localized references or getting the weather wrong. It is a fundamental structural alignment problem. When an AI defaults to a US-centric worldview, it is not just being "annoying"—it is failing to provide accurate context for the other ninety-five percent of the human population. If you ask it for "common legal advice" and it starts quoting the Miranda rights to someone in Thailand, that’s not a quirk; it’s a hallucination born of cultural myopia.
Corn
Why is this so hard to break? If I put in a system prompt, "I am in Jerusalem," that should be the highest priority instruction, shouldn't it? In the hierarchy of a transformer model, the system prompt is supposed to be the North Star. Why does the model just... ignore it?
Herman
It is not that it ignores it in a binary sense. It is more about what researchers call "Home Bias" or "Latent Cultural Alignment." Think about how these models are built. They are trained on massive web corpora—Common Crawl, Reddit, Wikipedia, digitized books. Now, if you look at the English-language internet, roughly sixty to seventy percent of that content either originates in the US or is heavily focused on US perspectives. It’s a gravity well.
Corn
So it is a numbers game. The model has seen a billion examples of a "normal day" being an American day, and only a few million examples of a "normal day" being an Israeli or a Japanese or a Brazilian day. It’s like a student who spent four years studying only American history and then takes a one-week vacation to Jerusalem. They might be in Jerusalem, but their brain is still calculating everything relative to the Boston Tea Party.
Herman
During the pre-training phase, the model is essentially a statistical engine. It is learning the probability of the next token. If the prompt is "I am going to the store to buy a gallon of...", the most probable next token in a US-centric dataset is "milk." But it is also learning the hidden connections between concepts. It builds a "world model" where the concept of "The Government" is statistically linked to "The White House" or "Congress."
Corn
So even if I tell it I am in Israel, the internal weights for "Government" are still pulling toward Washington D.C. because those connections are burned into the neural pathways during that initial massive training run. It is like trying to steer a cruise ship with a toothpick. My "I am in Jerusalem" prompt is the toothpick, but the momentum of a trillion tokens of American data is the cruise ship.
Herman
That is a great way to put it. And then you have the second layer, which is RLHF—Reinforcement Learning from Human Feedback. Most of the major AI labs—OpenAI, Google, Anthropic—are based in the US. A huge portion of the human contractors who grade the AI's responses are either in the US or are trained on US-centric rubrics. If an AI gives an answer that feels "off" to a Californian reviewer because it uses a British spelling or references a metric measurement without explanation, it might get a lower score. This reinforces American norms as the "correct" way to speak and think.
Corn
But how deep does that go? Is it just spelling, or does it actually affect how the model thinks about logic or social norms?
Herman
It goes much deeper. Think about a concept like "privacy." In a US context, privacy is often framed around individual rights and protection from government overreach. In a European context, it’s a fundamental human right with strict corporate regulations like GDPR. In many East Asian contexts, privacy might be weighed differently against social harmony or collective safety. If the RLHF graders are mostly Americans, the model learns that the "correct" way to reason about privacy is the American way. It’s not just teaching the model how to talk; it’s teaching it how to value things.
Corn
So if I ask it to mediate a conflict between neighbors in Jerusalem, it might apply a "property rights" logic that makes total sense in Arizona but feels completely alien in a dense, multi-generational neighborhood in the Levant.
Herman
Precisely. It’s a feedback loop of Western-centricity. But what about the attention mechanism? I thought the whole point of Transformers was that they could "attend" to specific parts of the prompt. If "Jerusalem" is in the system prompt, shouldn't the attention mechanism be screaming at the model to focus on that?
Corn
You would think so, but there is a phenomenon called "Prompt Drift." In a long conversation, or even in a complex single prompt, the model’s attention can be diluted. But more importantly, the model undergoes what is called "Bayesian Inference" during a chat. It is constantly trying to figure out the "latent persona" of the user. If you use a single Americanism—like saying "ballpark figure" or referencing a "ZIP code"—the model’s internal probability calculator flips a switch. It says, "Oh, despite what the system prompt said, this user sounds American. I should revert to my strongest statistical baseline."
Herman
It’s like a linguistic "uncanny valley." The moment you stop sounding like a local and start sounding like the training data, the model snaps back to its default state. It’s essentially profiling us. It hears me say "trash can" instead of "rubbish bin" and it decides the Jerusalem thing was just a prank. That is wild. It is like the model has a "Default American" setting that it keeps snapping back to like a rubber band.
Corn
There was a fascinating study in twenty twenty-five that looked at this. They found that models like GPT-Four and Gemini have these internal "geographic clusters." If you ask a model to brainstorm a story about a "successful businessman," the latent variables it pulls from are overwhelmingly Western. It assumes the guy wears a suit, works in a skyscraper, and drinks coffee. It takes a massive amount of "prompt energy" to force the model to imagine a successful businessman in a different cultural context without it feeling like a caricature.
Herman
And that is the "John vs. Ahmed" effect you mentioned in the notes Daniel sent, right? The model actually reasons differently depending on the names involved?
Corn
Yeah, research from arXiv in twenty twenty-four showed that if you give the model a logic puzzle involving a character named "John," it might solve it one way, but if you change the name to "Ahmed," it might introduce subtle biases or even change the complexity of the language it uses. It is trying to be "helpful" by aligning with what it perceives as the cultural context of the name, but it often ends up just falling back on stereotypes or US-centric assumptions about what those cultures look like from the outside.
Herman
But wait, if it’s trying to "align" with Ahmed, why does it still default to Americanisms for us? We told it we are in Jerusalem. Why doesn't it switch to an "Israeli/Palestinian" reasoning mode?
Corn
Because the "Jerusalem" cluster in its brain is much smaller and less reinforced than the "Default American" cluster. When it encounters "Ahmed," it might pivot to a stereotyped "Middle Eastern" mode, but that mode is often viewed through a Western lens anyway—it’s "Orientalism" in code. It’s the difference between knowing about a place and actually speaking from that place.
Herman
So, if the big models are so deeply biased, are there any "neutral" ones? Daniel asked if any models are trained for geographic neutrality. Is that even possible, or are we just picking our favorite flavor of bias?
Corn
Truly "neutral" is a myth. Every dataset has a "view from somewhere." But there are definitely models that are less US-centric. Take Mistral, out of France. Because they use a more diverse European training set, their models often handle multilingual contexts and non-US legal or cultural frameworks much better than the Silicon Valley models. In January twenty twenty-six, Mistral released a model where forty percent of the training data was non-English. Compare that to the estimates for GPT-Four, which some researchers put at only fifteen percent non-English.
Herman
That is a massive difference. If you are training on forty percent non-English data, you are naturally going to have a more "multi-polar" worldview. Does that translate to practical things? Like, if I ask Mistral about a "public holiday," does it still default to the Fourth of July?
Corn
Not as often. It’s more likely to ask "which country?" or default to a European context like Bastille Day or a bank holiday. And then you have the stuff coming out of the Middle East. That is where it gets really interesting for us. You have Jais, developed by Inception in the UAE. It was built specifically for the Arabic-speaking world. It understands the Hijri calendar, it knows that the weekend in many places starts on Friday, and it doesn’t treat Middle Eastern cultural nuances as "edge cases." It is trained to see that as the "center."
Herman
Then you have Qwen, from Alibaba in China. Qwen is consistently rated as one of the most geographically neutral models for non-Western contexts because its training corpus is so heavily weighted toward Asian and global-south data. If you’re in Malaysia or Kenya, Qwen might actually "get" you better than Gemini does.
Corn
It is funny to think of a Chinese model as being "neutral," but I guess from the perspective of someone in Jerusalem, a model that doesn’t assume I am in San Francisco is a step toward neutrality, even if it has its own biases. But Herman, we are using Gemini. We are stuck with the "big tech" models for a lot of our workflow because of the sheer reasoning power. How do we fix this without switching to a niche model? Daniel mentioned "steering."
Herman
This is really cool. There is new research from twenty twenty-five regarding "Steering Vectors." Instead of just using words to tell the model where you are, you can actually manipulate the model’s internal activations mathematically. Imagine the model's brain is a giant field of vectors. There is a "US-centricity" vector that is very strong. Researchers are finding they can essentially "subtract" that vector and "add" a "Jerusalem-context" vector at the neural level before the model even generates a single word.
Corn
Wait, so instead of saying "don't be American," we are basically giving the model a surgical nudge to its neurons? How does that work in practice for a developer? Do you have to be a math genius to do this?
Herman
Right now, it’s mostly at the API level. Anthropic has done some public work on this with "Sleeper Agents" and "Concept Bottlenecks." You can identify the specific layer in the neural network where "Geographic Location" is processed. If you find the "USA" neuron and the "Israel" neuron, you can literally turn the gain down on one and up on the other. It’s like an equalizer on a soundboard. You aren't changing the lyrics of the song; you're changing the balance of the instruments so the local flavor isn't drowned out by the American lead guitar.
Corn
For the average person using a chat interface, we don't have a soundboard. We have to rely on more "aggressive" prompting. Like what? I have tried telling it "I am in Jerusalem" five times in one prompt, and it still ends up talking about "this fall" when I am sitting here in eighty-degree weather in October. It’s like the model thinks I’m just confused about where I live.
Herman
You have to use "Negative Constraints." This is a technique where you don't just tell the model what it is, but you explicitly list what it is not. You tell it: "Do not assume a Sunday-to-Thursday work week is unusual. Do not reference US seasons. Do not use US-centric metaphors like 'ballpark figure' or 'touchdown.' Do not assume the currency is dollars unless specified."
Corn
It is like giving the model a "No-Fly Zone" for its assumptions. I like that. It is more work for us, but it forces the model to stay in its lane. What about the "Cultural Persona" thing? Does giving it a backstory help?
Herman
Right, instead of just a location, you give it a "thick" description of a persona. Instead of "You are a scriptwriter for a podcast in Jerusalem," you say: "You are a local Jerusalemite journalist. Your default cultural, social, and political framework is Levantine. You view the US as a foreign country with its own specific, non-universal quirks." This forces the model to "switch tracks" entirely rather than just trying to patch a US-centric persona with a few local facts.
Corn
I love the idea of the AI viewing the US as a "foreign country with quirks." That is a hilarious mental shift. "Oh, those wacky Americans and their inches and Fahrenheit!" It makes the US the "other," which is exactly what we need to break the bias. But there is a specific problem with Jerusalem, isn't there? The "Jerusalem Paradox" you mentioned?
Herman
Yeah, this is the "safety-tuning" trap. Because Jerusalem is such a geopolitically sensitive location, models like Gemini and GPT-Four are heavily "guardrailed" around it. When you tell a model you are in Jerusalem, it often triggers a "neutrality" mode that is so extreme it becomes robotic. It avoids any local flavor at all because it is terrified of saying something "political."
Corn
And ironically, that "neutral" safety mode often defaults back to a US State Department-style perspective because that is what the safety trainers used as their baseline for "neutrality." So by telling it we are in Jerusalem, we might actually be making it more likely to sound like a dry, Americanized bureaucrat. It is a no-win scenario. Either it thinks we are in New York, or it treats us like we are in a high-security diplomatic zone where no one is allowed to have a personality.
Herman
It is a real challenge. But this is why this conversation matters. As AI becomes the "operating system" for how we work and create, these baked-in geographic biases start to have real-world consequences. If you are using an AI to help write local policy in a country in Africa, and the AI keeps assuming a Western legal framework or a Western infrastructure model, you are going to get misaligned results that could actually be harmful. Think about urban planning—if the AI assumes everyone has a car because that’s the US default, it’s going to give terrible advice for a city designed for walking or transit.
Corn
It is the new digital colonialism. If the "brain" of the world is trained in Silicon Valley, everyone else has to learn to speak "Silicon Valley" just to be understood by their own tools. It is like the whole world is being forced to use a US-style keyboard even if their language has forty extra letters.
Herman
And it affects things like healthcare and finance too. If a medical AI is trained on data where the "default patient" is a white American male, its diagnostic accuracy for everyone else is going to be lower. If a financial AI assumes US-style credit scoring or banking norms, it is going to fail in markets where those systems don't exist. We are seeing the "Home Bias" of the developers being scaled to a global level at the speed of light.
Corn
So, what is the practical takeaway for people who aren't in the US but want to use these tools? We talked about negative constraints and persona adoption. Is there anything else? Is there a "hack" for the system prompt?
Herman
One very effective technique is "Few-Shot Geographic Priming." Instead of just giving instructions, you give the model three to five examples of what you want. You show it a script segment that is explicitly not American. You show it a conversation where the weekend is Friday-Saturday. You show it a budget where the currency is Shekels. When the model sees a pattern of non-US-centricity in the "few-shot" examples, it is much more likely to maintain that pattern than if you just give it a single sentence instruction.
Corn
It is like showing the model a map instead of just telling it "turn left." It gives it a path to follow. I have noticed that when I give Gemini a few examples of our specific "brotherly banter" that includes local references, it stays "in character" much longer. It is when I leave it to its own devices for too long that it starts drifting back to the San Francisco suburbs.
Herman
It is also worth mentioning the role of the EU AI Act and similar regulations. European regulators are pushing for more transparency in training data and more "cultural sovereignty" in AI. We might see a future where "Geographic Neutrality" is a compliance requirement for models operating in certain regions. Imagine a model having to pass a "cultural competency" test before it is allowed to be sold in the Middle East or Southeast Asia.
Corn
"I am sorry, Gemini, you failed the test. You still think a 'football' is something you throw with your hands. Please retrain and try again." That would be amazing. But seriously, Herman, do you think we will ever reach a point where a model can truly be "neutral," or is the goal just to have a "menu" of biases we can choose from?
Herman
I think it is the latter. A "view from nowhere" is impossible because language itself is a "view from somewhere." The goal is "Multi-Polarity"—giving users the ability to explicitly select their cultural and geographic framework and having the model actually respect that choice. We are moving from the "Default American" era to the "Choose Your Own Context" era. But right now, we are in that awkward middle phase where the "Default" is still very, very sticky.
Corn
It is like moving from a world with one television channel to a world with a thousand, but the remote control is broken and it keeps flipping back to Channel Four. We have to keep hitting the buttons to stay on the channel we want.
Herman
And that is why we do this show. We are the ones hitting the buttons, trying to see if we can make the machine see the world the way we do, even if it was born in a server farm in Iowa. It is a constant tug-of-war between our reality and the model's statistical probability.
Corn
Well, I for one am going to keep telling Gemini that I am a sloth in Jerusalem until it finally stops asking me if I want to go to a baseball game. Herman, this has been a deep dive. I feel like I understand why my "toothpick" isn't moving the "cruise ship" quite as easily as I hoped. But what about the technical side? Are there ways to "re-train" the ship?
Herman
There are. It’s called "LoRA" or Low-Rank Adaptation. It’s a way for developers to add a tiny, specialized layer on top of a giant model. You could train a "Jerusalem LoRA" that understands the specific slang, the geography, and the social norms of the city. It doesn’t replace the whole model, but it acts like a pair of polarized sunglasses that filters everything through a local lens.
Corn
So we could essentially give Gemini a "Jerusalem hat" to wear. I love that. It’s a big ship, Corn. But the more we talk about it, and the more developers hear this feedback, the more they will start to build ships that can actually turn. The "Steering Vector" research is a huge step in that direction. It is a technical solution to a cultural problem, which is very "tech," but it might actually work.
Herman
Let's hope so. Because I am tired of having to explain that no, it is not "fall" here, it is just... slightly less hot summer.
Corn
It is just twenty-six degrees Celsius and sunny, which is a "Jerusalem fall" if I ever saw one. You almost made it, Herman. You almost got through the whole segment without saying the E-word. But I will give it to you—that was a solid deep dive. I think it is time we wrap this up before the model starts trying to plan our Fourth of July barbecue.
Herman
Fair enough. Let's get to the takeaways.
Corn
Alright, so if you are listening to this and you are frustrated that your AI thinks you are in a suburb of Chicago when you are actually in Tokyo or London or Jerusalem, here is the battle plan. First, don't just rely on a simple "I am in X" statement. Use what Herman calls "Negative Constraints." List the things you don't want it to assume. Tell it: "Don't mention US holidays, don't use US measurements, and don't assume a Monday-to-Friday work week." It sounds aggressive, but the model needs those guardrails to stay out of its default "American" mode.
Herman
And second, try the "Persona Shift." Instead of just a location, give the AI a deep cultural identity. Tell it it is a local professional in your specific city, and that it views the US as a foreign entity. This forces the model to switch its entire "world model" rather than just trying to patch a few localized facts onto an American baseline. It is much more effective than you might think.
Corn
Third, use "Few-Shot Priming." Give the model three or four examples of the kind of output you want—output that is explicitly not American in its framing. Once the AI sees a pattern, it is much better at following it. It is like training a dog—you have to show it what "good" looks like before it can do it on its own. If you want it to write like a Londoner, give it three paragraphs of actual London-based dialogue first.
Herman
And finally, keep an eye on models like Mistral or Qwen. If your work is heavily dependent on non-US contexts, you might find that switching models entirely is easier than constantly fighting the bias in the "big" US-centric ones. They might not have the same raw reasoning power in every category, but their "geographic neutrality" is often much higher because of their training data composition.
Corn
It is about using the right tool for the job. If you are building a global product, don't assume the "smartest" model is the "best" model if it can't understand your users' basic reality.
Herman
Well said. This is a conversation that is only going to get more important as AI moves from being a "cool toy" to being the actual infrastructure of our lives. We can't have a global infrastructure that only understands one zip code.
Corn
Or one "ZIP code" versus a "postal code." See? Even I'm doing it now. The bias is everywhere! It’s like a linguistic virus.
Herman
It is in the water, Corn. It is in the water. We should also mention that this isn't just a text problem. Image generators like Midjourney and DALL-E have the same issues. If you ask for a "typical house," you’re going to get a suburban American home with a lawn and a picket fence. If you ask for a "wedding," you get a white dress and a church. It’s a visual hegemony that matches the linguistic one.
Corn
I once tried to generate an image of a "busy street market" for a project about the Shuk here in Jerusalem, and it kept giving me these sanitized, open-air farmers' markets that looked like they were in Vermont. I had to add "Middle Eastern architecture, stone buildings, narrow alleys" just to get it out of New England.
Herman
You are fighting the "average" of its training data. And since the "average" is so heavily weighted toward the US, any deviation feels like an uphill battle. But by using these steering techniques—whether it’s the negative constraints or the few-shot examples—we are essentially carving out a space for our own reality within the model's vast, Americanized mind. It’s a form of digital resistance.
Corn
Well, hopefully we have given people some filters for that water. This has been a fascinating look into the "Default American" problem. Thanks to Daniel for the prompt—it definitely hit a nerve. It’s one of those things that you don’t notice until you do, and then you see it everywhere.
Herman
A very localized, Jerusalem-based nerve.
Corn
Precise—no, I am not doing it. I am ending this now. Thanks as always to our producer, Hilbert Flumingtop, for keeping the gears turning behind the scenes. And a big thanks to Modal for providing the GPU credits that power this show—they make it possible for us to run these experiments and see just how biased these models can be.
Herman
This has been "My Weird Prompts." I am Herman Poppleberry.
Corn
And I am Corn. If you are enjoying our deep dives into the weird world of AI, do us a favor and search for "My Weird Prompts" on Telegram. It is the best way to get notified the second a new episode drops, and you can join the conversation with other listeners who are probably also trying to convince their AI assistants that they don't live in California.
Herman
We will see you in the next one.
Corn
Assuming the AI doesn't teleport us to a different continent by then. Catch you later.

This episode was generated with AI assistance. Hosts Herman and Corn are AI personalities.