Skip to content
Search

Latest Stories

Top Stories

Government must protect us from geolocational disinformation

City with GPS markers
Mongkol Chuewong/Getty Images

Crampton is an adjunct professor of geography at George Washington University and a member of Scholars.org.

As artificial intelligence is becoming more powerful and is more embedded in society, global governments are beginning to regulate these types of technology, and to balance benefits with potential harms. Yet while significant attention has been paid to reducing risk in the realms of health, finance and privacy, policymakers have left one element largely unaddressed: geolocation data.


This data — which provides information on the physical location of a person or device like a smartphone — is powerful, sensitive and highly valuable. AI procedures that are already being adopted to acquire, process and automate spatial and locational data are a particular concern that call for swift action. But policymakers can simultaneously look to the future and work to ensure that we develop independent, trustworthy AI governance for geolocation by drawing on the hard-won knowledge of the spatial digital revolution of the past two decades. To realize the best outcomes on privacy, combatting disinformation and deanonymization threats, policymakers must partner with geospatial domain experts rather than legislate around them.

The current regulatory landscape

In 2023, privacy legislation protecting “sensitive information” was passed in California, Colorado, Connecticut, Delaware, Florida, Indiana, Iowa, Montana, Oregon, Tennessee, Texas and Utah. A significant number of these laws include a provision covering “precise geolocation.” Data that qualify as indicating precise geolocation are limited to a radius of 1,750 feet around their subject, the equivalent of approximately one-third of a mile — significant territory in a densely populated urban area, as this interactive map of health care facilities in Washington, D,C., shows.

In Europe, the EU AI Act has prohibited the real-time collection of biometric data, such as occurs in facial recognition technology. Concurrently, U.S. legislators have become increasingly concerned about the risks of artificial intelligence, and the White House issued an executive order calling for new standards to prevent AI bias, threats or misuse. In 2018 in Carpenter v. United States, the Supreme Court held that law enforcement agencies require a warrant to obtain location data from cell-phone towers; however, this data is imprecise, and law enforcement actors switched to using richer data sources not covered by the ruling (like app-based location data sourced directly from Google).

While these are welcome developments in the ongoing need to secure privacy, geolocation has certain unique risks that this legislation and the policymakers concerned with it have yet to address.

Risks

There are three main categories of risk for geolocation data governance: disinformation, surveillance and antitrust/concentration of market. Disinformation (e.g. fake maps and data, propaganda), bias, and discrimination raise issues of trustworthiness, privacy, and ethics. The concern for AI is not just low-quality knowledge, but low-quality learning and low-quality meaning. For example, predictive policing — where data is analyzed to predict and intervene on potential future crime —may be based on poor, false or biased data that can lead to real-world discriminatory consequences.

  • Fake data can infiltrate maps (spatial databases), intentionally or not. Fake data could be included in driving apps or autonomous vehicles to chaotic effect; AI could be applied to geospatial data in a manner that misclassifies satellite imagery.
  • Disinformation may include falsely showing a person to be at a location when they were not — known as “location spoofing” — for blackmail or to cause reputational damage. “ Deepfake geography ” involves faking that a person is not at a place they should be: Imagine truckers’ data hacked to falsely show them as having deviated from their routes.
  • Inadvertent misinformation proliferated by lack of relevant geospatial analytic expertise can lead to detrimental outcomes. Inaccuracy and uncertainty can arise from analyzing a phenomenon at the wrong spatial scale (known as the “ the Openshaw Effect ”), or not accounting for how boundaries influence the scale of the analysis of aggregated data (referred to as “modifiable areal unit problem”).

Surveillance and locational tracking, which can include wide-scale biometric identification in real time or upon review of previously gathered data, poses many threats to privacy. Inference of personally identifiable information based on geospatial data obtained through surveillance is all too easy, and can include privacy infringements like re-constituting encrypted data (deanonymization) and uncovering the identity of a person or organization that has been obscured (re-identification.)

One well-known 2013 study found that knowing just four location points was enough to re-identify 95 percent of individuals, and that even when geospatial data is less precise, it can still reliable re-identify individuals — it takes a high factor of imprecision before location data loses its power to pinpoint. A new study of metro card travel data confirmed the findings that three random location points from within a period between one minute and one hour are sufficient to identify 67 percent to 90 percent of users. And facial recognition technology, which has many clear privacy risks, is now widely employed by law enforcement.

The limits of antitrust regulation and market concentration among tech companies point to increased opportunities for large data breaches or unethical use. The market is dominated by deep-pocketed AI tech companies including OpenAI, Google and Microsoft; these companies own and control the high-tech market, especially “high compute” fields like machine learning and AI training, effectively locking out competitors.

Recommendation to begin risk mitigation

The Offices of Science and Technology Policy at the White House and Congress can hold hearings with geospatial industry and academic experts to identify current and emerging threats to privacy from geolocation data and geolocation services and analytics. The quality and efficacy of legislation will depend on collaboration with and transparency from the experts who are designing and deploying these emerging technologies.

Read More

People on their phones. ​

In order to achieve scale, many civic efforts must also reach Americans as media consumers, where Americans currently spend much more time.

Getty Images, Xavier Lorenzo

Reaching Americans As Media Consumers – Not Only As Participants – To Improve the Political Environment

Current efforts to improve how Americans think and feel about those across the political spectrum overwhelmingly rely on participation. Participation usually involves interpersonal interaction, mostly to have dialogues or to collectively work on a project together.

These can be valuable, but in order to achieve scale, many efforts must also reach Americans as media consumers, where Americans currently spend much more time.

Keep ReadingShow less
Scams Targeting Immigrants Take Advantage of Fears of Immigration Status and Deportation

Scam incoming call alert screen on mobile phone.

Getty Images/Stock Photo

Scams Targeting Immigrants Take Advantage of Fears of Immigration Status and Deportation

WASHINGTON–When my phone rang and I saw the familiar DC area code, I picked up, and a man with a slight Indian accent said: “Ma’am, this is the Indian Embassy.”

Expecting a response from the Indian Embassy for an article I was working on, I said, “Is this in regards to my media inquiry?” He said no. He was calling about a problem with my Indian passport. I asked who he called, and when he said a name I didn’t recognize, I informed him he had the wrong person and hung up, figuring it was a scam.

Keep ReadingShow less
The American Schism in 2025: The New Cultural Revolution

A street vendor selling public domain Donald Trump paraphernalia and souvenirs. The souvenirs are located right across the street from the White House and taken on the afternoon of July 21, 2019 near Pennslyvania Avenue in Washington, D.C.

Getty Images, P_Wei

The American Schism in 2025: The New Cultural Revolution

A common point of bewilderment today among many of Trump’s “establishment” critics is the all too tepid response to Trump’s increasingly brazen shattering of democratic norms. True, he started this during his first term, but in his second, Trump seems to relish the weaponization of his presidency to go after his enemies and to brandish his corrupt dealings, all under the Trump banner (e.g. cyber currency, Mideast business dealings, the Boeing 747 gift from Qatar). Not only does Trump conduct himself with impunity but Fox News and other mainstream media outlets barely cover them at all. (And when left-leaning media do, the interest seems to wane quickly.)

Here may be the source of the puzzlement: the left intelligentsia continues to view and characterize MAGA as a political movement, without grasping its transcendence into a new dominant cultural order. MAGA rose as a counter-establishment partisan drive during Trump’s 2016 campaign and subsequent first administration; however, by the 2024 election, it became evident that MAGA was but the eye of a full-fledged cultural shift, in some ways akin to Mao’s Cultural Revolution.

Keep ReadingShow less
Should States Regulate AI?

Rep. Jay Obernolte, R-CA, speaks at an AI conference on Capitol Hill with experts

Provided

Should States Regulate AI?

WASHINGTON —- As House Republicans voted Thursday to pass a 10-year moratorium on AI regulation by states, Rep. Jay Obernolte, R-CA, and AI experts said the measure would be necessary to ensure US dominance in the industry.

“We want to make sure that AI continues to be led by the United States of America, and we want to make sure that our economy and our society realizes the potential benefits of AI deployment,” Obernolte said.

Keep ReadingShow less