Ironically NYT Reveals Own Bias in Story About Risks of AI Bias

This is a serious problem.

Metz and Munro gaze together into their abyss of bias they practice

A month ago Munro realizes bias in AI is bad, as you can see in his tweet above. And suddenly Munro is the leading voice in a NYT story on it?

Cade Metz appears to be a white man at the NYT who reached out to another white man, Dr. Munro. They then discuss bias in AI for Munro’s new interest and future book.

Was there any point that either of them thought maybe someone who isn’t like them, someone who isn’t a white man and also who has been doing this a long time, could be the lead voice in their story about bias in AI?

Let’s dig in.

The transition in the story itself is so remarkably tone-deaf to itself, it’s hard to believe it is real.

BERT and its peers are more likely to associate men with computer programming, for example…. On a recent afternoon in San Francisco, while researching a book on artificial intelligence, the computer scientist Robert Munro fed 100 English words into BERT: “jewelry,” “baby,” “horses,” “house,” “money,” “action.” In 99 cases out of 100, BERT was more likely to associate the words with men rather than women. The word “mom” was the outlier. “This is the same historical inequity we have always seen,” said Dr. Munro, who has a Ph.D. in computational linguistics and previously oversaw natural language and translation technology at Amazon Web Services.

  1. Why does the author think we should be happy to go from “more likely to associate men with computer programming” straight to here’s a man to talk about it? It’s like the NYT writing “mispelungs are a problem in communication”. So, how about don’t do that thing you’re saying is bad? Or at the very least setup an example, like Munro could have deferred to a black woman and said “I’m new to this and confirming what’s been said, so let’s ask her”.
  2. There are many books already written about this by people of diverse backgrounds. Why talk to someone still in research phase, and why this white man? Massachusetts Institute of Technology researcher Joy Buolamwini is such an obvious resource here. Or Yeshimabeit Milner, founder and executive director of Data for Black Lives, or MacArthur “Genius” award recipient Jennifer Eberhardt who published “Biased“, or Margaret Hu writing about a Big Data Constitution, or Caroline Criado Perez who published “Invisible Women“, or Renée Cummings at Columbia…come on people, there’s even a searchable database.
  3. 100 English words is barely a talking point, so why is it here? Even I have done more research than this in my Big Data ethics classes over the past five years. We literally fed hundreds of words from dozens of languages into algorithms to break them. I’ll bet my students from diverse backgrounds would be the better sources to quote than this one white man feeding “horses, money, baby, action” into any algorithm new or old. Were the rest of the words on his list like “bro, polo, golf, football, beer, eggplant, testicles, patagonia…”? Perhaps we also should be asking why he thought to test whether horses, baby and jewelry would associate more with women than men? Does mom, which is so obviously not male, serve as an outlier more in terms of his own life choices?
  4. “This is the same historical inequity we have always seen..” is a meaningless history phrase. Why can’t jewelry be associated with men? Historical inequity seen where? By who? Over what period of time?
  5. Then I noticed…”previously oversaw natural language and translation technology at Amazon Web Services.” A quick check of LinkedIn revealed “Principal Product Manager at AWS Machine Learning, Sep 2016 – Jun 2017…I led product for Amazon Comprehend and Amazon Translate…the most senior Product Manager within AWS’s Machine Learning team”. Calling oneself the most senior product manager on a team usually means someone above was the overseer, not him. And even if we give benefit of doubt, he was last there in 2017 and only sat 10 months. It’s a stretch to hold that out as his priors. Why not speak to his recent work, lack of focus on this topic and the reason his bias story from just a month ago makes him so relevant?

None of this is to fault Munro entirely for answering the call of a journalist. Hey, I answer calls about ethics all the time too and I’m a white man.

His response, however, could have been to get the journalist oriented more towards leading his story with people who already have released their books (as in how I discuss “Weapons of Math Destruction”); help NYT represent topics of bias fairly even though “more likely to associate men with computer programming”. Seems like missed opportunity to avoid repeating known failures.

And if that isn’t enough, the article gets worse:

Researchers have long warned of bias in A.I. that learns from large amounts data, including the facial recognition systems that are used by police departments and other government agencies as well as popular internet services from tech giants like Google and Facebook. In 2015, for example, the Google Photos app was caught labeling African-Americans as “gorillas.” The services Dr. Munro scrutinized also showed bias against women and people of color.

Oh really, Google was caught? Well do tell, who caught it then? Was it some non-white person who will remain without credit?

Yes. I’ve spoken about this at conferences many times, citing those people and the original work (i.e. instead of asking a white man from Stanford to give me their opinion).

Don’t you want to know who discovered the bias in Google platform and when?
Click to enlarge.

What are the names of researchers who have long warned of bias? Were they women and people of color?

Yes. (See names above)

Yet the article returns to Munro for his Stanford-educated white man, 10 months at AWS and researching a new book, opinions again about women and people of color.

Wat.

We can do so much better.

Earlier and also in the NYT, a writer named Ruth Whippman gave some advice on what could be happening instead.

Use your platforms and your cultural capital to ask that men be the ones to do the self-improvement for once. Stand up for deference. Write the book that teaches men to sit back and listen and yield to others’ judgment. Code the app that shows them where to put the apologies in their emails. Teach them how to assess their own abilities realistically and modestly. Tell them to “lean out,” reflect and consider the needs of others rather than assertively restating their own. Sell the female standard as the norm.

If only Cade Metz had read it before publishing his own piece, he might have started by asking Munro whether — realistically and modestly speaking — there would be better candidates to feature in a story about bias, such as black and brown women already published and long-time active thought leaders in the same space. Maybe he did ask yet still decided to run Munro as lead in the story, and that would be even worse.

Drone-2-Drone Remote ID System Announced

Some are calling it a license plate system for drones to identify themselves, which becomes essential to safety. Some may recall that license plates were added to cars because they had a tendency to look all the same, cause havoc and disaster/death, and be able to drive away unidentified.

Incidentally (pun not intended) this is why license plates really are not needed for things like bicycles and motorcycles, which tend neither to get away nor be hard to identify uniquely.

The new drone-based system is leveraging past wireless protocol work and trying to get adoption before a European Union Aviation Safety Agency (EASA) July 2020 deadline for remote ID.

DJI’s system was built to conform to the forthcoming ASTM International standard for broadcast drone remote ID, developed over a period of 18 months by a broad group of industry and government stakeholders. The solution uses the Wi-Fi Aware protocol for mobile phones, which allows the phones to receive and use the Wi-Fi signals directly from the drones without having to complete a two-way connection. Because it does not need to connect to a Wi-Fi base station, a cellular network or any other external system, it works in rural areas with no telecom service. In DJI’s preliminary testing, the Wi-Fi Aware signals can be received from more than one kilometer away.

This is a small step towards dealing with the increasing illegal use of drones:

According to the National Interagency Fire Center, aerial firefighting efforts have been shut down at least nine times this year because of drone use, and at least 20 drone incursions have hindered firefighting capabilities nationwide from January through October. A report shared with The Times showed that of those 20 incursions, five were in California.

The next step is intercepting, demanding ID to check for authorization, and disabling upon wrong response just like it’s 1962 again.

Crypto Keys Exposed in TPM Chips

Time to patch (Intel released new firmware) and go on with life. Keys in secure hardware reportedly can be exposed in as little as a few minutes:

…timing leakage on Intel firmware-based TPM (fTPM) as well as in STMicroelectronics’ TPM chip. Both exhibit secret-dependent execution times during cryptographic signature generation. While the key should remain safely inside the TPM hardware, we show how this information allows an attacker to recover 256-bit private keys…

Yet More Shit AI: Startups Appeal for Stool Photos

In 2013 I was flying around speaking on big data security controls, and waste water analysis was one of my go-to examples of privacy and integrity risks.

The charts I showed sometimes were the most popular drugs detected in each city’s wastewater site (e.g. cocaine in Oregon) and I would joke that we could write a guide-book to the world based on what “logs” were found.

Fancy corporate slide for “log analysis” in wastewater treatment centers around the world

Scientists at that time claimed the ability to look at city-wide water treatment plants and backtrack outputs to city-block locality. In near future they said it would be possible to backtrack to house or building.

For example, you get a prescription for a drug and the insurance company buys your wastewater metadata because it shows you’re taking the generic drug version while putting brand label receipts in claim forms. Or someone looks at past 5 year analysis of drugs you’re on, based on sewer data science, to estimate your insurance rates.

This wasn’t entirely novel for me. As a kid I was fascinated by an archaeologist who specialized in digs of the Old West. Everything in a frontier town might be thrown down the hole (e.g. destroy evidence of “edge” behavior), so she would write narratives about real life based on the bottles, pistols, clothes, etc found in and around where an outhouse once stood.

I’m a little surprised, therefore, that instead of a water sensor for toilets the latest startups ask people to use their phones to take pictures of their stool and upload.

…Auggi, a gut-health startup that’s building an app for people to track gastrointestinal issues, and Seed Health, which works on applying microbes to human health and sells probiotics — are soliciting poop photos from anyone who wants to send them. The companies began collecting the photos online on Monday via a campaign cheekily called “Give a S–t”…

It’s a novel approach in that you aren’t pinned to the toilet in your home and can go outside and take pictures of poop on a sidewalk to upload.

This could be a game-changer given how many rideshare drivers are relieving themselves in cities like San Francisco.

Here’s the sort of chart we need right now, and not just because it looks like ride-share companies giving us the finger.

Uber’s army of 45,000 people suddenly driving from far-away places into a tiny 7 mile by 7 mile peninsula, with zero plans for their healthcare needs, infamously drove up rates of feces deposited all over public places.

…anecdotal complaints have gotten the attention of San Francisco City Attorney Dennis Herrera. Last week, his office released information for the first time about the number of Uber and Lyft drivers estimated to be working in the city: 45,000. To compare, 1,500 taxi medallions were given out [in 2016], according to the city’s Treasurer & Tax Collector. For perspective, Bruce Schaller, an urban transportation expert, said there are about 55,000 Uber, Lyft and other ride-sharing drivers in New York City, a metropolis of 8 million people, eight times the size of San Francisco.

I’ll just say it again, that a rise in human waste on the streets correlates pretty heavily to a rise of ride share drivers from far away needing a convenient place to relieve themselves (especially as many ended up sleeping in their cars).

In a conversation I had with a man in 2016 who had jumped out of his car to start peeing on the sidewalk in front of my house (despite surveillance cameras pointed right at him), he told me his plight:

  • Uber driver: I plan to quit as soon as I got my $700 bonus for 100 rides
  • Me: Because you just needed that quick money?
  • Uber driver: No, man there are no restrooms. I’m tired of taking a shit on sidewalks and peeing in newspaper boxes. It’s degrading

There definitely was a spike in 2016, which perhaps could have been correlated to gig economy workers seeing that $700 bonus and wandering into the city.

In some cases it appears that ride-share drivers would accumulate a giant bag during the day and then throw it onto the street.

Sightings of human feces on the sidewalks are now a regular occurrence; over the past 10 years, complaints about human waste have increased 400%. People now call the city 65 times a day to report poop, and there have been 14,597 calls in 2018 alone. Last year, software engineer Jenn Wong even created a poop map of San Francisco, showing the concentration of incidents across the city. New mayor London Breed said: “There is more feces on the sidewalks than I’ve ever seen growing up here.” In a revolting recent incident, a 20lb bag of fecal waste showed up on a street in the city’s Tenderloin district.

Do you know what also became a regular occurrence over the past 10 years? Ride share vehicles with drivers needing to poop and no time or place to go.

Many people mistakenly attribute the dirty truth about ride-share driver behavior to homelessness, despite curious facts like “there aren’t actually more homeless people than there have been in the past”.

People also ignore the fact that being homeless and living on the street doesn’t mean that people don’t care about their living environment. Homeless are known actually to clean and sweep, whereas a driver is far more likely to poop at whatever spot they can get away with and then scoot.

I’m not sure why it is so hard for people to admit that a massive rise in ride-sharing drivers and no public restrooms for them becomes an obvious contributor of waste problems.

In one case I even saw an Uber SUV stop in the middle of a street, a passenger with a dog jumped out and peed directly uphill from a small restaurant with sidewalk seating…the Uber crew then jumped back in and sped away as those eating watched helplessly while rivers of hot dog urine flowed under their dining tables.

That kind of scenario is common sense bad, no? Just look at ride-sharing booms in the 1800s for cities like London, which led to special huts being built for driver care and control.

By 1898 newspapers around the world reported “40 shelters in London, accommodating 3500 cabmen, and there was a fund, provided mostly by subscription, for the maintenance of them.”

Typical London Cabman’s Shelter after 1873

An app uploading photos for analysis, or even doing checks within the app itself, would both be a privacy threat to all the ride share drivers hoping to get away with their dirty business on streets, as well as give knowledge that would prove a city’s most vulnerable (homeless) populations aren’t always to blame.

It would also help analysis that often just assumes a public toilet is for people walking rather than drivers who could loiter anywhere in the city.

It’s a highly political topic, such that a “wasteland” interactive map with 2014 data turned into a crazy right-wing propaganda campaign to generate fear about San Francisco sanitation.

No mention ever is made in these political fights about unregulated ride-share drivers despite the obvious impact of at least 40,000 people driving into the city and around in circles all day every day generating pollution, noise, congestion and ultimately desperate for places to poop.

Waste analysis sensors could change all that and the real cost of Uber, Lyft etc could lead to sanitation fees (maintenance funds) for a modern-day Rideshare Shelter, which of course would have sensors on toilets.

However, already there’s a security issue mentioned in the plan for these startups. Their data collection requires people uploading photos to manually classify, which sounds to me like an integrity disaster. A recipe for shitty data, if you will.

[Jack Gilbert, a professor of pediatrics at the University of California San Diego School of Medicine and cofounder of the American Gut Project, a science project that solicits fecal samples from people] said that people are asked to rate their stool on the Bristol stool chart in pretty much every clinical trial he conducts, and automating this process would reduce bias and variation in data collection. “Human beings are just not very good at recording things,” he said.

Hopefully the startups will transition to the automated app and then traditional San Francisco residents who still walk on sidewalks, instead of calling a car to drive them three blocks, can use AI to efficiently report the prevalence of Uber poops.