Why Open-Source AI is Faster, Safer and More Intelligent than Google or OpenAI

A “moat” historically meant a physical method to reduce threats to a group intended to fit inside it. Take for example the large Buhen fortress on the banks of the Nile. Built by Pharaoh Senwosret III around 1860 BCE, it boasted a high-tech ten meter high wall next to a three meter deep moat to protect his markets against Nubians who were brave enough to fight against occupation and exploitation.

Hieroglyphics roughly translated: “Just so you know, past this point sits the guy who killed your men, enslaved your women and children, burnt your crops and poisoned your wells. Still coming?”

Egyptian Boundary Stele of Senwosret III, ca. 1878-1840 B.C., Middle Kingdom. Quartzite; H. 160 cm; W. 96 cm. On loan to Metropolitan Museum of Art, New York (MK.005). http://www.metmuseum.org/Collections/search-the-collections/591230

Complicated, I suppose, since being safe inside such a moat meant protection against threats, yet being outside was defined as being a threat.

Go inside and lose freedom, go outside and lose even more? Sounds like Facebook’s profit model can be traced back to stone tablets.

Anyway, in true Silicon Valley fashion of ignoring complex human science, technology companies have been expecting to survive an inherent inability to scale by relying on building primitive “moats” to prevent groups inside from escaping to more freedom.

Basically moats used to be defined as physically protecting markets from raids, and lately have been redefined as protecting online raiders from markets. “Digital moats” are framed for investors as a means to undermine market safety — profit from users enticed inside who then are denied any real option to exit outside.

Unregulated highly-centralized proprietary technology brands have modeled themselves as a rise of unrepresentative digital Pharoahs who are shamelessly attempting new forms of indentured servitude despite everything in history saying BAD IDEA, VERY BAD.

Now for some breaking news:

Google has been exposed by an alleged internal panic memo about profitability of future servitude, admitting “We Have No Moat, And Neither Does OpenAI

While we’ve been squabbling, a third faction has been quietly eating our lunch. I’m talking, of course, about open source. Plainly put, they are lapping us. Things we consider “major open problems” are solved and in people’s hands today. […] Open-source models are faster, more customizable, more private, and pound-for-pound more capable. They are doing things with $100 and 13B params that we struggle with at $10M and 540B. And they are doing so in weeks, not months.

One week! Stunning pace of improvement. https://lmsys.org/blog/2023-03-30-vicuna/

It’s absolutely clear this worry and fret from Google insiders comes down to several key issues. The following paragraph in particular caught my attention since it feels like I’ve been harping about this for at least a decade already:

Data quality scales better than data size
Many of these projects are saving time by training on small, highly curated datasets. This suggests there is some flexibility in data scaling laws. The existence of such datasets follows from the line of thinking in Data Doesn’t Do What You Think, and they are rapidly becoming the standard way to do training outside Google.

There has to be common sense about this. Anyone who thinks about thinking (let alone writing code) knows a minor change is more sustainable for scale than complete restarts. The final analysis is that learning improvements grow bigger faster and better through fine-tuning/stacking on low-cost consumer machines instead of completely rebuilding upon each change using giant industrial engines.

…the model can be cheaply kept up to date, without ever having to pay the cost of a full run.

You can scale a market of ideas better through a system designed for distributed linked knowledge with safety mechanisms, rather than planning to build a new castle wall every time a stall is changed or new one opened.

Building centralized “data lakes” was a hot profit ticket in 2012 that blew-up spectacularly just a few years later. I don’t think people realized social science theory like “fog of war” had told them not to do it, but they definitely should have walked away from “largest” thinking right then.

Instead?

OpenAI was born in 2015 on the sunset phase of a wrong model mindset. Fun fact: I once was approached and asked to be CISO for OpenAI. Guess why I immediately refused and instead went to work on massively distributed high-integrity models of data for AI (e.g. W3C Solid)?

…maintaining some of the largest models on the planet actually puts us at a disadvantage.

Yup. Basically confidentiality failures that California breach law SB1386 hinted at way back in 2003, let alone more recent attempts to stop integrity failures.

Tech giants have vowed many times to combat propaganda around elections, fake news about the COVID-19 vaccines, pornography and child exploitation, and hateful messaging targeting ethnic groups. But they have been unsuccessful, research and news events show.

Bad Pharaohs.

Can’t trust them, as the philosopher David Hume sagely warned in the 1700s.

To me the Google memo reads as if pulled out of a dusty folder: an old IBM fret that open communities running on Sun Microsystems (get it? a MICRO system) using wide-area networks to keep knowledge cheaply up to date… will be a problem for mainframe profitability that depends on monopoly-like exit barriers.

Exiting times, in other words, to be working with open source and standards to set people free. Oops, meant to say exciting times.

Same as it’s ever been.

There is often an assumption that operations should be large and centralized in order to be scalable, even though such thinking is provably backwards.

I suspect many try to justify such centralization due to cognitive bias, not to mention hedging benefits away from a community and into a just small number of hands.

People sooth fears through promotion of competition-driven reductions; simple “quick win” models (primarily helping themselves) are hitched to a stated need for defense, without transparency. They don’t expend effort on wiser, longer-term yet sustainable efforts of more interoperable, diverse and complex models that could account for wider benefits.

The latter models actually scale, while the former models give an impression of scale until they can’t.

What the former models do when struggling to scale is something perhaps right out of ancient history. Like what happens when a market outgrows the slowly-built stone walls of a “protective” monarchist’s control.

Pharoahs are history for a reason.

The AI Trust Problem Isn’t Fakes. The AI Trust Problem Is Fakes.

See what I did there? It worries me that too many people are forgetting almost nobody really has been able to tell what is true since… forever. I gave a great example of this in 2020: Abraham Lincoln.

A print of abolitionist U.S. President Abraham Lincoln was in fact a composite, a fake. Thomas Hicks had placed Lincoln’s unmistakable head on the distinguishable body of Andrew Jackson’s rabidly pro-slavery Vice President John Calhoun. A very intentionally political act.

The fakery went quietly along until Stefan Lorant, art director for London Picture Post magazine, noticed a very obvious key to unlock Hick’s puzzle — Lincoln’s mole was on the wrong side of his face.

Here’s a story about Gary Marcus, a renowned AI expert, basically ignoring all the context in the world:

His immediate concern is that the internet will be flooded with false photos, videos and text, and the average person will “not be able to know what is true anymore.”

Will be flooded?

That is literally what the Internet has done since its origin. The printing press flooded the average person. The radio flooded the average person. It’s not like the Internet, in true reinvention of the wheel fashion, grew in popularity because it came with an inherent truth filter.

The opposite, bad packets were always there for bad actors and — despite a huge amount of money invested for decades into “defenses” — many bad packets continue to flow.

Markets always are dangerous, deceptive places if left without systems of trust formed with morality (as philosopher David Hume explained rather clearly in the 1700s, perhaps too clearly given the church then chastised him for being a thinker/non-believer).

“Magritte’s original title of this painting was L’usage de la parole I (The Use of Speech I), which implies that we should question the veracity of the words, not the image.” Source: BBC

So where does our confidence and ability to move forward stem from? Starting a garden (pun not intended) requires a plan to assess and address, curate if you will, risky growth. We put speed limiters into use to ensure pivots, such as making a turn or changing lanes, won’t be terminal.

Historians, let alone philosophers and anthropologists, might even argue that having troubles with the truth has been the human condition across all communication methods for centuries if not longer. Does my “professor of the year” blog post from 2008 or the social construction of reality ring any bells?

Americans really should know exactly what to do, given it has such a long history of regulating speech with lots of censorship; from the awful gag rule to preserve slavery, or banning Black Americans from viewing Birth of a Nation, all the way to the cleverly named WWII Office of Censorship.

What’s that? You’ve never heard of the U.S. Office of Censorship, or read its important report from 1945 saying Americans are far better off because of their work?

This is me unsurprised. Let me put it another way. Study history when you want to curate a better future, especially if you want to regulate AI.

Not only study history to understand the true source of troubles brewing now, growing worse by the day… but also to learn where and how to find truth in an ocean of lies generated by flagrant charlatans (e.g. Tesla wouldn’t exist without fraud, as I presented in 2016).

If more people studied history for more time we could worry less about the general public having skills in finding truth. Elon Musk probably would be in jail. Sadly the number of people getting history degrees has been in decline, while the number of people killed by a Tesla skyrockets. Already 19 dead from Elon Musk spreading flagrant untruths about AI. See the problem?

The average person doesn’t know what is true, but they know who they trust; a resulting power distribution is known by them almost instinctively. They follow some path of ascertaining truism through family, groups, associations, “celebrity” etc. that provide them a sense of safety even when truth is absent. And few (in America especially) are encouraged to steep themselves in the kinds of thinking that break away from easy, routine and minimal judgment contexts.

Just one example of historians at work is a new book about finding truth in the average person’s sea of lies, called Myth America. It was sent to me by very prominent historians talking about how little everyone really knows right now, exposing truths against some very popular American falsehoods.

This book is great.

Yet who will have the time and mindset to read it calmly and ponder the future deeply when they’re just trying to earn enough to feed their kids and cover rising medical bills to stay out of debtor prison?

Also books are old technology so they are read with heaps of skepticism. People start by wondering whether to trust the authors, the editors and so forth. AI, as with any disruptive technology in history, throws that askew and strains power dynamics (why do you think printing presses were burned by 1830s American cancel culture?).

People carry bias into their uncertainty, which predictably disarms certain forms of caution/resistance during a disruption caused by new technology. They want to believe in something, swimming towards a newly fabricated reality and grasping onto things that appear to float.

It is similar to Advanced Fee Fraud working so well with email these days instead of paper letters. An attacker falsely promises great rewards later, a pitch about safety, if the target reader is willing (greedy) to believe in some immediate lies coming through their computer screen.

Thus the danger is not just in falsehoods, which surround us all the time our whole lives, but how old falsehoods get replaced with new falsehoods through a disruptive new medium of delivery: fear during rapid changes to the falsehoods believed.

What do you mean boys can’t wear pink, given it was a military tradition for decades? Who says girls aren’t good at computers when they literally invented programming and led the hardware and software teams where quality mattered most (e.g. Bletchly Park was over 60% women)?

This is best understood as a power shift process that invites radical even abrupt breaks depending on who tries to gain control over society, who can amass power and how!

Knowledge is poweful stuff; creation and curation of what people “know” is often thus political. How dare you prove the world is not flat, undermining the authority of those who told people untruths?

AI can very rapidly rotate on falsehoods like the world being flat, replacing known and stable ones with some new and very unstable, dangerous untruths. Much of this is like the stuff we should all study from way back in the late 1700s.

It’s exactly the kind of automated information explosion the world experienced during industrialization, leading people eventually into world wars. Here’s a falsehood that a lot of people believed as an example: fascism.

Old falsehoods during industrialization fell away (e.g. a strong man is a worker who doesn’t need breaks and healthcare) and were replaced with new falsehoods (e.g. a strong machine is a worker that doesn’t need strict quality control, careful human oversight and very narrow purpose).

The uncertainty of sudden changes in who or what to believe next (power) very clearly scares people, especially in environments unprepared to handle surges of discontent when falsehoods or even truths rotate.

Inability to address political discontent (whether based in things false or true) is why the French experienced a violent disruptive revolution yet Germany and England did not.

That’s why the more fundamental problem is how Americans can immediately develop methods for reaching a middle ground as a favored political position on technology, instead of only a left and right (divisive terms from the explosive French Revolution).

New falsehoods need new leadership through a deliberative and thoughtful process of change, handling the ever-present falsehoods people depend upon for a feeling of safety.

Without the U.S. political climate forming a strong alliance, something that can hold a middle ground, AI can easily accelerate polarization that historically presages a slide into hot war to handle transitions — political contests won by other means.

Right, Shakespeare?

The poet describes a relationship built on mutual deception that deceives neither party: the mistress claims constancy and the poet claims youth.

When my love swears that she is made of truth
I do believe her though I know she lies,
That she might think me some untutored youth,
Unlearnèd in the world’s false subtleties.
Thus vainly thinking that she thinks me young,
Although she knows my days are past the best,
Simply I credit her false-speaking tongue;
On both sides thus is simple truth suppressed.
But wherefore says she not she is unjust?
And wherefore say not I that I am old?
O, love’s best habit is in seeming trust,
And age in love loves not to have years told.
Therefore I lie with her and she with me,
And in our faults by lies we flattered be.

American Bemoans “Big Dumb and Blind” Truck Options, Demands Import of Chicken Little Trucks

You may recall my 2018 post about the supply-chain ethics of Toyota light trucks used for militant political destabilization.

Basically when technology gets battle-tested with life-and-death results, the optimal tools for the job become somewhat self-evident.

If driving around to shoot things is the job, then a light truck (before light electric-motorbikes arrived) repeatedly came out favored in many theaters of conflict. Night after night Americans tended to see in their news feed how armed militias around the world terrorized civilians thanks to the Toyota’s Hilux (instead of motorbikes).

Source: NYT “Without a Motorcycle in Kandahar, ‘You Are Like a Prisoner'”

And… somehow that turned to jealousy, then a demand in Kansas.

A southern Californian who moved recently has posted to the Wichita press that he really, really needs a light truck to feel free.

Many’s the time I’ve turned on the nightly news and seen Taliban or ISIS militants tooling around in mini-trucks, mostly Toyotas, with machine guns bolted to the bed “Rat Patrol” style. Every time I see that, I say to myself (or anyone unlucky enough to be in earshot) “There, that’s the truck I want” — minus the machine gun, which I’d only need if I were driving Kris Kobach in a parade.

There’s a lot to unpack in this opinion piece.

First, a snide reference to a notoriously political extremist Kansas’ Kris Kobach (KKK) is legit. I can’t ignore it.

Kobach campaigned for office by literally swinging a .50 caliber machine gun on a Jeep towards children already traumatized by school shootings.

Sad fact: the Kansas parade organizers had to point out to team Kobach that although the parade entry form expressly prohibits throwing candy (an unapologetic denial of freedom), it also shouldn’t have to prohibit someone terrorizing children with guns.

Kobach’s tone-deaf response to public outcry (from a bunch of gun-toting life-long Republicans, mind you) was he flashed his prohibition-era white privilege card — claimed his exact replica machine gun was to stoke hysteria and fear, yet unable to actually shoot bullets (or candy). He argued a tiny legal loophole exists in the Constitution just for people like him — privileged rich white boy — that determines who gets to play domestic terrorist and who doesn’t.

There surely is no worse politician than this Harvard-trained radical xenophobic nativist lawyer named Kobach, although I’m told the also Harvard-trained Meatball DeSandwich may prove me wrong.

According to [junior officer DeSantis’ supported] narrative, the dead men bound their hands and feet, stuck cloth deep down their own throats, fashioned nooses from strips of material, climbed on their washbasins with the noose around their neck and stepped off.

Hello there Harvard law graduates Kobach and DeSantis. Talk about a supply-chain that is sorely lacking any safety!

Anyway, back to talking about the problem with Toyotas.

Second, the author’s recent move from southern California to Kansas (that he freely admits) is foundational to the thesis of his complaint.

I grew up in very rural Kansas, living off the land with all the heaping amounts of trucks and guns you’d perhaps expect. Then, after a detour in between (e.g. travel through Europe and Asia), I moved out to California.

When I first read the jealousy-laced upwardly-mobile piece on importing trucks, I DID NOT hear the familiar voice of modest, hard-working Kansans grounded in utility and sacrifice.

That shrill gimme gimme, consumption competition just for public perception narrative is unmistakably, definitely southern California culture.

Mini-trucks — mostly Toyotas, but also Ford Couriers and Chevy Luvs — were once ubiquitous on the streets and freeways of southern California, where I lived from the late 1970s to the late 1990s. […] I once transported enough salvaged solid oak hardwood flooring to redo our entire kitchen.

Once. He ONCE carried a lot of wood. For a kitchen remodel.

Excuse me as I clean up, as I just vomited my home grown apple pie all over myself.

Driving pickup trucks on freeways everyday just so one day, ONE day, you can carry a load of fancy nancy kitchen remodel flooring, is the most Californian thing ever.

That is not a pickup truck job. Get a station wagon. Or rent a box truck. It was just a day and some kitchen clean flooring. Exactly backwards thinking.

Try hauling rotting wet hay, forest felled lumber, manure and dead animals every day and then — just one day — needing to drive 60 miles high-speed on a highway to the nearest hospital because your hand was dangerously mangled by tools. Tractors don’t get you there on time.

That is a pickup truck used for a real life lived every day.

See the problem with how the wrong “once” context feeds into constant truck worship?

More to the point, the whole lack-of-thought piece about light trucks centers on a fetishization of terrorists as power-projection (see Kobach point above).

…when it comes to buying a pickup truck, we might once again be as free as the Taliban.

Southern Californian white man claims he doesn’t feel safe sitting in his oak floored kitchen because he needs to buy some frivolous and unnecessary vehicle to park outside. He demands the kind of unconstrained consumption that he dreams comes under the fear-based toxic masculinity of terror groups.

Freedom versus fear. News at 11.

This chicken little author says he wants the “best” trucks for society, yet it is based on a measure of utility by armed groups causing societal disruption (not to mention a one-time kitchen remodel or new lounger).

That is really him thinking about what is best to address his own weird fears, and has little to do with helping anyone else. He doesn’t even need a truck, big or little.

Third, and perhaps most tellingly, the big stated reason for getting rid of big trucks is two-fold: they’re called dangerous to people/planet and they are called artificially protected by market regulation.

…because of an antiquated trade policy levying a 25% tariff on imported light trucks, in retaliation for a European tariff on U.S. chicken.

So the author is saying regulations on trade need to be removed and somehow the market will magically start to fix the environment, without any reason to do so.

How? By importing the light trucks favored for home remodels and terrorism…? It doesn’t make any sense and reads like The Onion. Basically the underlying white guy libertarian trope has been turned into a paper-thin argument of societal benefit from small trucks.

It’s the sad rhetorical tactic of taking things that others’ care about (people/planet) and slapping them like an easy label over a completely disconnected personal political platform.

You want a truck that people who hate government use? And you hate the government too? Write an op-ed that says removing government regulation on light trucks will achieve X, where X is whatever other people say they want.

As much as I am a fan of light trucks, way back in 2000 I became the first person in California to import a Japanese four door light truck (the DMV in Santa Cruz was VERY confused and took hours to classify my light truck as a car instead of big truck).

In other words I have very quietly and modestly achieved the exact thing that this author has been yelling at his TV can’t be done. My import feat never seemed like anything worth writing about, from 20 years ago to today.

Incidentally, remember the Naan-truck story of 2010? That’s the kind of stuff worth writing about.

I’d do an import again right now to prove my point, except it’s super easy to find light trucks in California despite any need for them being long gone. Ride an electric bike! Even my four-door light truck classified by Santa Cruz DMV as a car is nothing special anymore, as you’ll see in a minute.

I thus find none of this author’s published arguments convincing let alone amusing.

Bottom line is he shows a side of America that conflates constant ugly power projection by unfiltered media with some kind of personal call to grand utopian posturing about ending government regulations.

The real story here is nervous white men in their upscale remodeled kitchens watching relentless fear-driven news feeds of trucks and guns produces a culture desperate for such trucks and guns falsely tied to sense of safety, devoid of actual rational relationship to service, duty or utility.

Unkindness shouldn’t be a political platform.

Pushing anti-societal terror platforms as somehow kind to the planet is the kind of bogus mental gymnastics we see out of California fraud factories such as Tesla. Their big truck, marketed as an apocalyptic tool meant for battle with all humanity, has design flaws so egregiously bad it wisely has been banned by safety experts even before release.

Pretending societal benefits flow from owning a Toyota known for anti-government warfare… doesn’t sufficiently hide an underlying unhinged libertarian screed that someone wants cool toys to project supreme power over everyone including the government they heavily depend upon.

Now, I should give the author credit here too. He has made me realize Tesla clearly missed the boat on their true target market. They would have been far better off just putting their little badge on the Toyota light truck (like everything else Tesla “makes”) and selling them into the chicken Californians hiding from society in posh suburban bunkers of Kansas.

Perhaps I can’t say this all any better than Hyundai? Their answer already has been… go buy a Santa Cruz and get a life, dude. Maybe try fishing instead of watching scary news?

“Just enough truck”. In 2000 I imported the first Japanese 4-door light truck to California, thanks to the Santa Cruz DMV. In 2022 Hyundai announced general availability of the Santa Cruz light truck. A great option for someone still too scared to ride an electric bike. Source: Car and Driver

U.S. Warrantless Search Has Dropped Significantly

In a new WSJ report, a harsh critic of warrentless searches holds the line.

“It still adds up to more than 300 warrantless searches for Americans’ phone calls, text messages, and emails every day,” stated NYU’s Elizabeth Goitein [senior director of the Brennan Center for Justice’s Liberty & National Security Program]. “One warrantless search is too many.”

One is too many?

I’m not following her logic, which is basically that she believes the Fourth Amendment doesn’t say anything about victims.

The right of the people to be secure in their persons, houses, papers, and effects, against unreasonable searches and seizures, shall not be violated, and no Warrants shall issue, but upon probable cause, supported by Oath or affirmation, and particularly describing the place to be searched, and the persons or things to be seized.

“Unreasonable” could be about victims.

If we were talking stopping deaths, in other words, then a zero goal makes sense. There’s no recovery from death. By comparison a goal of zero searches doesn’t make sense in the context of safety. A recovery from a warrantless search mistake seems entirely possible (even given rights violations), and there are surely examples of it being a lesser problem than death.

There are searches for missing people that are called off when it was a mistake. I don’t see people saying there should never be another search made, in that sense, because even a mistaken search still is intended to fight FOR victims’ rights.

It’s a bit like a critic of searches saying there should be zero people running out emergency exits. In a sense that sounds right, people running out the emergency exits can be a very bad thing. Yet instead of focusing on zero deaths from fires that caused use of exits, I see a fixation on stopping people from using the emergency mechanism even under a reasonable purpose of surviving in an emergency.

In 1911 there were 146 people dead within 20 minutes due to a lack of emergency exits in the NYC Triangle Factory fire. Someone arguing after that even one emergency exit use is too many… doesn’t seem to address an underlying infrastructure doctrine for something reasonably engineered to save lives in emergencies.

I also would cite here lessons from the 1905 Grover Shoe Factory fire (killing 58), yet suspiciously nobody in America seems to remember it… despite me putting details in every presentation I give on ethics in computer science.

…inspectors could not see between the two layers of a lap joint…

And that gets right into the second problem with the hard-line of criticism. The warrantless searches under Section 702 of FISA are purported to have saved lives.

Assuming there is any validity at all for this claim (e.g. the NSA Director says “We saved lives” so it’s plausibly true), then removing all the warrantless searches presupposes the opposite, they aren’t actually needed to save lives. It emphasizes removing them without explaining why or how lives would be saved some other way, or even why lives lost is a reasonable tradeoff for having no emergency exits.

Food for thought.

Critics will of course point to the FBI being itself subjected to corrupt and racist leadership, used to surveil and smear political opponents (e.g. Hoover targeting MLK under the pretense of “foreign” influence, leading to assassination).

On the flip side, federal prosecutors do find and prosecute Americans under foreign influence.

Former Transportation Secretary Ray LaHood admitted to federal prosecutors that he intentionally excluded from his financial disclosures a $50,000 loan he obtained while in office from a billionaire foreign donor, a document released by the Justice Department said Wednesday. During an interview with the FBI in 2017, LaHood initially denied receiving the loan, but later acknowledged the payment after being shown a copy of the $50,000 check he received in 2012, according to a non-prosecution agreement LaHood signed with federal prosecutors in Los Angeles.

Oops. Lately this LaHood guy, after his non-prosecution agreement with the FBI, has been a very loud critic of “Section 702 acquired information”, a twist that is perhaps best explained by EPIC.

Anyway, back to the story. The WSJ report on the ODNI report reveals that the FBI conducted about 200,000 warrantless searches of 120,000 Americans’ calls, emails & text messages in 2022.

That’s a significant drop from 3.4 million searches in 2021. Apparently the FBI is 1) complying better with restrictions due to reforms put in place, but also 2) foreign spies are behaving differently, drawing less 702 attention.