AI – artificial intelligence not artificial insemination – is the latest of quickly forgotten hype over technology, replacing our machine learning, blockchain, Metaverse, 3D goggles and so many other long-forgotten hype.
When most people talk about AI, they’re talking about Large Language Models. These models use machine learning to make computer programs produce sentences that sound like they were written by a person. They take words and phrases from what people type or say and generate a response. It has impressed some people, but the results are often rough and incorrect.
Wikipedia defines Large Language Models this way:
A large language model (LLM) is a computerized language model, embodied by an artificial neural network using an enormous amount of “parameters” (“neurons” in its layers with up to tens of millions to billions “weights” between them), that are (pre-)trained on many GPUs in relatively short time due to massive parallel processing of vast amounts of unlabeled texts containing up to trillions of tokens (parts of words) provided by corpora such as Wikipedia Corpus and Common Crawl, using self-supervised learning or semi-supervised learning,[1] resulting in a tokenized vocabulary with a probability distribution. LLMs can be upgraded by using additional GPUs to (pre-)train the model with even more parameters on even vaster amounts of unlabeled texts.[2]
The invention of the transformer algorithm, either unidirectional (such as used by GPT models) or bidirectional (such as used by BERT model), allows for such massively parallel processing.[3] Due to all above, most of the older (specialized) supervised models for specific tasks became outdated.[4]
In an implicit way, LLMs have acquired an embodied knowledge about syntax, semantics and “ontology” inherent in human language corpora, but also inaccuracies and biases present in the corpora.[4]
At one level, that sounds quite impressive, but in reality, the output of programs like ChatGPT is, at best, mediocre. In my experience, ChatGPT has very limited knowledge of numerous subject matters. Its understanding of contemporary or specialized topics is lacking and falls short when compared to information found in a conventional encyclopedia. The answers generated by ChatGPT are bland, lacking depth and refinement. Furthermore, they frequently contain numerous mistakes and inaccuracies. The fear of expressing something insensitive or illegal constrains ChatGPT, leading it to avoid engaging in discussions or providing information on certain topics. As a result, the range of conversations and knowledge it can actively engage with becomes significantly limited.
But what I object most to the hype of Artificial Intelligence is how the market leaders want to cement their position in law by tightly controlling the technology, aiming to restrict competition through government regulation and extensive lobbying efforts. The leading manufacturers of AI systems actively seek to impose a ban on open-source models, thereby hindering individuals and institutions from embracing this technology for their own advantage. They often cite various conjectured threats and fears surrounding the potential misuse of AI. However, the irony is that regardless of how successful these AI lobbyists may be, those operating beyond the jurisdiction of United States law will likely persist in freely developing AI, ultimately allowing it to permeate back into the United States.
Useful software, particularly open source software, moves around the internet like water. You can only resist it’s penetration temporarily. Governments may attempt to build barriers in an effort to suppress it, but the source code effortlessly permeates the vast expanses of the internet, evading the clutches of regulatory bodies. It brings to mind the enduring controversy surrounding the DeCSS library, which enabled Linux users and others to freely view and duplicate unencrypted DVD movies, expanding the domain of online sharing. In a futile endeavor, numerous regulators and movie companies endeavored to eradicate DeCSS from the web, yet it persistently eluded their grasp. In fact, it has now become a commonplace component of Linux distributions, further illustrating the ineffectiveness of their efforts. While the AI advocates may strive to proscribe the unrestricted utilization and dissemination of Large Language Models, their attempts will prove fruitless, as these models will continue to be employed and exchanged without the government’s control. Even if a brief period of suppression were to occur, it would merely hinder progress in the United States and select countries, rather than stifle it completely.
Large language models are likely to become an every day part of computing, especially as open source makes them widely available. There will be a lot of fear and loathing over AI but I doubt they’ll change life nearly as much as their promoters and the Luddites make them out be. Large language models with be an evolution, making computing better in ways behind the scenes but won’t change much in every day life in society.
Those of a certain age can remember the illegal prime number shown as an image, that for a number of years was considered a crime to distribute, at least in the United States (this image’s colors plus C0 is the code to decrypt most DVDs). Worked out so well for the government, which eventually decided this was so stupid and forgot about it by the mid-2000s.
Last night when I was doing one of those Zoom chats with John Wolcott and the Pine Bush gang we started talking about how problematic commercial software is and how you can get stuck with files that are hard to access once you upgrade your computer. Ever since I’ve been running Linux I’ve never had that problem – old software can usually be re-compiled on modern machines as source code is available and most Linux apps use well documented and fully open file formats.
Computers, computers, computers wherever you turn.
Those chips are so loaded with hot information
You’d think they would burn.
Some of it’s factual, actual,
Some of it’s made of thin air.
Whatever gets in a computer
Stays there.
You can put almost anything in there that comes to your mind.
The programmer gets lost in the shuffle, the scuffle,
The dope stays behind.
Some of its factual, actual,
Some of it is double-faced.
Whatever gets in a computer
Isn’t erased.
Our lives have been fed to computers, every thought, every dream,
Everything that we’ve bought that has rusted or busted
Or split at the seam,
Every up, every down,
Every howl, every glimmer of luck.
When something gets in a computer
It’s stuck.
The stuff that we have in our heads is a different affair.
We’ve hoarded and sorted, amended and bended
And let in the air.
But computer banks grow like a cancer,
They can always produce a wrong answer
And they never are troubled with doubt.
And once you get in a computer
You never get out.
I don’t use Microsoft Windows much. But when I do use it, the thing that is most noticeable is the enormous security and features updates that take forever to update. Literally, it took my laptop 25 minutes of unusable time complete with multiple reboots to get working the other day, meaning I would have missed my Zoom meeting if not for my phone.
In Linux, all updates can be done in the background using apt-get or a similar package manager. Linux doesn’t require reboots during updates, nowadays even kernels can in many cases be hot swapped. Most other software and libraries can be updated in the background. Yes, every two years there are big distribution updates for Ubuntu which often require a single reboot to load the latest kernel and features but it’s not usually mandated right away, and the reboot is just an ordinary reboot with no delays.
The windows update process really should be faster, requiring less files be downloaded for patching bugs and should occur in the background. The windows infrastructure should be redesigned to allow libraries to be updated behind the scenes without reboots and if a library has breaking changes, it should only require you to restart that app. Likewise, windows should tell you exactly what is updating during system updates, not keeping you in the dark – maybe hide the real technical information but at least give you a much better idea what is happening in the background (patching the security bug that could xxxx) along with estimated time, percent done and progress.