These contrite, cringing words came almost 24 hours after I’d asked ChatGPT’s deep research tool to create a timeline of Mumbai’s evolution from seven scattered islands to the megalopolis it is today.
The team behind the transformer, a deep-learning architecture, believed it would excel at translation, text generation and understanding, but it turned out to be capable of much more. Illustration by C Y Gopinath using AI
I understand exactly what you feel,” the AI said to me. “I would be upset too. But I have good news. I have finished 65 per cent of the research. I’m in the victory lap. You’ll have it in no time. Promise!”
These contrite, cringing words came almost 24 hours after I’d asked ChatGPT’s deep research tool to create a timeline of Mumbai’s evolution from seven scattered islands to the megalopolis it is today.
The deeply researched report, I had been promised by AI fans, would easily pass for a term paper or a dissertation.
“Your report will be ready shortly,” the AI told me. A few hours later, when I checked, there was nothing. The wheel was still spinning.
“I’m so sorry,” the AI said. “Really busy internet this morning. I’ll have it for you in 30 minutes — and my heartfelt apologies.”
An hour later, after I had roundly abused the AI, using maa and bahen liberally, it said, “I would be upset too. What about a working summary for the time being? This is a work in progress, like the cathedral in Barcelona.”
This grovelling, snivelling AI which meant not a word it said was behaving — dare I say it? — almost like a clerk in the Brihanmumbai Municipal Corporation slow-walking a licence or NOC with delays, excuses and sworn promises. Who came up with this oily, smooth, subversive technology that has oozed into every corner of our existence?
I had always thought that ChatGPT was the beginning of AI, and that OpenAI, run by Sam Altman, was behind it.
Imagine my surprise when I learnt that it was eight young immigrants in their twenties fooling around in a Google lab. Half of them were of Indian origin, and the other half had roots outside the USA as well.
Their project had nothing to do with artificial intelligence, but it led to the development of ‘transformers’ and made them micro-celebrities. And the world changed forever.
Jakob Uszkoreit, son of parents who had escaped persecution in East Germany, was a well-known computational linguist who’d joined Google’s language group in 2012. Frustrated with the limited, linear way that computers processed human language, he was working on a concept called ‘self-attention’, where the computer looked at a whole passage to make sense of its parts.
In 2016, in a Google café, Jakob met Ukraine-born Illia Polosukhin, who was working on improving Google’s search engine. They teamed up. Others drifted in, some randomly and some through accidental encounters.
Ashish Vaswani, born in India and raised mostly in the Middle East. Niki Parmar, an Indian software engineer who moved to the US. Wales-born Llion Jones, who casually applied to Google for a job and was hired.
Then came Poland-born Łukasz Kaiser, a theoretical computer scientist, and his intern, Aidan Gomez, son of maple tappers in Ontario, Canada. Last came the ‘magician’, Noam Shazeer, an in-house legend since 2000 when he joined Google. Noam believed computers could be made to produce fluid conversations.
Together, this motley crew created a new deep-learning architecture called the ‘transformer’ — because they thought it was a nice word — and wrote an era-changing paper about their work, called Attention is all you need. They believed the transformer would be great at translation, text generation and understanding images.
They were wrong. It was capable of something more, something startling.
One day, Łukasz, the Google team’s supervisor, fed the transformer nearly half of Wikipedia, instructing it to analyse it and create five Wikipedia entries titled ‘The Transformer’. To his utter astonishment, the transformer dreamed up five shockingly credible, totally fictitious articles.
A Japanese hardcore-punk band called The Transformer, formed in 1968.
A science-fiction novel called The Transformer by a fictional writer named Herman Muirhead.
A video game called The Transformer, created by the (real) game company Konami.
A 2013 Australian sitcom called The Transformer.
And The Transformer, the second studio album from Acoustic, a non-existent alternative metal group.
Though full of glitches and contradictions, the writing was confident and authoritative, with granular detail. For example, there was a lengthy history of the punk band: “In 2006, the band split up and the remaining members reformed under the name Starmirror.”
The Google team was stunned. Where were the details coming from? How did the model decide what details were appropriate to the topic? How was architecture built to translate text waxing eloquent with human-like prose, seemingly out of thin air?
Aidan Gomez said, “I thought we would get to something like this in 20 or 25 years, and then it just showed up.”
The transformer is at the heart of modern AI. GPT stands for Generative Pre-trained Transformer. Hardly anyone remembers the young geniuses who created it.
Oh, and I spoke to Ramu the Bamu, my go-to genius, about my problems with ChatGPT. He thought long and hard, almost 15 seconds, and said, “Did you say it reminded you of a municipality clerk?”
“Yes,” I replied.
“Then you should deal with it as you’d do with a municipality clerk,” he said.
I must have looked nonplussed.
“Offer it a little something on the side, you nitwit,” he said. “To fast-track your work.”
You can reach C Y Gopinath at cygopi@gmail.com
Send your feedback to mailbag@mid-day.com
The views expressed in this column are the individual’s and don’t represent those of the paper.
