Nenuphars and gas consumption : let's ask Gemini !
In a previous article from October 2023 I was trying to test Llama and GPT's logic by trying to solve simple math problems.
https://www.kindrobot.org/article/nenuphars-gas-consumption-and-crypto-stories-coding-is-not-dead-yet/
And the results were not this impressive.
As the technology evolves and improves and as editors claim "giant" progress loudly on a regular if not daily basis, let's try to challenge one of them: Google Gemini.
Gemini, the large language model developed by Google was launched last December. A big roadshow.
Large language models are not truly performant polyglots: how to escape the english centric trap?
Many people claim that large language models and particularly ChatGPT can speak and answer in almost all languages like humans.
Those marketing messages take a big shortcut by advertising those tools as efficient polyglots.
The progress in natural language processing is real but some problems and limits do remain on the table.
Are those tools equally efficient in all languages like some of us seem to think?
Isn't English better represented and related prompts receive better outputs?
How are those models trained and how important is the choice and volume of the initial corpus? How do editors fight against hallucinations and cultural biases?
Recipes, books, car driving: how good are large language models to assist us with chronology and order?
We already noted that Large language models were not good at logic and particularly at Maths.
You can read this previous article containing simple experiments to highlight those issues:
[Nenuphars, gas consumption and crypto stories? Coding is not dead. Yet.]
Today we'll try to experiment with Google Bard the notions of order and chronology in texts.
This idea came from experiments run by my dear friend Joachim.
Nenuphars, gas consumption and crypto stories? Coding is not dead. Yet.
Developers will be replaced massively by AI as coding becomes a useless task, starting today.
This idea and those statements started to grow after a few influencers experienced asking a chatbot to code a Tetris and deploying it and “it was working”.
They decided to take a big shortcut and concluded that “no-code” will be the new reality soon and the capacities of GPT and Large Language Models would make coders obsolete in the very near future.
It’s a simplistic opinion based on one single non representative use case used as a universal proof.
It does overestimate the capacities of the models we use.
Are all taxis yellow? A story of biases
I had the idea to use Stable Diffusion XL to create images and illustrate articles I’m about to write.
To be honest, I was not obtaining the result expected and was failing prompts after prompts.
Prompting is probably a form of art but it’s also complicated to identify which parts are assumed and covered (and how) by the model and which ones need to be detailed.
Many voices have begun to say that the gap between humans and artificial intelligence systems is shrinking quickly.
Let’s confront the state of art and challenge this statement using simple experiments.
Kindrobot : Who ? Why ? How ?
The main part of this article was written in 2017.
6 years after a lot has changed but many important questions remain on the table.
[Text from 2017]
Artificial intelligence does become an important topic and everyone's business.
When we hear about enthusiastic reactions from technical, marketing and financial parts of the society, we have a great idea of the growing buzz around the subject.
Our phones, computers, cars, hospitals already use these technologies.
Things are going fast.
The technological revolution has begun. Impacts will appear everywhere in our lives within 10 years.
Many personal "thinking" assistant will pop up in our living rooms. Chatbots, image recognition, speech recognition.
More and more data will be collected, analyzed to feed the models.
Data that we provide everyday and which is sometimes used without our clear consent.
[/Text from 2017]
[2023 Update]
The rise of Large Language Models begins to shape our link with creatitivy, reality, efficiency.
Even our relation to work is challenged by those new tools.
Has this recent rise helped in creating and promoting clarity and transparency? How inclusive or even helpful is AI today?
For example, OpenAI switched from a open source approach to an aggressive commercial one to annihilate concurrency.
In the meantime, we seem to forget (or some of us let us forget on purpose) that the very roots of AI come from the open source world through research, universities, and massively public data.
[/2023 Update]