AI & ML


AI tools: the good, the bad, and the ugly

29 March 2023 AI & ML

Like most ChatGPT users, I’ve been blown away by the sheer human-like quality of the language model that has been produced by OpenAI. Currently it is an amazing product. At Altron Karabina, the data and analytics team are experimenting with using the tool to verify and optimise code, such as SQL stored procedures and Python scripts. Salespeople are also using the tool to produce succinct summaries of complex products to help clients understand them better.

The good

The promise of AI tools is that they can supplement human endeavour. A lawyer, financial adviser, marketing manager or programmer who can use ChatGPT, LaMDA, or other AI tools effectively is going to be more productive, and possibly more inventive than one that cannot. AI technology has the potential to supercharge human work, and there is no doubt that the potential is there to create more jobs and more prosperity, not less.

The bad

However, technology such as ChatGPT, like all human inventions, is a double-edged sword and can be used for good or ill. ChatGPT declares these concerns itself in almost every chat. In 2019 OpenAI decided to withhold the release of GPT-2, a previous version of its language model, citing concerns about the tool, which could generate convincing news articles, being too easy to use for misinformation purposes.

There are several issues. Firstly, the tech itself can produce misleading, wrong or damaging information. AI models can be impacted negatively by the content that they learn from.

OpenAI has put in a lot of effort to try and avoid the negative effects of training language models on a large corpus of documents, such as Twitter, Wikipedia, and various sections of the internet, by having humans tag text that is violent, racist, misogynistic or otherwise unacceptable, to ensure that it doesn’t contaminate the user experience with GPT-3. It’s good to see that organisations like OpenAI can be self-regulating in this regard, but can we be sure that all AI businesses will do the same? The EU is preparing regulations that might help with this, the AI Act, expected to be made law in 2023, but the impact remains to be seen (https://artificialintelligenceact.eu/).

The ugly

The last concern is the damage that can be done in the process of creating the technology itself. As I write, Microsoft, OpenAI and GitHub are defending a class action lawsuit that alleges that the corpus of programming code that was used to train GPT-3 contains licenced code that should not be profited from. Matthew Butterick, who filed the suit, states that the creation of CoPilot, GitHub’s GPT-3 based coding assistant, is “software piracy on an unprecedented scale”.

As a throwback to the previous concern, the coding Q&A; site, Stack Overflow, has banned AI-generated answers to programming questions, saying “these have a high rate of being incorrect”. This concern must also apply to the millions of non-coding articles and books that GPT-3 has been trained on – who really owns the poems, scripts and movie scripts generated by ChatGPT?

In addition, Time reported in January 2023 that the very act of labelling some of the internet’s most offensive content to reduce any potential toxic output from GPT-3, has also caused damage. This work is generally outsourced to countries where labour is cheap, and a Kenyan company that paid workers less than $2 an hour for the task of labelling troubling material, seems to have had issues with employees who allege emotional trauma dealing with the nature of the content they vetted. OpenAI is by no means the only AI company that uses low-cost labour for this purpose – last year, Time published another story about the same Kenyan company performing labelling work for Meta in the article ‘Inside Facebook’s African Sweatshop’.

The verdict

The impact of AI is far-reaching. Like many human inventions since fire itself, we will have to guard it carefully to ensure that it keeps us warm, rather than burning out of control.

Legislation will slowly appear that will help to curb some of these issues with AI tools, but in the meantime, if AI solutions are being deployed in a business, it may be worth adding an additional ethics gate to the development process to debate these risks before release.

For more information visit https://altronkarabina.com/




Share this article:
Share via emailShare via LinkedInPrint this page

Further reading:

The dream of Edge AI
Altron Arrow Editor's Choice AI & ML
AI technology carries a great promise – the idea that machines can make decisions based on the world around them, processing information like a human might. But the promise of AI is currently only being fulfilled by big machines.

Read more...
MAX78000 neural network accelerator chip
Altron Arrow AI & ML
The hardware-based convolutional neural network accelerator enables even battery-powered applications to execute AI inferences.

Read more...
Nanomaterials to build next-gen AI hardware?
AI & ML
From improving scientific analyses and imaging capabilities, to predictive maintenance and monitoring operations in industrial settings, artificial intelligence is becoming ever more present in modern-day society.

Read more...
Microchip launches MPLAB ML development suite
AI & ML
Microchip’s unique solution is first to support 8-, 16- and 32-bit MCUs and 32-MPUs for machine learning at the edge.

Read more...
ToF sensor enables AI applications
Altron Arrow AI & ML
The VL53L7CH from STMicroelectronics is the perfect Time-of-Flight sensor enabling AI applications, with ultrawide 90° diagonal FoV and low power consumption.

Read more...
Analogue compute platform to accelerate Edge AI
Altron Arrow Editor's Choice AI & ML
Microchip has teamed up with Intelligent Hardware Korea to develop an analogue compute platform to accelerate Edge AI/ML inferencing using Microchip’s memBrain non-volatile in-memory compute technology.

Read more...
World’s most powerful open LLM
AI & ML
With a staggering 180 billion parameters, and trained on 3,5 trillion tokens, Falcon 180B has soared to the top of the Hugging Face Leaderboard for pretrained LLMs.

Read more...
Advancing quality control
Avnet Silica AI & ML
As manufacturing processes continue to become more sophisticated, the importance and effectiveness of advanced DVI solutions escalate, presenting opportunities for improved quality control.

Read more...
Give your edge AI model a performance boost
AI & ML
Join this webinar from STMicroelectronics to learn how to create an edge AI application easily on an STM32 MCU using the NVIDIA TAO toolkit.

Read more...
Three reasons why AI, ML add value for SMMEs only if the basics are in place
AI & ML
There is much chatter around artificial intelligence (AI) and the subfield of machine learning (ML), which can be confusing for SMME owners who may believe that they need to climb on the bandwagon.

Read more...