Alluring and useful they may be, but the AI interfaces’ potential as gateways for fraud and intrusive data gathering is huge – and is only set to grow

Concerns about the growing abilities of chatbots trained on large language models, such as OpenAI’s GPT-4, Google’s Bard and Microsoft’s Bing Chat, are making headlines. Experts warn of their ability to spread misinformation on a monumental scale, as well as the existential risk their development may pose to humanity. As if this isn’t worrying enough, a third area of concern has opened up – illustrated by Italy’s recent ban of ChatGPT on privacy grounds.

The Italian data regulator has voiced concerns over the model used by ChatGPT owner OpenAI and announced it would investigate whether the firm had broken strict European data protection laws.

Continue reading…

Leave a Reply

Your email address will not be published. Required fields are marked *

You May Also Like

Novelist Elif Shafak: ‘I’ve always believed in inherited pain’

The award-winning Turkish-British writer, whose new book explores love and politics in…

Prison officer job ad banned over ‘negative racial stereotype’

ASA rules that UK Facebook ad showing white officer with black prisoner…

Labour criticises cuts after leaked MoD report says army low on troops

Opposition says ‘a proper defence strategy’ is required after ‘a decade of…