Alluring and useful they may be, but the AI interfaces’ potential as gateways for fraud and intrusive data gathering is huge – and is only set to grow

Concerns about the growing abilities of chatbots trained on large language models, such as OpenAI’s GPT-4, Google’s Bard and Microsoft’s Bing Chat, are making headlines. Experts warn of their ability to spread misinformation on a monumental scale, as well as the existential risk their development may pose to humanity. As if this isn’t worrying enough, a third area of concern has opened up – illustrated by Italy’s recent ban of ChatGPT on privacy grounds.

The Italian data regulator has voiced concerns over the model used by ChatGPT owner OpenAI and announced it would investigate whether the firm had broken strict European data protection laws.

Continue reading…

Leave a Reply

Your email address will not be published. Required fields are marked *

You May Also Like

Gove gives reprieve to housebuilders that failed to sign up to safety scheme

Sister company to firm that refurbished Grenfell among builders that failed to…

Patients dying outside A&E, hospital corridors as makeshift wards – and it’s only October | Gaby Hinsliff

Rishi Sunak’s threats of ‘efficiencies’ crown a decade of Conservative neglect of…

Covert cameras and alleged hacking: how bust payments company Wirecard ‘hired spies and lawyers to silence critics’

Legal claim provides glimpse into world of corporate espionage and reputation management…

Resident Evil fans warned to be careful on social media as new game leaks

FANS are sending out a warning to others that footage of the…