(Reuters) — Rapid advances in artificial intelligence (AI) such as Microsoft-backed OpenAI's ChatGPT are complicating governments' efforts to agree laws governing the use of the technology.
Here are the latest steps national and international governing bodies are taking to regulate AI tools:
AUSTRALIA
* Planning regulations
Australia will make search engines draft new codes to prevent the sharing of child sexual abuse material created by AI and the production of deepfake versions of the same material, the country's internet regulator said on Sept. 8.
BRITAIN
* Planning regulations
Britain's data watchdog said on Oct. 10 it had issued Snap Inc (NYSE:SNAP)'s Snapchat with a preliminary enforcement notice over a possible failure to properly assess the privacy risks of its generative AI chatbot to users, particularly children.
The country's competition authority set out seven principles on Sept. 18 designed to make developers accountable, prevent Big Tech tying up the tech in their walled platforms, and stop anti-competitive conduct like bundling.
The proposed principles, which come six weeks before Britain hosts a global AI safety summit, will underpin its approach to AI when it assumes new powers in the coming months to oversee digital markets.
CHINA
* Implemented temporary regulations
China issued a set of temporary measures effective Aug. 15, requiring service providers to submit security assessments and receive clearance before releasing mass-market AI products.
EUROPEAN UNION
* Planning regulations
EU lawmaker Brando Benifei, who is leading negotiations on the bloc's AI Act, on Sep. 21 urged member countries to compromise in key areas to reach an agreement by the end of the year. Lawmakers are thrashing out details
Read more on investing.com