News

‘Safety labels’ that clearly indicate AI risks and testing to be looked at: Josephine Teo

SINGAPORE - Users of generative artificial intelligence (AI) apps may soon see labels that clearly state how the AI should be used, its risks and how it was tested, as part of upcoming guidelines to make the technology easier to understand.

SINGAPORE - Users of generative artificial intelligence (AI) apps may soon see labels that clearly state how the AI should be used, its risks and how it was tested, as part of upcoming guidelines to make the technology easier to understand. Generative AI refers to AI that can create new content like text and images, and is less predictable than traditional AI. Likening the practice to safety labels on medication or household appliances, Minister for Digital Development and Information Josephine Teo said the effort aims to standardise how tech companies communicate transparency and testing.  Creators and deployers of generative AI should be clear with users on the data used, any risks and limitations of the model and how their systems have been tested, said Mrs Teo in a speech on July 15 to mark the start of Personal Data Protection Week at the Sands Expo and Convention Centre at Marina Bay Sands. The event ends on July 18. “We will recommend that developers and deployers be transparent with users by providing information on how the generative AI models and apps work,” said Mrs Teo, who is also Minister-in-charge of Smart Nation and Cybersecurity. Explaining the guidelines, Mrs Teo said: “This is a little bit like when you open a box of over-the-counter medication. There is always a sheet of paper to tell you about how the medication is supposed to be used and what are some of the side effects you may face. “This level of transparency is needed for AI models built using generative AI. That’s the recommendation.”  The guidelines will make clear safety benchmarks that should be tested before a form of AI is deployed, such as risks of spouting falsehoods, toxic statements and biased content. Mrs Teo added: “This is like when we buy household appliances and they come with a label that says it has been tested, because you cannot be expected to know whether the appliance is safe to use.”  The Infocomm Media Development Authority (IMDA) will start consultations with the industry on the guidelines, which it aims to roll out by early 2025 and to apply to all sectors, including healthcare and financial services, said a spokesman for IMDA. Separately, Mrs Teo said that in early 2025, businesses in Asean will have a guide on data anonymisation to help facilitate secure transfer of data across the region.  The guide is one of the outcomes of a meeting in February among officials in the region overseeing technology, who . IMDA also released a guide on privacy-enhancing technology in AI, which Mrs Teo said will help to address growing demands for more data to train AI without compromising on users’ privacy.  She referred to how the technology can help to protect personally identifiable information so that businesses can share data more securely.  Synthetic data, in particular, shows promise as a solution, as it creates realistic data for AI model training without using actual sensitive data, said Mrs Teo, who echoed experts’ concerns that AI innovation could lag because of a shortage of good data to train AI, due to privacy concerns, among others. The guide pointed to synthetic data as a solution, as it is modelled closely on real-world data and can help to speed up innovation while mitigating concerns about cyber-security incidents. Managing data in generative AI poses even more challenges for the industry compared with traditional AI, which is more predictable, said IMDA assistant chief executive and Personal Data Protection Commission deputy commissioner Denise Wong. She was speaking during a panel discussion on AI and data privacy next to representatives from tech organisations, including consulting firm Accenture and ChatGPT developer OpenAI. Ms Wong said: “The next challenge is – how do we do that in a generative AI space and what are the relevant principles? This is something the team and I will be discussing in consultation with the industry.” OpenAI’s head of privacy legal Jessica Gan Lee said data protection safeguards need to be developed at all stages of AI, from training and development to deployment. When asked about the risks of generative AI tools like ChatGPT, she said the key is to train AI models through diverse datasets “from all corners of the world”, incorporating multiple cultures, languages and sources, while finding ways to reduce processing of personal information.  Panellist Irene Liu, who is the regional strategy and consulting lead for finance, risk and compliance practice at Accenture, said: “A lot of the conversations are always centred around how organisations protect the data that they collect, but I feel a lot more focus has to be on the consumers themselves being responsible for the data that they provide.”  Not many are aware of the implications of sharing information online, such as when downloading programmes or when accepting cookies, she said, and suggested: “Can we make sure there is a level of education to consumers who are sharing this information, and understanding the implications of why they are sharing it as well?”

News Clusters