AI Safety: OpenAI bares its efforts to ensure how its models work for humans
OpenAI, the creator of the chatbot ChatGPT, has publicly spoken about the safety of AI and how it tries to keep its products safe for its users. The company had come under criticism following privacy breaches and started approaching the problem by rapidly releasing new iterations of its models.
Last week, Italy became the first Western country to put a temporary ban on the use of ChatGPT, citing privacy concerns. Although OpenAI claimed that it was confident that it was complying with existing laws, the incident has raised the possibility of other countries following suit and restricting usage of the AI till its safety is assured.
OpenAI speaks about its approach to AI safety
A blog post on the company website is perhaps a way to assuage fears about AI models, which have made significant strides in the past few months. Open AI also seems to have been trying to address the points made by U.S. President Joe Biden recently, where he put the onus on the product's safety on the companies building it.
In the post, Open AI states that it rigorously tests any new system before it is introduced to the public using external experts and uses reinforcement learning with human feedback to make improvements. The company claims that it tested its recent model, GPT-4, for six months before it was released publicly and called for regulation to ensure that the industry adopts such practices at large.

Realizing the limits of the lab testing conditions, OpenAI said that it could not predict how users might deploy its models and therefore takes a cautious and gradual approach to new AI systems.
By releasing its latest and most capable models through its services or an Application Programming Interface (API), OpenAI says it can monitor the misuse of services and take action immediately based on real-world data.
This has enabled OpenAI to develop nuanced policies against genuine risks from its technology while still allowing people to use it for many beneficial purposes. The company also said it is evaluating verification options to ensure that users above 18 or those above 13 years and with parental approval are accessing its services.
Open AI said that its newer model, GPT-4, is 82 percent less likely to respond to requests for hateful, harassing, and violent content. The company has put safeguards to prevent content such as Child Sexual Abuse Material (CSAM) from being uploaded to its platform.
Stating that OpenAI's business model is not based on building profiles of people and selling advertising to them, the company works to remove personal information from its training data sets. It is fine-tuning its models to reject requests for the personal information of private individuals.
Additionally, the company is also making progress in ensuring that the content provided by its chatbot is factually correct. GPT-4 is 40 percent more likely to produce factual content than its predecessor, the blog post said while adding that there was much work to do in this area to make AI safer for all.