ChatGPT : the Sacrificial Goat ?
During the past few weeks, a few countries have either already imposed some restrictions on use of ChatGPT or are planning to do so
Reason ?
They think ChatGPT is “ potentially “ dangerous considering that , if left to “ evolve “ without any constraints, the following scenario is highly probable :
Ø AIs cloning / re-producing their own ( better or worse ) versions without any prompt from a human
Ø AIs engaging in conversations / chats among themselves without human intermediation
Ø AIs acquiring “ human frailties “ but failing to acquire “ human wisdom “
Ø AIs setting for themselves “ Goals / Targets / Tasks “ which cause harm to humans
Some 1000 + geeks / scientists have proposed a 6 month pause on development of even more powerful AI
Sure, a few countries can “ ban “ use of one ChatGPT , coming out of one country, USA
But can anyone ban / regulate , some 10,000 ChatGPT equivalent AI, coming out of a 100 countries in next 6 months , some of these having characteristics described above ?
Here are a few which have sprung up within past few weeks :
I am not against the idea of “ REGULATING “ all , current and future AI
In fact, I strongly believe there is an URGEND NEED for such a regulation , which is evolved through a CONSENSUS among all the stakeholders and implemented / regulated / enforced through a UN regulatory body ( ala SECURITY COUNCIL )
I urge Shri Ashwini vaishnaw , IT Minister ( India ) to take a lead in evolving such a consensus, by circulating among the stakeholders ( with modifications
deemed necessary ) my following suggestion :
Ø Parekh’s Law of Chatbots …….. 25 Feb 2023
With regards,
Hemen Parekh
Related Readings :
US begins study of possible rules to regulate AI like ChatGPT ……….. Reuters / 12 Apr 2023
Extract :
The Biden administration said Tuesday it is seeking public comments on potential accountability measures for artificial intelligence (AI) systems as questions loom about its impact on national security and education.
ChatGPT, an AI program that recently grabbed the public's attention for its ability to write answers quickly to a wide range of queries, in particular has attracted U.S. lawmakers' attention as it has grown to be the fastest-growing consumer application in history with more than 100 million monthly active users.
The National Telecommunications and Information Administration, a Commerce Department agency that advises the White House on telecommunications and information policy, wants input as there is "growing regulatory interest" in an AI "accountability mechanism."
The agency wants to know if there are measures that could be put in place to provide assurance "that AI systems are legal, effective, ethical, safe, and otherwise trustworthy."
“Responsible AI systems could bring enormous benefits, but only if we address their potential consequences and harms. For these systems to reach their full potential, companies and consumers need to be able to trust them,” said NTIA Administrator Alan Davidson.
President Joe Biden last week said it remained to be seen whether AI is dangerous. "Tech companies have a responsibility, in my view, to make sure their products are safe before making them public," he said.
ChatGPT, which has wowed some users with quick responses to questions and caused distress for others with inaccuracies, is made by California-based OpenAI and backed by Microsoft Corp (MSFT.O).
NTIA plans to draft a report as it looks at "efforts to ensure AI systems work as claimed – and without causing harm" and said the effort will inform the Biden Administration's ongoing work to "ensure a cohesive and comprehensive federal government approach to AI-related risks and opportunities."
A tech ethics group, the Center for Artificial Intelligence and Digital Policy, asked the U.S. Federal Trade Commission to stop OpenAI from issuing new commercial releases of GPT-4 saying it was "biased, deceptive, and a risk to privacy and public safety."
China's New Draft Law Mandates "Security Assessement" For AI Products NDTV / 11 Apr 2023
Extract :
New AI products developed in China will have to undergo a security assessment before being released and must reflect "core socialist values", a sweeping new draft law by the country's internet regulator showed Tuesday.
The fresh regulations come as a flurry of Chinese companies rush to develop artificial intelligence services that can mimic human speech since San Francisco-based OpenAI launched ChatGPT in November, sparking a gold rush in the market.
Rapid advancements in AI have stoked global alarm over the technology's potential for disinformation and misuse, with deepfake images and people shown mouthing things they never said.
"Before providing services to the public that use generative AI products, a security assessment shall be applied for through national internet regulatory departments," the draft law, released by the Cyberspace Administration of China, reads.
The draft law -- dubbed "Administrative Measures for Generative Artificial Intelligence Services" -- aims to ensure "the healthy development and standardised application of generative AI technology", it read.
AI-generated content, it continued, needs to "reflect core socialist values, and must not contain content on subversion of state power".
It must also not contain, among other things, "terrorist or extremist propaganda", "ethnic hatred" or "other content that may disrupt economic and social order."
The Cyberspace Administration of China said it was seeking public input on the contents of the new regulations, which under Beijing's highly centralised political system are almost certain to become law.
"The new CAC draft document is one of the strictest measures for generative AI so far," Andy Chun, adjunct professor at City University of Hong Kong, told AFP.
Companies submitting security assessments will need to "be very careful to ensure each data source used for AI learning must be within guidelines, accurate, unbiased, and not infringe on IP rights of others," he said.
"Ensuring accuracy is hard. No generative AI system to date can do that," said Chun.
The regulatory crackdown comes as China's tech giants ramp up their efforts in the closely-watched sector.
Alibaba's cloud computing unit on Tuesday unveiled its own product called Tongyi Qianwen, which is expected to be rolled out across the tech giant's office workplace communications software and household appliances.
CEO Daniel Zhang said in a statement that the software came in a "technological watershed moment driven by generative AI and cloud computing".
And Baidu -- which operates the country's leading search engine -- released its own "Ernie Bot" AI chat product last month.
But investors were unimpressed by the bot's display of linguistic and maths skills at an unveiling, sending shares falling by as much as 10 percent.
ChatGPT is unavailable in China, but the American software is also gaining a base of Chinese users who use virtual private networks to get around the ban, deploying it to write essays and cram for exams.
And a 24-year-old Shanghai blogger caused a stir this month when he used AI technology to "resurrect" his dead grandmother, producing lifelike imagery of his interactions with the dead relative.
Beijing has announced ambitious plans to become a global leader in the field of AI by 2030, and consultancy group McKinsey estimates the sector could add about $600 billion every year to China's gross domestic product by then.
But it has also warned that deepfakes present a "danger to national security and social stability".
Beijing in January enforced new rules that would require businesses offering deepfake services to obtain the real identities of their users. They also require deepfake content to be appropriately tagged to avoid "any confusion".
UK
5 Core Principles of AI Ethics
Extract :
A key recommendation from the report calls for a cross-sector AI code to be formed, a code that a could be adopted around the globe.
“An ethical approach ensures the public trusts this technology and sees the benefits of using it. It will also prepare them to challenge its misuse,” writes Lord Tim Clement-Jones, the chairman of the House of Lords Select Committee on AI that commissioned the UK report.
The report includes 5 Core Principles:
• AI should be developed for the common good and benefit of humanity.
• AI should operate on principles of intelligibility and fairness.
• AI should not be used to diminish the data rights or privacy of individuals, families or communities.
• All citizens should have the right to be educated to enable them to flourish mentally, emotionally and economically alongside artificial intelligence.
• The autonomous power to hurt, destroy or deceive human beings should never be vested in AI.
Italy became the first Western country to ban ChatGPT. Here’s what other countries are doing
Extract :
· Italy last week became the first Western country to ban ChatGPT, the popular AI chatbot.
· ChatGPT has both impressed researchers with its capabilities while also worrying regulators and ethicists about the negative implications for society.
· The move has highlighted an absence of any concrete regulations, with the European Union and China among the few jurisdictions developing tailored rules for AI.
· Various governments are exploring how to regulate AI, and some are thinking of how to deal with general purpose systems such as ChatGPT.
https://medium.com/generative-
INDIA
Thank You, Ashwini Vaishnawji
No comments:
Post a Comment