AMIGO-MA bids well for Biden
Context :
Top
tech firms commit to AI safeguards amid fears over pace of change …
Guardian…
Extract :
Top
players in the development of artificial intelligence, including Amazon, Google, Meta, Microsoft and OpenAI, have
agreed to new safeguards for the fast-moving technology, Joe Biden announced on
Friday.
Among the guidelines brokered by the Biden administration are watermarks for AI content to make it
easier to identify and third-party testing of the technology that will try to spot
dangerous flaws.
The
president said AI brings “incredible opportunities”, as well as risks to
society and economy. The agreement, he said, would underscore three fundamental
principles – safety, security and trust.
The White House said seven US companies had agreed to the voluntary
commitments, which are meant to ensure their AI products are safe before they release them.
The voluntary commitments are not legally binding, but may create a stopgap while more comprehensive action is developed.
A surge of commercial investment in generative AI tools that can
write convincingly human-like text and churn out new images and other media has
brought public fascination as well as concern about their ability to trick
people and spread disinformation, among other dangers.
The
tech companies agreed to eight measures:
·
Using watermarking on audio and visual content
to help identify content generated by AI.
·
Allowing independent experts to try to push models into bad behavior
– a process known as “red-teaming”.
·
Sharing trust and safety information with
the government and other companies.
·
Investing in cybersecurity
measures.
·
Encouraging third parties to uncover security
vulnerabilities.
·
Reporting societal risks such as inappropriate uses
and bias.
·
Prioritizing research on
AI’s societal risks.
·
Using the most cutting-edge AI systems, known as frontier
models, to solve society’s greatest
problems.
·
The voluntary commitments are meant to be an immediate way of addressing risks ahead
of a longer-term push to get Congress to pass laws regulating the technology.
·
The Senate majority leader, Chuck Schumer, has said he will introduce legislation to
regulate AI.
But
some experts and upstart competitors worry that the type of regulation being
floated could be a boon for deep-pocketed first-movers led by OpenAI, Google
and Microsoft, as smaller players are elbowed out by the high cost of making
their AI systems known as large language models adhere to regulatory
strictures.
The UN secretary general, António Guterres, recently said the United Nations was “the ideal place” to adopt global standards and
appointed a board that will report back on options for global AI governance by
the end of the year.
The
United Nations chief also said he welcomed calls from some countries for the creation
of a new UN body to support global efforts to govern AI, inspired by such models
as the International Atomic Energy Agency or the
Intergovernmental Panel on Climate Change.
The White House said on Friday that it had already consulted on
the voluntary commitments with a number of countries.
US-India
ties: President Joe Biden's Science Adviser calls for AI collaboration with
India… Mint… 22 July
Extract :
The Science Advisor
to President , Arati
Prabhakar [engagement@ostp.eop.gov.]
on Friday said that
the United States and like-minded countries including India need to work
together for
shaping the course
of artificial intelligence.
MY TAKE :
My
55 Blogs on Artificial Intelligence ( as of 22 July 2023 )
With regards,
Hemen Parekh
www.hemenparekh.ai / 23
July 2023
TRAI has suggested that an independent statutory authority, ‘Artificial Intelligence and Data Authority of India ,’ be established immediately for development of responsible AI and regulation of use cases in India
FOUR US leaders in artificial intelligence (AI) announced on Wednesday (Jul 26) the formation of an industry group devoted to addressing risks that cutting edge versions of the technology may pose.
Anthropic, Google, Microsoft, and ChatGPT-maker OpenAI said the newly created Frontier Model Forum will draw on the expertise of its members to minimise AI risks and support industry standards.
The companies pledged to share best practices with each other, lawmakers and researchers.
“Frontier” models refer to nascent, large-scale machine-learning platforms that take AI to new levels of sophistication – and also have capabilities that could be dangerous.
“Companies creating AI technology have a responsibility to ensure that it is safe, secure, and remains under human control,” Microsoft president Brad Smith said in a statement.
“This initiative is a vital step to bring the tech sector together in advancing AI responsibly and tackling the challenges so that it benefits all of humanity.”
No comments:
Post a Comment