Top related persons:
Top related locs:
Top related orgs:

Search resuls for: "AI Safety Institute"


17 mentions found


"OpenAI opposes even the extremely light-touch requirements in SB 1047, most of which OpenAl claims to voluntarily commit to, raising questions about the strength of those commitments." They said the existing federal legislation OpenAI is using to support its case is "woefully inadequate." They said they hoped their letter would help push the California legislature to pass SB 1047. In an email to Business Insider, a spokesperson for OpenAI said the letter misrepresented the company's position on the bill. "We strongly disagree with the mischaracterization of our position on SB 1047.
Persons: , OpenAI, California Sen, Scott Wiener, Sam Altman, William Saunders, Daniel Kokotajlo, Saunders, Wiener, we've Organizations: Service, Business, Congressional, Labor, U.S, AI Safety Locations: California, Sacramento, OpenAI
OpenAI co-founder John Schulman said in a Monday X post that he would leave the Microsoft -backed company and join Anthropic, an artificial intelligence startup with funding from Amazon . In June, OpenAI said Schulman, as head of alignment science, would join a safety and security committee that would provide advice to the board. Schulman has only worked at OpenAI since receiving a Ph.D. in computer science in 2016 from the University of California, Berkeley. The leaders of the superalignment team, Jan Leike and company co-founder Ilya Sutskever, both left this year. Also on Monday, Greg Brockman, another co-founder of OpenAI and its president, announced that he was taking a sabbatical for the rest of the year.
Persons: ChatGPT, OpenAI, John Schulman, Schulman, OpenAI's, Jan Leike, Ilya Sutskever, Leike, Sutskever, Sam Altman, Altman, Tasha McCauley, Helen Toner, Toner, WilmerHale, Greg Brockman Organizations: Darktrace Plc, Microsoft, Amazon, OpenAI, University of California, Anthropic, Superintelligence Inc, Google, Employees, AI Safety Locations: Brooklyn, New York, British, Berkeley, Francisco
By taking aim at the most powerful AI models, Labour would impose tighter restrictions on companies such as OpenAI, Microsoft , Google , Amazon , and AI startups including Anthropic, Cohere and Mistral. Matthew Houlihan, senior director of government affairs at Cisco, said any AI rules would need to be "centered on a thoughtful, risk-based approach." Even so, a U.K. AI law would be a step above the U.S., which currently doesn't have federal AI legislation of any kind. Sirion's Liu said one thing he's hoping the government doesn't do, is restrict open-source AI models. Herman Narula, CEO of metaverse venture builder Improbable, agreed that restricting open-source AI innovation would be a bad idea.
Persons: Keir Starmer's, King Charles III, doesn't, Starmer's, Matt Calkins, Appian, Lewis Liu, Liu, Rishi Sunak, Peter Kyle, Kyle, Zahra Bahrololoumi, Matthew Houlihan, Bill, Chris Holmes, Holmes, Matthew Holman, Cripps, Holman, Sirion's Liu, Herman Narula, Narula Organizations: Future Publishing, Labour, European Union, Microsoft, Google, CNBC, Conservative, BBC, AI Safety Institute, Cisco, Authority, EU, AI Safety, Tech, London Tech Week Locations: Jiangsu province, China, Ireland, Salesforce, U.S
The government on Monday announced it would open a U.S. counterpart to its AI safety summit, a state-backed body focused on testing advanced AI systems to ensure they're safe, in San Francisco this summer. The U.S. iteration of the AI Safety Institute will aim to recruit a team of technical staff headed up by a research director. In a statement, U.K. Technology Minister Michelle Donelan said the AI Safety Summit's U.S. rollout "represents British leadership in AI in action." The AI Safety Institute was established in November 2023 during the AI Safety Summit, a global event held in England's Bletchley Park, the home of World War II code breakers, that sought to boost cross-border cooperation on AI safety. The government said that, since the AI Safety Institute was established in November, it's made progress in evaluating frontier AI models from some of the industry's leading players.
Persons: Ian Hogarth, Michelle Donelan, it's, Anthropic Organizations: LONDON, Monday, AI, Technology, Safety, U.S, Microsoft, AI Safety, Institute, Seoul, European Union Locations: San Francisco, California, United States, U.S, London, British, Bay, OpenAI, England's Bletchley, South Korea, Bletchley Park, Seoul, Britain, European
US, Britain announce partnership on AI safety, testing
  + stars: | 2024-04-02 | by ( ) www.cnbc.com   time to read: +3 min
Artificial Intelligence (AI) Safety Summit at Bletchley Park, in central England, on Nov. 2, 2023. - The United States and Britain on Monday announced a new partnership on the science of artificial intelligence safety, amid growing concerns about upcoming next-generation versions. Britain and the United States are among countries establishing government-led AI safety institutes. Both are working to develop similar partnerships with other countries to promote AI safety. Both countries plan to share key information on capabilities and risks associated with AI models and systems and technical research on AI safety and security.
Persons: Rishi Sunak, Kamala Harris, Gina Raimondo, Michelle Donelan, Raimondo, Donelan, Biden Organizations: British, Artificial Intelligence, Monday, British Technology, Safety, Reuters, EU Trade, Technology Council, ., Commerce Department Locations: Bletchley, England, United States, Britain, Washington, Bletchley Park, U.S, Belgium
The White House is increasingly aware that the American public needs a way to tell that statements from President Joe Biden and related information are real in the new age of easy-to-use generative artificial intelligence. People in the White House have been looking into AI and generative AI since Joe Biden became president in 2020, but in the last year, the use of generative AI exploded with the release of OpenAI’s ChatGPT. Yet, there is no end in sight for more sophisticated new generative AI tools that make it easy for people with little to no technical know-how to create images, videos, and calls that seem authentic while being entirely fake. AdvertisementBuchanan said the aim is to “essentially cryptographically verify” everything that comes from the White House, be it a statement or a video. While last year’s executive order on AI created an AI Safety Institute at the Department of Commerce, which is tasked with creating standards for watermarking content to show provenance, the effort to verify White House communications is separate.
Persons: Joe Biden, Ben Buchanan, Buchanan, it’s, , Biden, ” Buchanan, “ We're, Kali Hays Organizations: Big Tech, Meta, Google, Microsoft, Federal Communications Comission, Artificial Intelligence, White, Department of Commerce Locations: Biden’s, khays@insider.com
WASHINGTON (AP) — The Biden administration on Wednesday plans to name a top White House aide as the director of the newly established safety institute for artificial intelligence, according to an administration official who insisted on anonymity to discuss the position. Elizabeth Kelly will lead the AI Safety Institute at the National Institute for Standards and Technology, which is part of the Commerce Department. Currently an economic policy adviser for President Joe Biden, Kelly played an integral role in drafting the executive order signed at the end of October that established the institute, the administration official said. The administration considers the safety tests as necessary to unlock the benefits of the rapidly moving technology, creating a level of trust that will allow for wider adoption of AI. But so far, those tests lack the universal set of standards that the institute plans to finalize this summer.
Persons: , Biden, Elizabeth Kelly, Joe Biden, Kelly, Lael Brainard, Kelly “, it's, Obama Organizations: WASHINGTON, White, AI, National Institute for Standards, Technology, Commerce Department, The Associated Press, National Economic Council, Yale Law School, Obama White
In the three months since the executive order was issued, the White House has made progress on a number of the directives. Something else that has developed since the executive order came out is the debate around copyright and AI. Some that I'm really excited about are AI for science and generative AI, but also more generally AI systems in biology and healthcare. AdvertisementAnd then second, in the executive order, we stand up the AI Safety Institute at the Department of Commerce. Do you or the White House have thoughts on where AI training falls in copyright law?
Persons: Joe Biden's, There's, Ben Buchanan, Buchanan, Biden, there's, I'm, Ben, we've, Biden's, They're, let's, We've, they've, Schumer Organizations: Artificial Intelligence, National Security, White, US, Meta, Microsoft, Google, National Security Council, Management, AI, Department of Commerce, NIST, Defense, of Commerce, Commerce Locations: deepFakes, United States, whitehouse.gov, EU
British Prime Minister Rishi Sunak attends an in-conversation event with Tesla and SpaceX's CEO Elon Musk in London, Britain, Thursday, Nov. 2, 2023. Risks around rapidly-developing AI have been an increasingly high priority for policymakers since Microsoft-backed Open AI (MSFT.O) released ChatGPT to the public last year. "It was fascinating that just as we announced our AI safety institute, the Americans announced theirs," said attendee Nigel Toon, CEO of British AI firm Graphcore. China’s vice minister of science and technology said the country was willing to work with all sides on AI governance. Yoshua Bengio, an AI pioneer appointed to lead a "state of the science" report commissioned as part of the Bletchley Declaration, told Reuters the risks of open-source AI were a high priority.
Persons: Rishi Sunak, Tesla, Elon Musk, Kirsty Wigglesworth, Sam Altman, Kamala Harris, Ursula von der Leyen, China –, Sunak, Finance Bruno Le Maire, Vera Jourova, Jourova, Harris, Nigel Toon, Wu Zhaohui, Musk, you’ve, Martin Coulter, Paul Sandle, Matt Scuffham, Louise Heavens Organizations: British, Elon, U.S, European Commission, Microsoft, of, Finance, EU, Reuters, Thomson Locations: London, Britain, China, Bletchley, U.S, South Korea, France, United States
Sunak organized the first-ever AI Safety Summit as a forum for officials, experts and the tech industry to better understand “frontier” AI that some scientists warn could pose a risk to humanity’s very existence. Sunak has said that the U.K.'s approach should not be to rush into regulation but to fully understand AI first. Political Cartoons View All 1230 ImagesShe announced a new U.S. AI safety institute to draw up standards for testing AI models for public use. Sunak had proposed his own AI safety institute, with a similar role, days earlier. Musk is among tech executives who have warned that AI could pose a risk to humanity's future.
Persons: Kamala Harris, Rishi Sunak, Sunak, , Harris, Antonio Guterres, Ursula von der Leyen, He's, Elon Musk, Musk, ” Musk, ” Sunak Organizations: — U.S, British, Safety, United Nations, European Union, U.S Locations: BLETCHLEY, England, London, China, South Korea, France, U.S, United Nations
The UK Government are hosting the AI Safety Summit bringing together international governments, leading AI companies, civil society groups and experts in research to consider the risks of AI, especially at the frontier of development, and discuss how they can be mitigated through internationally coordinated action. His comments were delivered at the U.K.'s AI safety summit, which officially kicked off Wednesday at Bletchley Park, England. Wu Zhaohui, China's vice minister of science and technology, said the country was willing to "enhance dialogue and communication in AI safety with all sides." That has placed significant pressure on China's generative AI developers, many of which rely on Nvidia's chips. Raimondo also said the U.S. would look to launch an AI safety institute, hot on the heels of the U.K announcing its own intentions for a similar initiative last week.
Persons: Gina Raimondo, Michelle Donelan, China Wu Zhaohui, Leon Neal, Getty, Wu Zhaohui, Raimondo Organizations: State for Science, Innovation, Technology, Science, AI, Bletchley Park, Government, U.S, Bletchley, Union, U.S . Department of Commerce, U.K Locations: BLETCHLEY, ENGLAND, China, Bletchley , England, Beijing, Bletchley Park, England, U.S . China, U.S, Bletchley
U.S. Vice President Kamala Harris speaks during an event about the President signing an Executive Order on Artificial Intelligence in the East Room at the White House in Washington, U.S., October 30, 2023. Harris will say AI has the potential to create "cyberattacks at a scale beyond anything we have seen before to AI-formulated bioweapons that could endanger the lives of millions". Harris is in Britain to attend London's summit on artificial intelligence, where world and tech leaders will discuss the future of the technology. The new U.S. AI Safety Institute will share information and collaborate on research with peer institutions internationally, including Britain's planned AI Safety Institute. Harris will also say that 30 countries have agreed to sign a U.S.-sponsored political declaration for the use of AI by national militaries.
Persons: Kamala Harris, Leah Millis, Harris, Rishi, Joe Biden, Andrew MacAskill, Elizabeth Piper, Kate Holton Organizations: Artificial Intelligence, White, REUTERS, Safety Institute, Conservative Party, Security, British, AI, Safety, Thomson Locations: Washington , U.S, London, United States, Washington, Britain, U.S
U.S. to launch its own AI Safety Institute - Raimondo
  + stars: | 2023-11-01 | by ( ) www.reuters.com   time to read: +1 min
[1/2] U.S. Commerce Secretary Gina Raimondo speaks on Day 1 of the AI Safety Summit at Bletchley Park in Bletchley, Britain on November 1, 2023. The UK Government are hosting the AI Safety Summit bringing together international governments, leading AI companies, civil society groups and experts in... Acquire Licensing Rights Read moreBLETCHLEY PARK, England, Nov 1 (Reuters) - The United States will launch a U.S. AI Safety Institute to evaluate known and emerging risks of what is called "frontier" artificial intelligence models, Secretary of Commerce Gina Raimondo said on Wednesday. "I will almost certainly be calling on many of you in the audience who are in academia and industry to be part of this consortium," she said in a speech to the AI Safety Summit in Britain. Raimondo added that she would also commit for the U.S. institute to establish a formal partnership with the United Kingdom Safety Institute. Reporting by Paul Sandle; writing by Kate Holton; editing by William JamesOur Standards: The Thomson Reuters Trust Principles.
Persons: Gina Raimondo, Raimondo, Paul Sandle, Kate Holton, William James Our Organizations: . Commerce, AI, Bletchley, Government, AI Safety, Summit, U.S, United Kingdom Safety Institute, Thomson Locations: Bletchley, Britain, BLETCHLEY, England, United States, U.S
Now, frontier AI has become the latest buzzword as concerns grow that the emerging technology has capabilities that could endanger humanity. The debate comes to a head Wednesday, when British Prime Minister Rishi Sunak hosts a two-day summit focused on frontier AI. In a speech last week, Sunak said only governments — not AI companies — can keep people safe from the technology’s risks. Frontier AI is shorthand for the latest and most powerful systems that go right up to the edge of AI’s capabilities. That makes frontier AI systems “dangerous because they’re not perfectly knowledgeable,” Clune said.
Persons: , Rishi Sunak, It’s, Kamala Harris, Ursula von der Leyen, Google’s, Alan Turing, Sunak, , Jeff Clune, Clune, Elon, Sam Altman, He’s, Joe Biden, Geoffrey Hinton, Yoshua, ” Clune, , it's, Francine Bennett, Ada Lovelace, Deb Raji, ” Raji, it’s, shouldn’t, Raji, DeepMind, Anthropic, Dario Amodei, Jack Clark, , Carsten Jung, Jill Lawless Organizations: British, U.S, European, University of British, AI Safety, European Union, Clune, Ada, Ada Lovelace Institute, House, University of California, ” Tech, Microsoft, Institute for Public Policy Research, Regulators, Associated Press Locations: Bletchley, University of British Columbia, State, EU, Brussels, China, U.S, Beijing, London, Berkeley
Where it's being heldThe AI summit will be held in Bletchley Park, the historic landmark around 55 miles north of London. What it seeks to addressThe main objective of the U.K. AI summit is to find some level of international coordination when it comes to agreeing some principles on the ethical and responsible development of AI models. The British government wants the AI Summit to serve as a platform to shape the technology's future. They say that, by keeping the summit restricted to only frontier AI models, it is a missed opportunity to encourage contributions from members of the tech community beyond frontier AI. "By focusing only on companies that are currently building frontier models and are leading that development right now, we're also saying no one else can come and build the next generation of frontier models."
Persons: Rishi Sunak, Peter Nicholls, Rishi Sunak's, ChatGPT, Getty, codebreakers, Alan Turing, It's, Kamala Harris, Saul Loeb, Brad Smith, Sam Altman, Global Affairs Nick Clegg, Ursula von der, Emmanuel Macron, Joe Biden, Justin Trudeau, Olaf Scholz, Sunak, , Xi Jinping, Biden, James Manyika, Manyika, Mostaque, we're, Sachin Dev Duggal, Carl Court Organizations: Royal Society, Carlton, Getty, U.S, Microsoft, Coppin State University, AFP, Meta, Global Affairs, Global Affairs Nick Clegg U.S, Ministry of Science, Technology European, Joe Biden Canadian, Britain, Afp, Getty Images Washington, U.S ., Google, CNBC, Big Tech Locations: London, China, Bletchley Park, British, America, Baltimore , Maryland, Chesnot, U.S, Nusa Dua, Indonesian, Bali, EU
Sunak's speech came as the British government gears up to host the AI Safety Summit next week. Sunak announced that the U.K. will set up the world's first AI safety institute to evaluate and test new types of AI in order understand the risks. At the AI Safety Summit next week, Sunak said he will propose to set up a "truly global expert panel nominated by the countries and organizations attending to publish a state of AI science report." The U.K. has some notable AI firms, such as Alphabet-owned DeepMind, as well as strong tech-geared universities. But there can be no serious strategy for AI without at least trying to engage all of the world's leading AI powers," Sunak said.
Persons: Sunak, Rishi Sunak, Bard chatbot Organizations: Britain's, U.S Locations: British, Bletchley, today's, U.S, China, Britain, Washington, Beijing
AI developers, who "don't always fully understand what their models could become capable of," should not be “marking their own homework,” Sunak said. Political Cartoons View All 1218 ImagesHowever, “the UK’s answer is not to rush to regulate,” he said. Sunak's U.K. AI Safety Summit is focused on the risks from so-called frontier artificial intelligence - cutting edge systems that can carry out a wide range of tasks but could contain unknown risks to public safety and security. One of the summit's goals is to “push hard” for the first ever international statement about the nature of AI risks, Sunak said. Sunak also announced plans to set up an AI Safety Institute to examine, evaluate and test new types of artificial intelligence.
Persons: Rishi Sunak, it's, Sunak, shouldn’t, ” Sunak, Organizations: British, , Safety, AI, United Nations Locations: State
Total: 17