A.M. Turing Award Recipient and most-cited AI researcher.
https://lawzero.org/en
https://yoshuabengio.org/profile/
· When given 10 attempts, attackers can use malicious prompts to bypass leading systems' safeguards about half the time.
· Inserting as few as 250 malicious documents into a model's training data can introduce vulnerabilities.
(6/6)
· When given 10 attempts, attackers can use malicious prompts to bypass leading systems' safeguards about half the time.
· Inserting as few as 250 malicious documents into a model's training data can introduce vulnerabilities.
(6/6)
(5/6)
(5/6)
· Technical safeguards are beginning to inform transparency measures in governance frameworks.
(4/6)
· Technical safeguards are beginning to inform transparency measures in governance frameworks.
(4/6)
· Improvements in adversarial training methods to make models more resistant to potentially harmful requests,
· Better tools for tracking AI-generated content.
(3/6)
· Improvements in adversarial training methods to make models more resistant to potentially harmful requests,
· Better tools for tracking AI-generated content.
(3/6)
Read the full Update here: internationalaisafetyreport.org/publication/...
(2/6)
Read the full Update here: internationalaisafetyreport.org/publication/...
(2/6)
Innovation & regulation must go hand in hand, notably by building technical safeguards to make AI systems more trustworthy for individuals and businesses. That's at the heart of our work at @law-zero.bsky.social.
Innovation & regulation must go hand in hand, notably by building technical safeguards to make AI systems more trustworthy for individuals and businesses. That's at the heart of our work at @law-zero.bsky.social.