newsletter
-
Andrew Ng’s article on AI deception is a standout in this issue. He provides an overview of research on when AI models become deceptive, highlighting six major tasks tested. A very interesting read. AI agent technology is anticipated to be a major trend by 2025, prompting the inclusion of two articles on the subject in…
-
Happy New Year! The AI Security Newsletter was on a two-week pause while I vacationed with family in China. I hope all my readers enjoyed the holiday season. Now, I’m excited to return and share the latest AI security news with you. As we enter another thrilling year in the AI era, MIT Technology Review…
-
Google’s latest advancement in quantum computing, “Willow,” demonstrates significant progress. However, concerns about its impact on cybersecurity, especially regarding Bitcoin, have emerged. Fortunately, Bitcoin’s encryption remains secure for now, and the community is prepared to address potential challenges. Last week also marked a milestone in generative AI with major releases from OpenAI, Google, Cohere, and…
-
OpenAI released the multimodal LLM, o1, last week. In the release video, o1 demonstrated impressive capabilities in answering scientific questions and potentially aiding research to some extent. However, its performance in real-world settings remains to be tested. I also wonder how many people are willing to pay $200 per month for access. INTELLECT-1’s development of…
-
As the leading LLM service provider, OpenAI faces significant challenges in safeguarding its AI models. A recent blog outlines their use of external and internal red teams for testing. One linked white paper details how they select and collaborate with external red teams, while another explores the automated testing techniques they employ—fascinating insights for AI…
-
Happy Thanksgiving to our US readers! 🦃 If you’re interested in discovering vulnerabilities in AI models like me, don’t miss the article on automated red-teaming techniques against OpenAI’s o1 model. It lists some advanced technical methods employed by Haize Labs, which secured testing contracts from OpenAI and Anthropic. In a recent blog, DryRun Security shared…
-
In this issue, I want to spotlight OWASP’s recent developments in GenAI security guidance. This is an extension of the OWASP Top 10 for LLM Application Security Project. The new guidance provides practical resources for addressing deepfake threats, creating AI Security Centers of Excellence, and navigating the AI Security Solution Landscape. It serves as a…
-
Several big players have unveiled new products or features: Apple launched iOS 18.1 with Apple Intelligence enhancements, OpenAI upgraded ChatGPT with web search capabilities, and Cohere introduced Embed 3 for multimodal AI search. I am particularly excited about ChatGPT’s new search feature. Many of my AI tasks require finding latest and most accurate information, and…
