Rishi Iyengar
It sounds like the beginning of a nightmare scenario that artificial intelligence doomsayers have been warning about: This month, Silicon Valley AI company Anthropic said it had developed a model so dangerous that the company had decided against releasing it to the public.
The model, known as Claude Mythos Preview, is a general-purpose language model like Anthropic’s Claude or OpenAI’s ChatGPT. But during testing, it showed an ability to find and exploit so-called “zero day” vulnerabilities—an industry term that refers to previously undiscovered holes in a system’s software. The model “could reshape cybersecurity” because it found “thousands of high-severity vulnerabilities” in “every major operating system and web browser,” Anthropic said. It made those claims in a blog post announcing that it would open up Mythos only to a few dozen companies and critical infrastructure operators. That collective, which Anthropic named Project Glasswing, includes Amazon Web Services, Apple, Google, JPMorganChase, Microsoft, and Nvidia as companies that will receive early access to the model to patch vulnerabilities in their systems.
No comments:
Post a Comment