ASIC joins global regulators monitoring Anthropic’s Mythos AI


Australia’s markets regulator has publicly confirmed it is watching the development of Anthropic’s Mythos model alongside peer regulators worldwide, adding to a rapidly expanding international regulatory response that began with the Bank of England, the US Federal Reserve, and the Treasury Department. ECB President Lagarde has warned no governance framework is yet in place.


The Australian Securities and Investments Commission (ASIC) confirmed on Monday that it is monitoring the development of Anthropic’s frontier AI model Mythos and its potential implications for the Australian financial market, Reuters reported.

“ASIC is closely monitoring these developments along with peer regulators to assess possible implications for the Australian market,” an ASIC spokesperson said.

“ASIC engages closely with other regulators, government agencies and the financial sector to understand and respond to changing technologies.”

The 💜 of EU tech

The latest rumblings from the EU tech scene, a story from our wise ol’ founder Boris, and some questionable AI art. It’s free, every week, in your inbox. Sign up now!

The regulator added that it expected financial services licensees to “be on the front foot” to safeguard their customers and clients.

The ASIC statement is the latest in a cascade of global regulatory responses to Mythos, the advanced AI model that Anthropic launched on 7 April 2026 under a restricted access programme called Project Glasswing.

Anthropic claimed the model successfully identified and exploited zero-day vulnerabilities in every major operating system and web browser, a capability the company says is intended to accelerate defensive security work but which regulators have identified as a potential systemic risk if threat actors accessed the model’s capabilities.

The response from financial regulators has been rapid and unusually coordinated for a technology event. Bank of England Governor Andrew Bailey, speaking at Columbia University in New York, warned that Mythos could “crack the whole cyber risk world open” and called on regulators to urgently assess the extent to which the model can identify and exploit vulnerabilities in financial infrastructure.

The Bank of England’s Cross Market Operational Resilience Group (CMORG) and its AI Taskforce subsequently scheduled meetings to discuss Mythos within weeks. European Central Bank President Christine Lagarde told Bloomberg TV that there is currently no governance framework “to actually mind those things”, a frank admission that the regulatory infrastructure has not kept pace with the technology.

In the United States, Treasury Secretary Scott Bessent and Federal Reserve Chair Jerome Powell convened an urgent meeting of major bank CEOs to discuss Mythos’s cyber risk implications.

The meeting, held while bank chiefs were already in Washington for a Financial Services Forum board meeting, was confirmed by CNBC. JPMorgan Chase CEO Jamie Dimon was the only major bank CEO who could not attend.

A Treasury spokesperson subsequently confirmed the meeting and said Treasury plans to lead further sessions with regulators and institutions on an ongoing basis.

On the commercial side, major US banks have begun internal testing of Mythos for defensive purposes. Goldman Sachs CEO David Solomon told analysts on a quarterly earnings call that the bank has access to the model and has “hypersensitivity” to the enhanced capabilities of new AI systems.

JPMorgan Chase was named as an initial Project Glasswing partner, alongside approximately 40 companies including Amazon, Apple, Google, Microsoft, and Nvidia.

Anthropic has committed $100 million in credits to these partners and $4 million to open-source security organisations, with the explicit goal of building defensive capacity ahead of any public capability release.

The core risk that regulators are assessing is structural rather than individual. Financial institutions run technology stacks that layer decades-old legacy systems with modern cloud infrastructure, creating accumulated technical debt and undiscovered vulnerabilities.

The banking sector’s heavy reliance on a small number of consolidated cloud providers means that a sufficiently capable AI model exploiting vulnerabilities in those providers’ systems could cascade across the entire financial system.

IBM Senior Vice President Rob Thomas has publicly criticised Anthropic’s restricted-access approach, arguing that “security improves more often through scrutiny than through concealment.”

Anthropic’s CEO Dario Amodei has defended the restricted rollout, writing that “the dangers of getting this wrong are obvious, but if we get it right, there is a real opportunity to create a fundamentally more secure internet and world.”

Anthropic’s relationship with the US government remains complicated by a separate dispute. The Department of Defense designated Anthropic a supply chain risk to national security, a classification the company has contested in court.

A federal appeals court denied Anthropic’s request to temporarily block the designation, leaving it barred from DoD contracts, though a separate preliminary injunction allows the company to continue working with other government agencies while the legal challenges proceed.



Source link

Leave a Reply

Subscribe to Our Newsletter

Get our latest articles delivered straight to your inbox. No spam, we promise.

Recent Reviews



Researchers at the University of Washington have developed a new prototype system that could change how people interact with artificial intelligence in daily life. Called VueBuds, the system integrates tiny cameras into standard wireless earbuds, allowing users to ask an AI model questions about the world around them in near real time.

The concept is simple but powerful. A user can look at an object, such as a food package in a foreign language, and ask the AI to translate it. Within about a second, the system responds with an answer through the earbuds, creating a seamless, hands-free interaction.

A Different Approach To AI Wearables

Unlike smart glasses, which have struggled with adoption due to privacy concerns and design limitations, VueBuds takes a more subtle approach. The system uses low-resolution, black-and-white cameras embedded in earbuds to capture still images rather than continuous video.

These images are transmitted via Bluetooth to a connected device, where a small AI model processes them locally. This on-device processing ensures that data does not need to be sent to the cloud, addressing one of the biggest concerns around wearable cameras.

To further enhance privacy, the earbuds include a visible indicator light when recording and allow users to delete captured images instantly.

Engineering Around Power And Performance Limits

One of the biggest challenges the research team faced was power consumption. Cameras require significantly more energy than microphones, making it impractical to use high-resolution sensors like those found in smart glasses.

To solve this, the team used a camera roughly the size of a grain of rice, capturing low-resolution grayscale images. This approach reduces battery usage and allows efficient Bluetooth transmission without compromising responsiveness.

Placement was another key consideration. By angling the cameras slightly outward, the system achieves a field of view between 98 and 108 degrees. While there is a small blind spot for objects held extremely close, researchers found this does not affect typical usage.

The system also combines images from both earbuds into a single frame, improving processing speed. This allows VueBuds to respond in about one second, compared to two seconds when handling images separately.

Performance Compared To Smart Glasses

In testing, 74 participants compared VueBuds with smart glasses such as Meta’s Ray-Ban models. Despite using lower-resolution images and local processing, VueBuds performed similarly overall.

The report showed participants preferred VueBuds for translation tasks, while smart glasses performed better at counting objects. In separate trials, VueBuds achieved accuracy rates of around 83–84% for translation and object identification, and up to 93% for identifying book titles and authors.

Why This Matters And What Comes Next

The research highlights a potential shift in how AI-powered wearables are designed. By embedding visual intelligence into a device people already use, the system avoids many of the barriers faced by smart glasses.

However, limitations remain. The current system cannot interpret color, and its capabilities are still in early stages. The team plans to explore adding color sensors and developing specialised AI models for tasks like translation and accessibility support.

The researchers will present their findings at the Association for Computing Machinery Conference on Human Factors in Computing Systems in Barcelona, offering a glimpse into a future where everyday devices quietly become intelligent assistants.



Source link