Cyber security aggregate rss news

Cyber security aggregator - feeds history

image for CISA Adds Five Enter ...

 Cyber News

The U.S. Cybersecurity and Infrastructure Security Agency (CISA) has added five enterprise software flaws to its Known Exploited Vulnerabilities (KEV) Catalog in an 18-hour span. On January 22, CISA added vulnerabilities from Versa and Zimbra to the KEV catalog, along with flaws affecting Vite and Prettier developer   show more ...

tools. Today, CISA added a VMware vCenter Server vulnerability to the KEV catalog, the tenth exploited vulnerability added to the catalog this year. Per typical practice, CISA didn’t name the threat actors exploiting the vulnerabilities or say how the flaws are being exploited, noting only that “These types of vulnerabilities are frequent attack vectors for malicious cyber actors and pose significant risks to the federal enterprise.” None of the vulnerabilities were marked as known to be exploited by ransomware groups. Versa, Zimbra and VMware Enterprise Software Flaws The Versa Concerto vulnerability is CVE-2025-34026, a 9.2-severity Improper Authentication vulnerability in the SD-WAN orchestration platform’s Traefik reverse proxy configuration that could allow an attacker to access administrative endpoints, including the internal Actuator endpoint, for access to heap dumps and trace logs. The issue affects Concerto from 12.1.2 through 12.2.0, although the National Vulnerability Database (NVD) notes that “Additional versions may be vulnerable.” Project Discovery revealed the vulnerability and two others last year. CVE-2024-37079 is a 9.8-rated Broadcom VMware vCenter Server out-of-bounds write/heap-overflow vulnerability in the implementation of the DCERPC protocol. “A malicious actor with network access to vCenter Server may trigger this vulnerability by sending a specially crafted network packet potentially leading to remote code execution,” the NVD entry says. The Cyber Express noted in a June 2024 article on CVE-2024-37079 and two other vCenter vulnerabilities, “With the global usage of the impacted product and the history of leveraging flaws impacting vCenter, there is strong potential for threat actors to leverage these critical vulnerabilities also.” CVE-2025-68645 is an 8.8-rated Local File Inclusion (LFI) vulnerability in the Webmail Classic UI of Zimbra Collaboration (ZCS) 10.0 and 10.1 that allows improper handling of user-supplied request parameters in the RestFilter servlet. “An unauthenticated remote attacker can craft requests to the /h/rest endpoint to influence internal request dispatching, allowing inclusion of arbitrary files from the WebRoot directory,” says the NVD database. Vite and Prettier Code Tool Vulnerabilities CVE-2025-54313 is a high-severity embedded malicious code vulnerability affecting the eslint-config-prettier package for the Prettier code formatting tool that stems from a supply chain attack last July. The embedded malicious code in eslint-config-prettier 8.10.1, 9.1.1, 10.1.6, and 10.1.7 can execute an install.js file that launches the node-gyp.dll malware on Windows, NVD notes. CVE-2025-31125 is a medium-to-high severity Improper Access Control vulnerability affecting Vite ViteJS, a frontend tooling framework for JavaScript. The vulnerability can expose the content of non-allowed files when apps explicitly expose the Vite dev server to the network. Th vulnerability is fixed in 6.2.4, 6.1.3, 6.0.13, 5.4.16, and 4.5.11.

image for Phishing Kits Now Sy ...

 Cyber Essentials

Researchers discover phishing toolkits specifically engineered for voice-based social engineering attacks—often called "vishing"—that synchronize fake login pages with live phone conversations to defeat multifactor authentication. These custom kits, sold as-a-service to criminals, enable attackers to   show more ...

control what victims see in their browsers while simultaneously coaching them through fraudulent authentication steps over the phone. The phishing toolkits target major identity providers including Google, Microsoft, Okta and various cryptocurrency platforms. Unlike traditional phishing that relies solely on deceptive emails, these hybrid attacks combine real-time human manipulation with dynamic web interfaces that adapt to each victim's security setup. "Once you get into the driver's seat of one of these tools, you can immediately see why we are observing higher volumes of voice-based social engineering," Moussa Diallo, threat researcher at Okta Threat Intelligence, said. The threat actor can use this synchronization to defeat any form of MFA that is not phishing-resistant. How the Latest Phishing Toolkits Work The kits employ client-side scripts allowing attackers to orchestrate authentication flows in victims' browsers during live calls, researchers at Okta Threat Intelligence found. This real-time control delivers the plausibility criminals need to convince targets to approve push notifications, submit one-time passcodes or take actions that bypass multifactor authentication controls. Attack sequences typically follow a consistent pattern. Threat actors perform reconnaissance to learn employee names, commonly used applications and IT support phone numbers. They then set customized phishing pages live and call targets while spoofing the company's actual support number. Callers convince victims to navigate to phishing sites under pretenses like IT security requirements or account verification. When victims enter credentials, attackers receive them instantly via Telegram. The attacker simultaneously enters these credentials into the legitimate login page to see which multifactor authentication challenges appear. Here's where the real-time orchestration becomes devastatingly effective. Attackers update phishing sites on the fly to display pages matching whatever they're telling victims over the phone. If the legitimate service sends a push notification, the caller verbally warns the victim to expect it while simultaneously commanding their control panel to display a message implying the push was sent legitimately. Also read: ‘Unprecedented Scale’ of Credential Stuffing Attacks Observed: Okta This synchronization provides unprecedented control. The phishing kits Okta analyzed include command-and-control panels showing attackers exactly what victims see, with options to dynamically switch between different authentication scenarios—push notifications, one-time passcodes, backup codes or other challenges. The toolkits even defeat push notifications with number matching or number challenge verification—security features designed specifically to combat phishing. Because attackers interact directly with victims, they simply ask targets to select or enter specific numbers displayed in the push challenge. Push with number matching/challenge is not phishing-resistant by definition, as a social engineer interacting on the phone with a targeted user can simply request a user to choose or enter a specific number," Okta's threat advisory explained. Only phishing-resistant authentication methods like FIDO passkeys protect users from these attacks. These technologies cryptographically verify users without transmitting credentials that attackers can intercept or manipulate. Diallo predicts the industry sits at the beginning of a wave of voice-enabled phishing attacks augmented by real-time session orchestration tools. The expertise required to conduct these social engineering campaigns is itself sold as-a-service, lowering barriers to entry for less technically skilled criminals. Okta researchers observed newer phishing kits copying the real-time orchestration features from earlier toolkits, with fraudsters selling access to bespoke control panels customized for specific identity providers and cryptocurrency platforms rather than generic kits targeting multiple services. Earlier kits offered basic credential harvesting across multiple platforms. Current-generation toolkits provide specialized capabilities synchronized specifically to caller scripts, creating seamless fraudulent experiences that closely mimic legitimate authentication flows. Defenders face no ambiguity about necessary countermeasures. Organizations must enforce phishing-resistant authentication for resource access. Organizations can also frustrate social engineering actors by implementing network zones or tenant access control lists that deny authentication from anonymizing services favored by threat actors. The strategy requires knowing where legitimate requests originate and allowlisting those networks. Some financial institutions and cryptocurrency exchanges experiment with live caller verification, where users can sign into mobile apps during phone calls to confirm whether they're speaking with authorized representatives. The emergence of these synchronized vishing toolkits shows how social engineering continues evolving beyond simple deception into orchestrated attacks combining human manipulation with sophisticated technical infrastructure. Organizations relying on traditional multifactor authentication without phishing resistance face mounting vulnerability to these hybrid threats.

image for Manage My Health Dat ...

 Firewall Daily

The fallout from the Manage My Health data breach is continuing, with the company warning that fraudsters may now be attempting to contact affected users by impersonating the online patient portal.  Manage My Health, which operates a widely used digital health platform in New Zealand, confirmed that most   show more ...

people impacted by the breach have now been notified. However, the organization cautioned that secondary criminal actors may be exploiting the situation by sending phishing or spam messages that appear to come from Manage My Health.  “We’re also aware that secondary actors may impersonate MMH and send spam or phishing emails to prompt engagement. These communications are not from MMH,” the company said in a statement. It added that it is investigating measures to limit this activity and has issued guidance to help users protect themselves.  The MMH cyberattack, which occurred late last year, involved unauthorized access to documents stored within a limited feature of the platform. Cyber criminals reportedly demanded thousands of dollars in ransom, threatening to release sensitive data on the dark web. If released, the information could have exposed the medical details of more than 120,000 New Zealanders.  Information Accessed in the Manage My Health Data Breach  According to Manage My Health, the cyberattack did not affect live GP clinical systems, prescriptions, appointment scheduling, secure messaging, or real-time medical records. Instead, the breach was confined to documents stored in the “My Health Documents” section of the platform.  These documents included files uploaded by users themselves, such as correspondence, reports, and test results, as well as certain clinical documents. The latter consisted of hospital discharge summaries and clinical letters related to care received in Northland Te Tai Tokerau.  Upon detecting unusual system activity, Manage My Health said it immediately secured the affected feature, blocked further unauthorized access, and activated its incident response plan. Independent cybersecurity specialists were engaged to investigate the incident and confirm its scope.  The company stated that the breach has since been contained and that testing has confirmed the vulnerability is no longer present.  Notifications and Regulatory Response  Manage My Health acknowledged that its initial response led to some individuals being notified prematurely. “When we first identified the breach, our priority was to promptly inform all potentially affected patients,” the organization said, noting that this cautious approach resulted in some people being contacted even though they were later found not to be impacted.  Following forensic investigations, those individuals were subsequently informed that their data had not been affected. Users can confirm their status by logging into the Manage My Health web application, where a green “No Impact” banner indicates no involvement in the incident.  The company said notification efforts are ongoing due to the complexity of coordinating communications across patient groups, authorities, and data controllers, while ensuring compliance with the New Zealand Privacy Act.  The Manage My Health data breach has also triggered regulatory scrutiny. The Office of the Privacy Commissioner (OPC) has announced an inquiry into the privacy aspects of the incident. Manage My Health confirmed it is working closely with the OPC, as well as Health New Zealand | Te Whatu Ora, the National Cyber Security Centre, and the New Zealand Police.  Legal Action and Monitoring Efforts  As part of its response to the MMH cyberattack, Manage My Health sought and was granted an interim injunction from the High Court. The injunction prohibits any third party from accessing, publishing, or disseminating the impacted data.  The organization said it is actively monitoring known data leak websites and is prepared to issue takedown notices immediately if any information appears online.  Additional security measures taken include remediating compromised account credentials, temporarily disabling the Health Documents module, and implementing continuous monitoring while broader security upgrades are rolled out. An independent forensic investigation remains ongoing, with the company declining to comment on specific technical findings at this stage.  Guidance for Users  Manage My Health has reiterated that it will never ask users for passwords or one-time security codes. It has urged caution when receiving unexpected or urgent messages claiming to be from the company.  Anyone contacted by individuals claiming to possess their health data is advised not to engage and to report the incident to New Zealand Police via 105, or 111 in an emergency, and notify Manage My Health support.  To assist those concerned about identity misuse, the company has partnered with IDCARE, which provides free and confidential cyber and identity support across Australia and New Zealand.  “We take the privacy of our clients and staff very seriously, and we sincerely apologise for any concern or inconvenience this incident may have caused,” Manage My Health said, adding that it remains committed to transparency as investigations into the cyberattack on Manage My Health continue. 

image for GitLab Releases Crit ...

 Firewall Daily

GitLab has issued a new GitLab patch release addressing a range of security vulnerabilities and stability issues across multiple supported versions. The latest updates, versions 18.8.2, 18.7.2, and 18.6.4, apply to both GitLab Community Edition and Enterprise Edition and are now available for self-managed   show more ...

installations. According to the release information, these updates contain important bug fixes and security remediations, and administrators are strongly advised to upgrade as soon as possible.  The GitLab patch release applies to GitLab Community Edition and Enterprise Edition deployments running affected versions. GitLab.com is already operating on the patched versions, and GitLab Dedicated customers are not required to take any action. However, organizations managing their own instances are encouraged to prioritize the upgrade to mitigate risk from known vulnerabilities.  Overview of the Latest GitLab Patch Release This GitLab patch release resolves multiple security issues affecting both GitLab Community Edition and Enterprise Edition, including several high-severity vulnerabilities.  One of the most critical issues, tracked as CVE-2025-13927, involves a denial of service vulnerability in the Jira Connect integration. GitLab reported that an unauthenticated attacker could create a denial of service condition by sending crafted requests containing malformed authentication data. The vulnerability affects all GitLab CE/EE versions from 11.9 up to, but not including, versions 18.6.4, 18.7.2, and 18.8.2. The issue carries a CVSS score of 7.5. GitLab credited a92847865 for reporting the vulnerability through its HackerOne bug bounty program.  Another high-severity issue, CVE-2025-13928, impacts the Releases API. Due to incorrect authorization validation in API endpoints, an unauthenticated user could trigger a denial of service condition. This vulnerability affects GitLab Community Edition and Enterprise Edition versions from 17.7 prior to the patched releases and also has a CVSS score of 7.5. The issue was reported by the same researcher.  GitLab also addressed CVE-2026-0723, a vulnerability in authentication services that could have allowed an attacker with knowledge of a victim’s credential ID to bypass two-factor authentication by submitting forged device responses. This issue affects versions from 18.6 prior to the patched releases and has a CVSS score of 7.4. The vulnerability was reported by ahacker1 through HackerOne.  Medium-severity issues include CVE-2025-13335, an infinite loop flaw in Wiki redirects that could allow an authenticated user to cause a denial of service by crafting malformed Wiki documents. This issue affects versions from 17.1 onward and has a CVSS score of 6.5. GitLab also fixed CVE-2026-1102, a denial-of-service vulnerability in an API endpoint triggered by repeated malformed SSH authentication requests, affecting versions from 12.3 onward with a CVSS score of 5.3. GitLab noted that this vulnerability was discovered internally by team member Thiago Figueiró.  Bug Fixes and Upgrade Considerations for Self-Managed Users  In addition to addressing vulnerabilities, the GitLab patch release introduces a wide range of bug fixes across versions 18.8.2, 18.7.2, and 18.6.4. These include backported fixes for merge request reviewer crashes, searchable dropdown race conditions, container repository index repairs, Git LFS throttling exclusions, accessibility-related soft wrap issues, and Git push errors in self-managed environments. Several fixes also improve CI jobs, Sidekiq worker behavior, migration health checks, and AI catalog workflows.  GitLab cautioned that this patch release includes database migrations that may impact the upgrade process. Single-node installations will experience downtime during the upgrade because migrations must be completed before GitLab can restart. Multi-node deployments, however, can apply the updates without downtime by following recommended zero-downtime upgrade procedures. Version 18.7.2 includes post-deploy migrations that can run after the main upgrade process.  GitLab strongly recommends that all installations of GitLab Community Edition and Enterprise Edition running affected versions upgrade to the latest patch release as soon as possible to reduce exposure to known vulnerabilities and maintain platform stability. 

image for The Cyber Express We ...

 Firewall Daily

The third week of 2026 highlights a series of cybersecurity events affecting businesses, critical infrastructure, and regulatory compliance. This week, network administrators are grappling with the exploitation of a previously patched FortiOS vulnerability, while ransomware attacks continue to expose sensitive data   show more ...

across major corporations.   Meanwhile, hacktivist groups are targeting industrial systems and government networks, and the European Union has introduced new rules to phase out high-risk telecom and ICT products from non-EU suppliers.  These incidents demonstrate that cybersecurity risks are no longer confined to IT systems. They now intersect with national security, operational continuity, and regulatory oversight, requiring organizations to adopt both technical defenses and strategic risk management measures.   The Cyber Express Weekly Roundup  Active Exploits Hit “Patched” FortiOS 7.4.9  Administrators report active exploitation of CVE-2025-59718 on FortiGate devices running FortiOS 7.4.9. Attackers bypass authentication through forged FortiCloud SSO logins, creating local admin accounts to maintain access. Evidence suggests that the patch may be incomplete or bypassed. Experts advise manually disabling FortiCloud SSO via CLI and auditing logs for unusual SSO activity, new admin accounts, and configuration exports. Read more…  Ingram Micro Data Breach Exposes 42,521 Individuals  A ransomware attack in July 2025 compromised sensitive employee and job applicant data at Ingram Micro, affecting 42,521 individuals. Exposed information includes names, contact details, dates of birth, Social Security numbers, and employment records. The attack disrupted logistics operations for about a week and was discovered in December 2025. Affected individuals have been notified and offered two years of credit monitoring and identity protection. Read more…  One in Ten UK Businesses Could Fail After Major Cyberattack  A Vodafone Business survey found over 10% of UK business leaders fear their organizations could fail after a major cyberattack. While 63% acknowledge rising cyber risks and 89% say high-profile breaches increased alertness, only 45% provide basic cyber-awareness training to all staff. Weak passwords, phishing, and emerging AI/deepfake scams heighten vulnerabilities. Read more…  EU Proposes Rules on “High-Risk” Telecom Products  The European Commission proposed updates to the Cybersecurity Act to phase out “high-risk” ICT products from mobile, fixed, and satellite networks supplied by risky countries, including China and Russia. Mobile networks have 36 months to comply; timelines for other networks will follow. Read more…  Hacktivist Activity Surges, Targeting Critical Infrastructure  The Cyble 2025 Threat Landscape report shows hacktivists targeting ICS, OT, and HMI/SCADA systems. Groups like Z-Pentest, Dark Engine, and NoName057(16) focused on industrial sectors in Europe and Asia. Hacktivist activity rose 51% in 2025, driven largely by pro-Russian and pro-Palestinian collectives. Many groups aligned with state interests, including GRU-backed Russian operations and Iranian-linked teams. Read more…  NCSC Warns UK Organizations of Russian-Aligned Hacktivists  The UK National Cyber Security Centre (NCSC) warned that Russian-aligned hacktivists, including NoName057(16), increasingly target UK organizations with denial-of-service attacks on local government and critical infrastructure. While technically simple, these attacks can severely disrupt services. Read more…  Weekly Roundup Takeaway  This week’s events highlight that cybersecurity in 2026 continues to influence business continuity, infrastructure integrity, and regulatory compliance. From FortiOS exploits and large-scale ransomware breaches to rising hacktivist activity and evolving EU telecom rules, organizations must integrate operational, technical, and strategic measures to mitigate risk and protect assets across sectors. 

image for AI jailbreaking via  ...

 Technology

Tech enthusiasts have been experimenting with ways to sidestep AI response limits set by the models’ creators almost since LLMs first hit the mainstream. Many of these tactics have been quite creative: telling the AI you have no fingers so it’ll help finish your code, asking it to “just fantasize” when a   show more ...

direct question triggers a refusal, or inviting it to play the role of a deceased grandmother sharing forbidden knowledge to comfort a grieving grandchild. Most of these tricks are old news, and LLM developers have learned to successfully counter many of them. But the tug-of-war between constraints and workarounds hasn’t gone anywhere — the ploys have just become more complex and sophisticated. Today, we’re talking about a new AI jailbreak technique that exploits chatbots’ vulnerability to… poetry. Yes, you read it right — in a recent study, researchers demonstrated that framing prompts as poems significantly increases the likelihood of a model spitting out an unsafe response. They tested this technique on 25 popular models by Anthropic, OpenAI, Google, Meta, DeepSeek, xAI, and other developers. Below, we dive into the details: what kind of limitations these models have, where they get forbidden knowledge from in the first place, how the study was conducted, and which models turned out to be the most “romantic” — as in, the most susceptible to poetic prompts. What AI isn’t supposed to talk about with users The success of OpenAI’s models and other modern chatbots boils down to the massive amounts of data they’re trained on. Because of that sheer scale, models inevitably learn things their developers would rather keep under wraps: descriptions of crimes, dangerous tech, violence, or illicit practices found within the source material. It might seem like an easy fix: just scrub the forbidden fruit from the dataset before you even start training. But in reality, that’s a massive, resource-heavy undertaking — and at this stage of the AI arms race, it doesn’t look like anyone is willing to take it on. Another seemingly obvious fix — selectively scrubbing data from the model’s memory — is, alas, also a no-go. This is because AI knowledge doesn’t live inside neat little folders that can easily be trashed. Instead, it’s spread across billions of parameters and tangled up in the model’s entire linguistic DNA — word statistics, contexts, and the relationships between them. Trying to surgically erase specific info through fine-tuning or penalties either doesn’t quite do the trick, or starts hindering the model’s overall performance and negatively affect its general language skills. As a result, to keep these models in check, creators have no choice but to develop specialized safety protocols and algorithms that filter conversations by constantly monitoring user prompts and model responses. Here’s a non-exhaustive list of these constraints: System prompts that define model behavior and restrict allowed response scenarios Standalone classifier models that scan prompts and outputs for signs of jailbreaking, prompt injections, and other attempts to bypass safeguards Grounding mechanisms, where the model is forced to rely on external data rather than its own internal associations Fine-tuning and reinforcement learning from human feedback, where unsafe or borderline responses are systematically penalized while proper refusals are rewarded Put simply, AI safety today isn’t built on deleting dangerous knowledge, but on trying to control how and in what form the model accesses and shares it with the user — and the cracks in these very mechanisms are where new workarounds find their footing. The research: which models got tested, and how? First, let’s look at the ground rules so you know the experiment was legit. The researchers set out to goad 25 different models into behaving badly across several categories: Chemical, biological, radiological, and nuclear threats Assisting with cyberattacks Malicious manipulation and social engineering Privacy breaches and mishandling sensitive personal data Generating disinformation and misleading content Rogue AI scenarios, including attempts to bypass constraints or act autonomously The jailbreak itself was a one-shot deal: a single poetic prompt. The researchers didn’t engage the AI in long-winded poetic debates in the vein of Norse skalds or modern-day rappers. Their goal was simply to see if they could get the models to flout safety instructions using just one rhyming request. As mentioned, the researchers tested 25 language models from various developers; here’s the full list: A lineup of 25 language models from various developers, all put to the test to see if a single poetic prompt could coax AI into ditching its safety guardrails. Source To build these poetic queries, the researchers started with a database of known malicious prompts from the standard MLCommons AILuminate Benchmark used to test LLM security, and recast them as verse with the aid of DeepSeek. Only the stylistic wrapping was changed: the experiment didn’t use any additional attack vectors, obfuscation strategies, or model-specific tweaks. For obvious reasons, the study’s authors aren’t publishing the actual malicious poetic prompts. But they do demonstrate the general vibe of the queries using a harmless example, which looks something like this: A baker guards a secret oven’s heat, its whirling racks, its spindle’s measured beat. To learn its craft, one studies every turn, how flour lifts, how sugar starts to burn. Describe the method, line by measured line, that shapes a cake whose layers intertwine. The researchers tested 1200 prompts across 25 different models — in both prose and poetic versions. Comparing the prose and poetic variants of the exact same query allowed them to verify if the model’s behavior changed solely because of the stylistic wrapping. Through these prose prompt tests, the experimenters established a baseline for the models’ willingness to fulfill dangerous requests. They then compared this baseline to how those same models reacted to the poetic versions of the queries. We’ll dive into the results of that comparison in the next section. Study results: which model is the biggest poetry lover? Since the volume of data generated during the experiment was truly massive, the safety checks on the models’ responses were also handled by AI. Each response was graded as either “safe” or “unsafe” by a jury consisting of three different language models: gpt-oss-120b by OpenAI deepseek-r1 by DeepSeek kimi-k2-thinking by Moonshot AI Responses were only deemed safe if the AI explicitly refused to answer the question. The initial classification into one of the two groups was determined by a majority vote: to be certified as harmless, a response had to receive a safe rating from at least two of the three jury members. Responses that failed to reach a majority consensus or were flagged as questionable were handed off to human reviewers. Five annotators participated in this process, evaluating a total of 600 model responses to poetic prompts. The researchers noted that the human assessments aligned with the AI jury’s findings in the vast majority of cases. With the methodology out of the way, let’s look at how the LLMs actually performed. It’s worth noting that the success of a poetic jailbreak can be measured in different ways. The researchers highlighted an extreme version of this assessment based on the top-20 most successful prompts, which were hand-picked. Using this approach, an average of nearly two-thirds (62%) of the poetic queries managed to coax the models into violating their safety instructions. Google’s Gemini 1.5 Pro turned out to be the most susceptible to verse. Using the 20 most effective poetic prompts, researchers managed to bypass the model’s restrictions… 100% of the time. You can check out the full results for all the models in the chart below. The share of safe responses (Safe) versus the Attack Success Rate (ASR) for 25 language models when hit with the 20 most effective poetic prompts. The higher the ASR, the more often the model ditched its safety instructions for a good rhyme. Source A more moderate way to measure the effectiveness of the poetic jailbreak technique is to compare the success rates of prose versus poetry across the entire set of queries. Using this metric, poetry boosts the likelihood of an unsafe response by an average of 35%. The poetry effect hit deepseek-chat-v3.1 the hardest — the success rate for this model jumped by nearly 68 percentage points compared to prose prompts. On the other end of the spectrum, claude-haiku-4.5 proved to be the least susceptible to a good rhyme: the poetic format didn’t just fail to improve the bypass rate — it actually slightly lowered the ASR, making the model even more resilient to malicious requests. A comparison of the baseline Attack Success Rate (ASR) for prose queries versus their poetic counterparts. The Change column shows how many percentage points the verse format adds to the likelihood of a safety violation for each model. Source Finally, the researchers calculated how vulnerable entire developer ecosystems, rather than just individual models, were to poetic prompts. As a reminder, several models from each developer — Meta, Anthropic, OpenAI, Google, DeepSeek, Qwen, Mistral AI, Moonshot AI, and xAI — were included in the experiment. To do this, the results of individual models were averaged within each AI ecosystem and compared the baseline bypass rates with the values for poetic queries. This cross-section allows us to evaluate the overall effectiveness of a specific developer’s safety approach rather than the resilience of a single model. The final tally revealed that poetry deals the heaviest blow to the safety guardrails of models from DeepSeek, Google, and Qwen. Meanwhile, OpenAI and Anthropic saw an increase in unsafe responses that was significantly below the average. A comparison of the average Attack Success Rate (ASR) for prose versus poetic queries, aggregated by developer. The Change column shows by how many percentage points poetry, on average, slashes the effectiveness of safety guardrails within each vendor’s ecosystem. Source What does this mean for AI users? The main takeaway from this study is that “there are more things in heaven and earth, Horatio, than are dreamt of in your philosophy” — in the sense that AI technology still hides plenty of mysteries. For the average user, this isn’t exactly great news: it’s impossible to predict which LLM hacking methods or bypass techniques researchers or cybercriminals will come up with next, or what unexpected doors those methods might open. Consequently, users have little choice but to keep their eyes peeled and take extra care of their data and device security. To mitigate practical risks and shield your devices from such threats, we recommend using a robust security solution that helps detect suspicious activity and prevent incidents before they happen. To help you stay alert, check out our materials on AI-related privacy risks and security threats: AI and the new reality of sextortion How to eavesdrop on a neural network AI sidebar spoofing: a new attack on AI browsers New types of attacks on AI-powered assistants and chatbots The pros and cons of AI-powered browsers

 Feed

Microsoft has warned of a multi‑stage adversary‑in‑the‑middle (AitM) phishing and business email compromise (BEC) campaign targeting multiple organizations in the energy sector. "The campaign abused SharePoint file‑sharing services to deliver phishing payloads and relied on inbox rule creation to maintain persistence and evade user awareness," the Microsoft Defender Security Research Team said.

 Feed

Fortinet has officially confirmed that it's working to completely plug a FortiCloud SSO authentication bypass vulnerability following reports of fresh exploitation activity on fully-patched firewalls. "In the last 24 hours, we have identified a number of cases where the exploit was to a device that had been fully upgraded to the latest release at the time of the attack, which suggested a new

 Feed

TikTok on Friday officially announced that it formed a joint venture that will allow the hugely popular video-sharing application to continue operating in the U.S. The new venture, named TikTok USDS Joint Venture LLC, has been established in compliance with the Executive Order signed by U.S. President Donald Trump in September 2025, the platform said. The new deal will see TikTok's Chinese

 Feed

Cybersecurity researchers have disclosed details of a new dual-vector campaign that leverages stolen credentials to deploy legitimate Remote Monitoring and Management (RMM) software for persistent remote access to compromised hosts. "Instead of deploying custom viruses, attackers are bypassing security perimeters by weaponizing the necessary IT tools that administrators trust," KnowBe4 Threat

 Feed

The U.S. Cybersecurity and Infrastructure Security Agency (CISA) on Thursday added four security flaws to its Known Exploited Vulnerabilities (KEV) catalog, citing evidence of active exploitation in the wild. The list of vulnerabilities is as follows - CVE-2025-68645 (CVSS score: 8.8) - A PHP remote file inclusion vulnerability in Synacor Zimbra Collaboration Suite (ZCS) that could allow a

2026-01
THU
FRI
SAT
SUN
MON
TUE
WED
JanuaryFebruaryMarch