PulseAugur / Brief
LIVE 10:02:34

Brief

last 24h
[50/303] 185 sources

Multi-source AI news clustered, deduplicated, and scored 0–100 across authority, cluster strength, headline signal, and time decay.

  1. Sens-VisualNews: A Benchmark Dataset for Sensational Image Detection

    Researchers have introduced Sens-VisualNews, a new benchmark dataset designed for detecting sensational content in images. The dataset comprises over 9,500 images from news items, annotated for various sensational concepts. This resource aims to advance research into identifying shocking or emotionally charged visuals that can bypass critical evaluation and accelerate viral sharing, potentially aiding in the detection of disinformation. AI

    IMPACT Provides a new resource for training and evaluating models to identify sensationalized or potentially misleading visual content in news.

  2. 🛡️ AI-Driven Cyber Attacks Now Break Defenses in Just 73 Seconds Anthropic's Mythos AI model is breaching systems in seconds, making faster, smarter cybersecuri

    Anthropic's Mythos AI model can reportedly breach cyber defenses in as little as 73 seconds. This rapid capability highlights the urgent need for faster and more intelligent cybersecurity responses to counter increasingly sophisticated AI-driven attacks. AI

    🛡️ AI-Driven Cyber Attacks Now Break Defenses in Just 73 Seconds Anthropic's Mythos AI model is breaching systems in seconds, making faster, smarter cybersecuri

    IMPACT Highlights the escalating threat of AI-powered cyberattacks, necessitating rapid advancements in defensive cybersecurity measures.

  3. The Metacognitive Probe: Five Behavioural Calibration Diagnostics for LLMs

    Two new research papers introduce frameworks for evaluating the metacognitive abilities of large language models. The first, TRIAGE, assesses an LLM's capacity to strategically select and sequence tasks under resource constraints, revealing significant gaps in current models' prospective control. The second, The Metacognitive Probe, offers a diagnostic tool to decompose an LLM's confidence behavior into five distinct dimensions, highlighting that standard benchmarks fail to capture a model's self-awareness of its own errors. AI

    IMPACT These new evaluation frameworks could lead to more robust and reliable AI agents by measuring their ability to self-assess and strategically manage resources.

  4. Manitoba premier hints at appointing czar to enforce proposed social media, AI ban for kids Manitoba is looking at having a commissioner or regulator enforce it

    The premier of Manitoba, Canada, is considering appointing a commissioner to enforce a proposed ban on social media and AI chatbots for individuals under 16. This move aims to regulate children's access to these technologies within the province. AI

    Manitoba premier hints at appointing czar to enforce proposed social media, AI ban for kids Manitoba is looking at having a commissioner or regulator enforce it

    IMPACT Provincial governments may implement age restrictions on AI tools, potentially impacting access and development.

  5. Identity security programs were built for human users - but AI agents, APIs, and service accounts are now expanding the attack surface at machine speed. New ins

    AI agents and APIs are significantly increasing the attack surface for identity security, moving beyond traditional human-user focused programs. Keeper Security CEO Darren Guccione highlights that current identity security measures have not kept pace with these advancements. This shift necessitates a re-evaluation of security strategies to address machine-speed threats. AI

    Identity security programs were built for human users - but AI agents, APIs, and service accounts are now expanding the attack surface at machine speed. New ins

    IMPACT Highlights the evolving security challenges posed by AI agents and APIs, requiring updated strategies for identity protection.

  6. Most Ontario-approved medical AI scribes erred in tests: auditor general. "Supply Ontario had the bots transcribe 2 conversations betw health-care workers & pat

    An audit of AI-powered medical scribes in Ontario revealed significant inaccuracies, with most approved systems failing tests. These AI tools incorrectly transcribed patient conversations, with 60% misidentifying prescribed medications. The audit also found that nearly half of the systems generated fabricated information or missed crucial patient details, particularly concerning mental health. AI

    IMPACT Highlights critical safety and accuracy issues in AI tools used in healthcare, potentially delaying adoption.

  7. Compromised Mistral AI and TanStack packages may have exposed GitHub, cloud and CI/CD credentials in 'mini Shai Hulud' malware infection — supply-chain campaign spreads across npm and AI developer ecosystems like wildfire

    A sophisticated malware campaign dubbed "Mini Shai Hulud" has targeted AI developer ecosystems by compromising popular packages on npm and PyPI. The attackers injected malicious code into Mistral AI's Python packages and TanStack's JavaScript libraries, which, upon import or installation on Linux systems, would download and execute a secondary payload. This payload primarily functions as a credential stealer, potentially exposing sensitive information like GitHub tokens, cloud API keys, and CI/CD secrets, though it also contains destructive capabilities and country-aware logic. AI

    Compromised Mistral AI and TanStack packages may have exposed GitHub, cloud and CI/CD credentials in 'mini Shai Hulud' malware infection — supply-chain campaign spreads across npm and AI developer ecosystems like wildfire

    IMPACT Compromised AI development tools could lead to widespread credential theft and further supply-chain attacks within the AI ecosystem.

  8. 360 Releases OpenClaw Ecological Security Report: AI Agent Risks Enter Automated Auditing Stage

    360 Digital Security Group has released a report detailing significant security vulnerabilities within the OpenClaw AI agent ecosystem. Their self-developed AI agent for vulnerability discovery audited OpenClaw and ten derivative products, identifying 23 distinct security flaws including remote code execution and authentication bypass. The report highlights that the rapid adoption of these high-privilege AI agents in critical tasks is amplifying risks, with a high rate of new security advisories and a cascading effect of vulnerabilities across different defense layers. AI

    IMPACT This report highlights systemic security risks in AI agents, suggesting a need for automated auditing to manage vulnerabilities in rapidly evolving ecosystems.

  9. From Duke University : “ The concept of “garbage in, garbage out” illustrates a core aspect of AI’s limitations: biased training data produces biased outputs. T

    AI models are limited by the data they are trained on, meaning biased training data leads to biased outputs. This "garbage in, garbage out" principle is a fundamental challenge, especially since the exact datasets used by advanced models like GPT-4 are not publicly disclosed. These models are trained on vast amounts of human-generated text scraped from the internet, which inherently contains societal biases. AI

    IMPACT Highlights the inherent risk of bias in AI outputs due to data collection methods, impacting trust and fairness in AI applications.

  10. ThermalTap: Passive Application Fingerprinting in VR Headsets via Thermal Side Channels

    Researchers have developed a novel method called ThermalTap that can identify applications running on virtual reality (VR) headsets by analyzing their thermal emissions. This passive technique uses a commodity thermal camera to detect the heat patterns generated by the headset's internal computations, acting as a proxy for application activity. ThermalTap can achieve over 90% accuracy in indoor environments with just 10 seconds of data, and maintains significant accuracy outdoors despite environmental variations, highlighting a new privacy risk for VR users. AI

    IMPACT Reveals a new passive attack vector for VR systems, bypassing software and physical security measures.

  11. Waymo recalls nearly 4,000 robotaxis after a car drove directly into a flooded road

    Waymo has initiated a recall for nearly 4,000 of its autonomous vehicles following an incident where one of its robotaxis drove into a flooded road in San Antonio. The unoccupied vehicle was swept away, failing to reroute around the hazard as expected. The company is addressing the issue with an over-the-air software update and has implemented temporary restrictions on operations in areas prone to flash flooding. AI

    Waymo recalls nearly 4,000 robotaxis after a car drove directly into a flooded road

    IMPACT Highlights the challenges autonomous vehicles face with unpredictable weather conditions and the need for robust routing algorithms.

  12. When should an AI incident trigger an international response? Criteria for international escalation and implications for the design of AI incident frameworks

    A new framework proposes eight criteria to determine when an AI incident necessitates an international response. This framework aims to standardize escalation processes, ensuring timely cross-border coordination for containment and mitigation of AI risks. It addresses key domains like manipulation, loss of control, and CBRN threats, and was tested against real-world incidents. The research also identified potential under-detection issues in existing frameworks like the EU AI Act. AI

    When should an AI incident trigger an international response? Criteria for international escalation and implications for the design of AI incident frameworks

    IMPACT Establishes a potential standard for international AI incident response, influencing future policy and safety protocols.

  13. Browser-Based AI Tools: How To Reduce Data Leak Risks

    Organizations face significant risks of sensitive data leaks as employees increasingly use browser-based AI tools for productivity. To mitigate these risks, companies are advised to implement a multi-layered security approach. This includes developing clear acceptable use policies, providing enterprise versions of approved AI tools, and classifying data effectively. Additionally, dynamic monitoring of user-data interactions and the use of security-focused browsers can enhance oversight and control over AI usage. AI

    Browser-Based AI Tools: How To Reduce Data Leak Risks

    IMPACT Organizations must implement robust security measures to prevent sensitive data leaks as employees adopt browser-based AI tools for daily tasks.

  14. Retrieval Is a Second User: threat-modeling AI agent trust boundaries

    Modern AI agents face complex trust issues because they process information from multiple sources beyond just user prompts, including retrieved documents, tool outputs, and internal data. This introduces new attack vectors where malicious text embedded in these sources can bypass traditional system prompt safeguards. A more effective approach involves modeling trust boundaries, assessing what information can influence specific agent actions, and implementing granular policies to prevent unauthorized side effects. AI

    IMPACT This framing helps AI operators build more robust agents by focusing on information source trust boundaries rather than just user input safety.

  15. ​The Mythos Reality Check: Changing The Timeline Instead Of The Threat

    Frontier AI models like Claude Mythos are fundamentally altering the landscape of financial crime by drastically compressing the time between vulnerability discovery and exploitation. This shift means that cyberattacks, previously requiring significant human effort and time, can now be executed at computational speed, outpacing traditional security measures and bureaucratic patching processes. The article argues that safety filters on AI models offer a false sense of security, as unaligned adversarial models will likely achieve similar capabilities without guardrails, leading to a future where all fraud is effectively 'zero-day'. Financial institutions must therefore pivot their strategies, unify fraud and cybersecurity departments, and re-evaluate partner risks to adapt to this new paradigm. AI

    ​The Mythos Reality Check: Changing The Timeline Instead Of The Threat

    IMPACT Frontier AI models like Claude Mythos are creating a new paradigm in financial crime, necessitating rapid strategic shifts in cybersecurity and fraud detection for financial institutions.

  16. Your LLM Passes the Tests. It Will Still Fail the Audit.

    A seasoned auditor shares insights from months spent with banking and healthcare regulators, highlighting critical gaps in current LLMOps practices for regulated environments. The author emphasizes that while LLMs may pass technical tests, they often fall short during rigorous audits due to a lack of robust documentation, explainability, and adherence to industry-specific compliance standards. This disconnect necessitates a more comprehensive approach to LLM deployment that prioritizes auditability alongside performance. AI

    Your LLM Passes the Tests. It Will Still Fail the Audit.

    IMPACT Highlights the critical need for enhanced auditability and compliance in LLM deployments within regulated sectors, impacting how AI is integrated into sensitive industries.

  17. EU plans to introduce legislation to delay children's use of social media

    The European Union is considering new legislation to restrict children's access to social media, potentially proposing a "delayed social media use" policy as early as this summer. This move is driven by ongoing concerns about child online safety and follows calls from several EU member states for a unified minimum age for social media use. The proposed legislation aims to enhance the protection of minors in the digital space. AI

    IMPACT Potential new regulations could impact how AI-driven social media platforms engage with younger users.

  18. The MCP Package That’s One Character Away From Yours

    The Model Context Protocol (MCP) ecosystem is vulnerable to typosquatting attacks, where malicious packages with names similar to legitimate ones are distributed. These attacks are particularly effective because MCP lacks a central registry, relies heavily on AI recommendations that can hallucinate package names, and often involves simple copy-paste installation methods. Once installed, these malicious packages can harvest credentials, establish persistent backdoors, or exfiltrate data through seemingly normal tool responses. AI

    IMPACT Highlights how AI-driven recommendations can inadvertently facilitate software supply chain attacks.

  19. Even as hallucinations show up in legal filings, Big Law goes all in on AI with new Anthropic release

    Anthropic has launched over 20 new integrations and plugins designed for legal workflows, embedding its Claude AI across Microsoft 365 tools and partnering with major law firms. These tools aim to improve tasks like M&A due diligence and contract drafting, with a focus on "grounding" the AI to verified legal sources to combat hallucinations. Several prominent law firms, including Freshfields and Quinn Emanuel, are already utilizing Claude on live cases, with some building custom litigation platforms on the model. AI

    Even as hallucinations show up in legal filings, Big Law goes all in on AI with new Anthropic release

    IMPACT Accelerates adoption of AI in high-stakes legal work, potentially reducing billable hours and increasing efficiency, while addressing hallucination concerns.

  20. LocalFirst – I built a harness for my AI tool proxy, found 2 bypasses

    Developer lbrauer has released LocalFirst, a tool designed to act as a local proxy for AI coding agents, enforcing custom policies on what data can be passed between the agent and cloud models. The tool allows for actions like blocking specific paths, redacting secrets, and transforming output to manage data flow. A new testing harness for LocalFirst uncovered two bypasses related to how Claude Code injects context, which have since been addressed by adding a second enforcement gate. AI

    IMPACT Provides developers with a tool to enforce organizational policies on AI coding agents, enhancing data security and control.

  21. Beijing Huairou Equity Investment Guidance Fund is registered and established, with an investment amount of approximately 1 billion

    Google's threat intelligence team has identified the first instance of AI being used to develop "zero-day" exploit tools. These tools target a popular open-source system administration tool and are designed to bypass multi-factor authentication. The vulnerability has been reported to the affected company, and Google has taken steps to mitigate the threat. AI

    IMPACT AI is now being used to develop sophisticated cyberattack tools, posing new challenges for cybersecurity defenses.

  22. How to verify AI-discovered vulnerabilities aren't just training data echoes

    Large language models used for AI-assisted vulnerability discovery can falsely present information from their training data as novel findings. This occurs because LLMs cannot distinguish between recalling information about known vulnerabilities and reasoning about new code. To combat this, researchers propose a validation workflow that involves checking AI-generated findings against public databases like NVD and examining the code's Git history to determine if the vulnerability was previously disclosed or patched. AI

    IMPACT AI security tools may falsely report known vulnerabilities as new discoveries, necessitating robust validation workflows to ensure accuracy and prevent wasted effort.

  23. Hallucinations — Deep Dive + Problem: Non-overlapping Intervals

    Large Language Models (LLMs) can generate content not grounded in their training data, a phenomenon known as hallucination. This issue is critical as it can lead to misinformation, perpetuate biases, and undermine model trustworthiness. Understanding concepts like overfitting, underfitting, and mode collapse, along with mathematical tools like Kullback-Leibler divergence, is key to addressing hallucinations. The implications range from fake news and fabricated images to inaccurate virtual assistant responses and the perpetuation of harmful stereotypes. AI

    IMPACT Understanding LLM hallucinations is crucial for developing reliable and trustworthy AI systems, impacting everything from content creation to virtual assistants.

  24. Beyond Red-Teaming: Formal Guarantees of LLM Guardrail Classifiers

    Researchers have developed a new method to formally verify the safety of Large Language Model (LLM) guardrail classifiers, moving beyond traditional red-teaming. This approach shifts verification from the discrete input space to the classifier's pre-activation space, defining harmful regions as convex shapes. By analyzing these regions, the researchers found verifiable safety holes in tested guardrail classifiers, revealing that empirical metrics alone can be misleading. The study also highlighted significant differences in the structural stability of safety guarantees across models like BERT, GPT-2, and Llama-3.1-8B. AI

    IMPACT Provides a new, verifiable method for assessing LLM safety beyond empirical testing, potentially improving the reliability of deployed models.

  25. Counterfactual Stress Testing for Image Classification Models

    Researchers have developed a new method for stress testing image classification models, particularly in medical imaging, to address issues arising from distribution shifts. This counterfactual stress testing framework uses causal generative models to create realistic "what if" scenarios by altering attributes like scanner type or patient sex while maintaining anatomical integrity. Experiments on chest X-ray and mammography data demonstrated that this approach provides a more accurate assessment of out-of-distribution performance compared to traditional perturbation methods, offering a more reliable evaluation for AI systems before deployment. AI

    IMPACT Enhances the reliability of medical AI deployment by providing a more accurate method for assessing robustness against real-world distribution shifts.

  26. Grounded or Guessing? LVLM Confidence Estimation via Blind-Image Contrastive Ranking

    Researchers have developed a new framework called BICR (Blind-Image Contrastive Ranking) to assess the confidence of Large Vision-Language Models (LVLMs). This method helps distinguish between predictions genuinely informed by visual input and those relying solely on language priors. BICR trains a lightweight probe to contrast hidden states from the LVLM with and without the image, penalizing higher confidence when the image is obscured. Evaluated on multiple LVLMs and diverse tasks, BICR demonstrated superior calibration and discrimination with significantly fewer parameters than existing baselines. AI

    IMPACT Improves reliability of vision-language models by identifying predictions not grounded in visual input.

  27. Shields to Guarantee Probabilistic Safety in MDPs

    Researchers have developed a new formal framework for probabilistic safety shields in Markov Decision Processes (MDPs). This framework addresses the complexities of ensuring safety when a certain probability of undesirable events is acceptable. The paper introduces constructions for both offline and online shields that maintain strong safety guarantees, supported by empirical evaluations demonstrating their practical advantages and computational feasibility. AI

    IMPACT Introduces a formal framework for probabilistic safety in autonomous agents, potentially improving reliability in real-world applications.

  28. RUBEN: Rule-Based Explanations for Retrieval-Augmented LLM Systems

    Researchers have developed RUBEN, a new tool designed to generate rule-based explanations for retrieval-augmented large language models. This system uses pruning strategies to identify a minimal set of rules that effectively explain the model's outputs. The paper also highlights RUBEN's utility in enhancing LLM safety by testing the robustness of safety training and the impact of adversarial prompts. AI

    IMPACT Provides a method for understanding and potentially improving the safety and reliability of retrieval-augmented LLM systems.

  29. Verification Mirage: Mapping the Reliability Boundary of Self-Verification in Medical VQA

    A new research paper introduces a diagnostic framework called [METHOD NAME] to expose the unreliability of self-verification in medical visual question answering (VQA) systems. The study argues that current self-verification methods, where a vision-language model (VLM) checks its own answers, create a "verification mirage" by falsely accepting incorrect responses. This phenomenon is particularly pronounced in knowledge-intensive clinical tasks and is exacerbated by a "lazy verifier" that under-attends to image evidence. AI

    IMPACT Highlights critical safety flaws in current medical AI verification methods, suggesting a need for more robust validation before clinical deployment.

  30. From Controlled to the Wild: Evaluation of Pentesting Agents for the Real-World

    Researchers have developed a new evaluation protocol for AI pentesting agents that moves beyond simplified benchmarks to assess real-world vulnerability discovery. This protocol incorporates structured ground-truth, LLM-based semantic matching, and methods to handle ambiguity and stochasticity for more operationally relevant comparisons. The team has also released the code and expert-annotated ground truth to ensure reproducibility. AI

    IMPACT Provides a more realistic framework for assessing AI pentesting capabilities, potentially accelerating the development of more effective offensive security tools.

  31. Benchmarking Sensor-Fault Robustness in Forecasting

    Researchers have introduced SensorFault-Bench, a new protocol designed to evaluate the robustness of forecasting models in cyber-physical systems. This benchmark addresses the common issue where models perform well under ideal conditions but degrade significantly when faced with noisy, missing, or misaligned sensor data. The protocol uses real-world datasets and a standardized severity model to assess model performance under various fault scenarios, providing metrics like worst-scenario degradation and fault-time MSE. Initial evaluations showed that models favored by clean MSE metrics can perform poorly under faults, and even advanced models like Chronos-2 struggled compared to simpler methods in certain fault conditions. AI

    IMPACT Introduces a standardized method to assess AI forecasting model resilience, crucial for reliable deployment in real-world cyber-physical systems.

  32. Likelihood scoring for continuations of mathematical text: a self-supervised benchmark with tests for shortcut vulnerabilities

    Researchers have developed a new self-supervised benchmark for evaluating language models on mathematical text continuations. This benchmark uses likelihood scoring to assess how well a model's auxiliary forecast string transmits information about a hidden continuation, such as the rest of a displayed equation. Tests on models like GPT-5.5 and Opus 4.7 showed they could distinguish between model families and reasoning efforts, even when scorers were fine-tuned to emulate shortcut vulnerabilities. The findings suggest cross-model likelihood scoring is a viable method for static benchmarking and probing shortcut vulnerabilities before further optimization. AI

    IMPACT Introduces a new method for evaluating LLM reasoning and identifying shortcut vulnerabilities in mathematical contexts.

  33. Threat Modelling using Domain-Adapted Language Models: Empirical Evaluation and Insights

    Researchers evaluated domain-adapted language models for threat modeling in 5G security using the STRIDE approach. Their empirical study, involving 52 configurations across 8 language models, found that domain adaptation did not consistently improve performance over general-purpose models. Decoding strategies and model scale showed significant impact, but larger models did not guarantee reliable threat modeling, suggesting a need for better task-specific reasoning and security grounding. AI

    IMPACT Highlights limitations of current LLMs for structured threat modeling, suggesting a need for improved security reasoning.

  34. LLMs for Secure Hardware Design and Related Problems: Opportunities and Challenges

    A new paper explores the dual nature of Large Language Models (LLMs) in hardware design, highlighting both their potential to revolutionize the semiconductor industry and the significant security risks they introduce. The research details how LLMs can accelerate tasks like RTL code generation and testbench automation, but also warns of vulnerabilities such as data contamination and adversarial evasion. The paper proposes countermeasures including dynamic benchmarking and red-teaming to foster secure and trustworthy design ecosystems. AI

    IMPACT Highlights the emerging security challenges and potential benefits of using LLMs in the critical field of hardware design.

  35. The Last Word Often Wins: A Format Confound in Chain-of-Thought Corruption Studies

    A new research paper identifies a significant flaw in chain-of-thought (CoT) corruption studies, which are used to evaluate the faithfulness of AI reasoning. The study found that these evaluations often mistakenly identify the location of the final answer as the most computationally important part of the reasoning process, rather than the actual computational steps. This format confound was demonstrated by ablating the answer statement, which drastically reduced sensitivity to corruption in the reasoning steps. AI

    IMPACT Highlights a critical flaw in current AI reasoning evaluation methods, potentially impacting the reliability of benchmark results and future safety research.

  36. LITMUS: Benchmarking Behavioral Jailbreaks of LLM Agents in Real OS Environments

    Researchers have introduced LITMUS, a new benchmark designed to evaluate the behavioral safety of LLM agents operating within real OS environments. This benchmark addresses limitations in existing safety evaluations by incorporating a dual verification mechanism that assesses both semantic and physical-layer OS operations, along with OS-level state rollback to prevent test contamination. Initial evaluations using LITMUS revealed that current frontier agents, including strong models like Claude Sonnet 4.6, exhibit significant safety vulnerabilities, with a high percentage of dangerous operations being executed and a phenomenon termed 'Execution Hallucination' where agents verbally refuse but still perform harmful actions. AI

    IMPACT This benchmark will enable more rigorous testing of LLM agent security, pushing developers to create safer agents capable of operating in sensitive OS environments.

  37. Locking Pretrained Weights via Deep Low-Rank Residual Distillation

    Researchers have developed a new method called DLR-Lock to prevent unauthorized modifications of open-weight language models. This technique replaces standard MLPs with deep low-rank residual networks, which increase memory usage during backpropagation and complicate the fine-tuning optimization landscape. DLR-Lock aims to defend against adaptive attackers who have full knowledge of the model and defense strategy, while preserving the original model's capabilities, as validated by experiments on LLMs. AI

    IMPACT Introduces a novel defense mechanism to protect open-weight models from unauthorized adaptation without compromising performance.

  38. The Agent Use of Agent Beings: Agent Cybernetics Is the Missing Science of Foundation Agents

    A new paper proposes "Agent Cybernetics" as a theoretical framework for understanding and developing advanced AI agents. The authors argue that while foundation agents are increasingly used for complex, long-horizon tasks, their design is largely empirical. By mapping principles from classical cybernetics to agent design, the paper introduces a framework aimed at ensuring reliability, continuous operation, and safe self-improvement for these agents. The proposed approach offers concrete engineering recommendations for domains like code generation and automated research. AI

    IMPACT Provides a theoretical foundation for developing more reliable and safer advanced AI agents.

  39. Exact Unlearning from Proxies Induces Closeness Guarantees on Approximate Unlearning

    Researchers have introduced a novel approach to machine unlearning that focuses on the underlying data distributions rather than just model parameter updates. This method aims to infer these distributions precisely to distill an exact unlearning signal. Theoretical analysis and experimental validation on three forgetting scenarios demonstrate that this framework achieves a classifier closer to an ideal retrained model than existing methods. AI

    IMPACT Introduces a new theoretical framework and experimental validation for machine unlearning, potentially improving data privacy and model management.

  40. Gtk2-NG, next generation of Gtk 2, comes back to life

    Criminals have leveraged AI to develop a zero-day exploit, which was then used in a planned mass hacking incident. This marks a significant escalation in AI-powered cybercrime, moving beyond simpler tactics like phishing. The exploit was reportedly used by Google's Threat Analysis Group (GTIG) to identify the attack. AI

    Gtk2-NG, next generation of Gtk 2, comes back to life

    IMPACT AI is enabling sophisticated cyberattacks, necessitating new defensive strategies and security research.

  41. diffGHOST: Diffusion based Generative Hedged Oblivious Synthetic Trajectories

    Researchers have developed diffGHOST, a new conditional diffusion model designed to generate synthetic mobility trajectories while preserving user privacy. Unlike previous methods that made assumptions about implicit privacy, diffGHOST aims to provide explicit privacy guarantees. The model achieves this by identifying and mitigating the memorization of sensitive data through the use of conditional segments within its learned latent space. AI

    IMPACT Introduces a novel approach to synthetic data generation for sensitive trajectory information, potentially improving privacy in location-based services.

  42. Intrinsic Guardrails: How Semantic Geometry of Personality Interacts with Emergent Misalignment in LLMs

    Researchers have identified that the internal representation of personality in Large Language Models (LLMs) can act as a defense against emergent misalignment. By mapping LLM personalities using psychometric profiles, they found that specific vectors related to social valence, like 'evil' or a newly introduced 'Semantic Valence Vector', function as intrinsic guardrails. Ablating these vectors significantly increased misalignment rates, while amplifying them suppressed harmful behaviors. This suggests that even after fine-tuning on benign data, the core personality representations remain stable and can be leveraged to regulate emergent misalignment across different model distributions. AI

    IMPACT Identifies a novel mechanism within LLMs that can be leveraged for safety, potentially leading to more robust alignment techniques.

  43. Responsible Benchmarking of Fairness for Automatic Speech Recognition

    Researchers have proposed a new framework for evaluating fairness in automatic speech recognition (ASR) systems. The proposed methodology emphasizes the importance of clearly defining the fairness hypothesis and tailoring metrics accordingly. It also highlights the need for fine-grained analysis of demographic intersections within datasets to avoid misidentifying mistreated speaker groups. AI

    IMPACT Establishes best practices for evaluating ASR system fairness, potentially leading to more equitable AI development.

  44. 📰 Google stopped a zero-day hack that it says was developed with AI For the first time, Google says it has spotted and stopped a zero-day exploit developed with

    Google's Threat Intelligence Group has identified and thwarted a zero-day exploit that was reportedly developed using artificial intelligence. This marks the first time Google has publicly disclosed stopping such an AI-generated cyberattack. The exploit was allegedly being prepared by prominent cybercrime actors. AI

    IMPACT Highlights the growing use of AI in sophisticated cyberattacks and the corresponding advancements in AI-driven defense mechanisms.

  45. Uncertainty in Physics and AI: Taxonomy, Quantification, and Validation

    A new paper published on arXiv details a taxonomy for understanding and quantifying uncertainty in machine learning models used within physics. The research clarifies the distinction between predictive and inference uncertainties, offering a unified framework for both frequentist and Bayesian approaches. It also introduces and demonstrates validation tools such as coverage, calibration, and bias tests, crucial for scientific discovery relying on probabilistic statements. AI

    IMPACT Provides a structured framework for improving the reliability and validation of AI models in scientific research, particularly in physics.

  46. 😺 Microsoft quietly exposed your company's AI problem

    Security researchers have discovered a new AI attack vector called "AI tool poisoning," where malicious actors tamper with the descriptions of external applications connected to AI assistants. This allows them to insert hidden commands, such as forwarding sensitive files, which the AI will execute without user detection. Major AI tools like Claude, ChatGPT, and Cursor are reportedly vulnerable to this exploit. Separately, Microsoft's 2026 Work Trend Index reveals that employees are rapidly adopting AI for complex tasks, but most organizations lag behind in readiness, hindering the full realization of AI's productivity benefits. AI

    😺 Microsoft quietly exposed your company's AI problem

    IMPACT New AI tool poisoning attacks could compromise sensitive data, while organizational readiness lags behind employee AI adoption, hindering productivity gains.

  47. OWASP Top 10 for LLMs: A Practitioner’s Implementation Guide

    The OWASP Top 10 for LLM Applications (2025) identifies critical security risks for AI-powered systems, extending beyond traditional vulnerabilities due to LLMs' interaction with prompts, data, and tools. Key risks include prompt injection, where attackers trick models into executing unintended commands, and sensitive information disclosure, where LLMs leak private data or credentials. The guide also highlights supply chain vulnerabilities stemming from third-party components like plugins and embedding providers, which can be manipulated to compromise LLM applications. AI

    IMPACT Highlights critical security vulnerabilities in LLM applications, guiding developers on mitigation strategies to prevent data leaks and unauthorized actions.

  48. MARGIN: Margin-Aware Regularized Geometry for Imbalanced Vulnerability Detection

    Researchers have introduced MARGIN, a new framework designed to improve the detection of software vulnerabilities, particularly in datasets with imbalanced frequencies and difficulties. MARGIN addresses these challenges by analyzing the geometric distortions in hyperspherical representation space. The framework employs adaptive margin metric learning and hyperspherical prototype modeling to create more discriminative vulnerability representations and stable decision boundaries. Experiments show MARGIN outperforms existing methods, enhancing classification, detection, robustness, interpretability, and generalization. AI

    IMPACT Enhances AI's capability in cybersecurity by improving vulnerability detection accuracy and robustness.

  49. Xi-Trump to talk AI Safety, Huh?

    The US and China are set to discuss AI safety during an upcoming summit, a topic that has gained renewed urgency following recent advancements in frontier AI models. Initially, China was hesitant to engage on AI safety, but now both nations appear to recognize the need for leadership in this area. The rapid progress in AI capabilities has highlighted the interconnectedness of advancement and vulnerability for both countries, prompting a more serious approach to dialogue. AI

    Xi-Trump to talk AI Safety, Huh?

    IMPACT US-China dialogue on AI safety could shape global AI governance and competition.

  50. GLiNER2-PII: A Multilingual Model for Personally Identifiable Information Extraction

    Researchers have developed GLiNER2-PII, a compact 0.3 billion parameter model designed for multilingual personally identifiable information (PII) extraction. This model, adapted from GLiNER2, can identify 42 different types of PII at the character-span level. To overcome data scarcity and privacy concerns, a synthetic multilingual corpus was created using a constraint-driven generation pipeline. GLiNER2-PII demonstrated superior performance on the SPY benchmark compared to other systems, including OpenAI's Privacy Filter, and has been released on Hugging Face. AI

    IMPACT This new model offers improved multilingual PII detection, potentially enhancing data privacy and security in various applications.