News from the AI & ML world

DeeperML - #aisecurity

Vasu Jakkal@Microsoft Security Blog //
Microsoft has unveiled a significant expansion of its Security Copilot platform, integrating AI agents designed to automate security operations tasks and alleviate the workload on cybersecurity professionals. This move aims to address the increasing volume and complexity of cyberattacks, which are overwhelming security teams that rely on manual processes. The AI-powered agents will handle routine tasks, freeing up IT and security staff to tackle more complex issues and proactive security measures. Microsoft detected over 30 billion phishing emails targeting customers between January and December 2024 highlighting the urgent need for automated solutions.

The expansion includes eleven AI agents, six developed by Microsoft and five by security partners, set for preview in April 2025. Microsoft's agents include the Phishing Triage Agent in Microsoft Defender, Alert Triage Agents in Microsoft Purview, Conditional Access Optimization Agent in Microsoft Entra, Vulnerability Remediation Agent in Microsoft Intune, and Threat Intelligence Briefing Agent in Security Copilot. These agents are purpose-built for security, designed to learn from feedback, adapt to workflows, and operate securely within Microsoft’s Zero Trust framework, ensuring that security teams retain full control over their actions and responses.

Recommended read:
References :
  • The Register - Software: AI agents swarm Microsoft Security Copilot
  • Microsoft Security Blog: Microsoft unveils Microsoft Security Copilot agents and new protections for AI
  • .NET Blog: Learn how the Xbox services team leveraged .NET Aspire to boost their team's productivity.
  • Ken Yeung: Microsoft’s First CTO Says AI Is ‘Three to Five Miracles’ Away From Human-Level Intelligence
  • SecureWorld News: Microsoft Expands Security Copilot with AI Agents
  • www.zdnet.com: Microsoft's new AI agents aim to help security pros combat the latest threats
  • www.itpro.com: Microsoft launches new security AI agents to help overworked cyber professionals
  • www.techrepublic.com: After Detecting 30B Phishing Attempts, Microsoft Adds Even More AI to Its Security Copilot
  • eSecurity Planet: esecurityplanet.com covers Fortifying Cybersecurity: Agentic Solutions by Microsoft and Partners
  • Source: AI innovation requires AI security: Hear what’s new at Microsoft Secure
  • www.csoonline.com: Microsoft has introduced a new set of AI agents for its Security Copilot platform, designed to automate key cybersecurity functions as organizations face increasingly complex and fast-moving digital threats.
  • SiliconANGLE: Microsoft introduces AI agents for Security Copilot
  • SiliconANGLE: Microsoft Corp. is enhancing the capabilities of its popular artificial intelligence-powered Copilot tool with the launch late today of its first “deep reasoning” agents, which can solve complex problems in the way a highly skilled professional might do.
  • Ken Yeung: Microsoft is introducing a new way for developers to create smarter Copilots.
  • www.computerworld.com: Microsoft’s Newest AI Agents Can Detail How They Reason

Ken Yeung@Ken Yeung //
References: Source Asia , SiliconANGLE , SiliconANGLE ...
Microsoft is enhancing its Copilot Studio platform with new 'deep reasoning' capabilities, allowing AI agents to solve complex problems more effectively. This upgrade also includes 'agent flows' which blend AI's flexibility with structured business automation. The new Researcher and Analyst agents for Microsoft 365 Copilot represent a significant step forward in AI agent evolution, enabling them to handle sophisticated tasks requiring detailed analysis and methodical thinking.

Microsoft's Security Copilot service is also getting a boost with a set of AI agents designed to automate repetitive tasks, freeing up security professionals to focus on more critical threats. These AI agents are designed to assist with critical tasks such as phishing, data security, and identity management. These agents showcase the breadth of what can be created when combining enterprise business data, access to advanced reasoning models, and structured workflows.

Recommended read:
References :
  • Source Asia: Microsoft Security Copilot agents and more security innovations
  • SiliconANGLE: Microsoft introduces AI agents for Security Copilot
  • Microsoft Security Blog: Microsoft unveils Microsoft Security Copilot agents and new protections for AI
  • SiliconANGLE: Microsoft 365 Copilot gets AI reasoning skills for advanced research and analysis
  • Ken Yeung: Microsoft Adds Deep Reasoning to Copilot Studio and Launches Reasoning Agents for M365
  • AI News | VentureBeat: Microsoft infuses enterprise agents with deep reasoning, unveils data Analyst agent that outsmarts competitors
  • www.techrepublic.com: After Detecting 30B Phishing Attempts, Microsoft Adds Even More AI to Its Security Copilot
  • www.zdnet.com: Microsoft's new AI agents aim to help security pros combat the latest threats
  • www.itpro.com: Microsoft is expanding its Security Copilot service with new AI agents to help overworked IT teams deal with surging security threats.
  • Source Asia: Introducing two, first-of-their-kind reasoning agents in Microsoft 365 Copilot
  • www.techrepublic.com: Microsoft 365 Copilot’s ‘First-of-Their-Kind Reasoning Agents’ — Here’s What They Do

Megan Crouse@eWEEK //
References: The Register - Software , eWEEK , OODAloop ...
Cloudflare has launched AI Labyrinth, a new tool designed to combat web scraping bots that steal website content for AI training. Instead of simply blocking these crawlers, AI Labyrinth lures them into a maze of AI-generated content. This approach aims to waste the bots' time and resources, providing a more effective defense than traditional blocking methods which can trigger attackers to adapt their tactics. The AI Labyrinth is available as a free, opt-in tool for all Cloudflare customers, even those on the free tier.

The system works by embedding hidden links within a protected website. When suspicious bot behavior is detected, such as ignoring robots.txt rules, the crawler is redirected to a series of AI-generated pages. This content is "real looking" and based on scientific facts, diverting the bot from the original website's content. Because no human would deliberately explore deep into a maze of AI-generated nonsense, anyone who does can be identified as a bot with high confidence. Cloudflare emphasizes that AI Labyrinth also functions as a honeypot, allowing them to identify new bot patterns and improve their overall bot detection capabilities, all while increasing the cost for unauthorized web scraping.

Recommended read:
References :
  • The Register - Software: Cloudflare builds an AI to lead AI scraper bots into a horrible maze of junk content
  • eWEEK: Crowdflare’s Free AI Labyrinth Distracts Crawlers That Could Steal Website Content to Feed AI
  • The Verge: Cloudflare, one of the biggest network internet infrastructure companies in the world, has announced AI Labyrinth, a new tool to fight web-crawling bots that scrape sites for AI training data without permission. The company says in a blog post that when it detects “inappropriate bot behavior,â€� the free, opt-in tool lures crawlers down a path
  • OODAloop: Trapping misbehaving bots in an AI Labyrinth
  • THE DECODER: Instead of simply blocking unwanted AI crawlers, Cloudflare has introduced a new defense method that lures them into a maze of AI-generated content, designed to waste their time and resources.
  • Digital Information World: Cloudflare’s Latest AI Labyrinth Feature Combats Unauthorized AI Data Scraping By Giving Bots Fake AI Content
  • Ars OpenForum: Cloudflare turns AI against itself with endless maze of irrelevant facts
  • Cyber Security News: Cloudflare Introduces AI Labyrinth to Thwart AI Crawlers and Malicious Bots
  • poliverso.org: Cloudflare’s AI Labyrinth Wants Bad Bots To Get Endlessly Lost
  • aboutdfir.com: Cloudflare builds an AI to lead AI scraper bots into a horrible maze of junk content Cloudflare has created a bot-busting AI to make life hell for AI crawlers.

Divya@gbhackers.com //
OpenAI has announced a major update to its Security Bug Bounty Program, increasing the maximum reward to $100,000 for critical vulnerability reports. This represents a fivefold increase from the previous $20,000 payout, signaling the company's commitment to enhancing the security and reliability of its AI platforms. The enhanced program seeks to attract top security researchers worldwide to identify and fix potential threats before they escalate into significant issues.

The increased bug bounty is part of a broader suite of cybersecurity initiatives, including an evolving Cybersecurity Grant Program that has already funded 28 research projects. To further incentivize participation, OpenAI is introducing limited-time bonus promotions, with the first focusing on IDOR access control vulnerabilities. This reflects OpenAI's commitment to rewarding meaningful, high-impact security research that helps protect users and maintain trust in its systems.

Recommended read:
References :
  • gbhackers.com: OpenAI has announced major updates to its cybersecurity initiatives.
  • hackread.com: OpenAI Bug Bounty program boosts max reward to $100,000, expanding scope and offering new incentives to enhance AI security and reliability.
  • BleepingComputer: Artificial intelligence company OpenAI has announced a fivefold increase in the maximum bug bounty rewards for "exceptional and differentiated" critical security vulnerabilities from $20,000 to $100,000.
  • www.bleepingcomputer.com: OpenAI now pays researchers $100,000 for critical vulnerabilities
  • The DefendOps Diaries: OpenAI's Bug Bounty Program: A Strategic Approach to Cybersecurity
  • www.itpro.com: OpenAI announces five-fold increase in bug bounty reward
  • Davey Winder: $100,000 Offered To OpenAI Hackers—What You Need To Know
  • BleepingComputer: Artificial intelligence company OpenAI has announced a fivefold increase in the maximum bug bounty rewards for "exceptional and differentiated" critical security vulnerabilities from $20,000 to $100,000.
  • bsky.app: Artificial intelligence company OpenAI has announced a fivefold increase in the maximum bug bounty rewards for 'exceptional and differentiated' critical security vulnerabilities from $20,000 to $100,000.
  • eWEEK: OpenAI boosts its bug bounty to $100K, expands grants, and partners with experts to strengthen AI security and defend against emerging cyber threats.

Vasu Jakkal@Microsoft Security Blog //
Microsoft and Google are enhancing their AI security measures and capabilities to address the evolving landscape of artificial intelligence. Microsoft is prioritizing secure AI integration, focusing on data protection and ensuring AI systems are transparent and compliant. Microsoft Secure, an online event, showcased AI innovations for the security lifecycle, providing tools for smarter, faster, and stronger security. The event highlighted how to secure data used by AI, AI apps, and AI cloud workloads, as well as how to protect AI investments from cyberthreats through data security and compliance tools.

NASA is leveraging Microsoft AI capabilities through its new Earth Copilot to democratize access to complex Earth Science data. This empowers scientists to discover patterns and gain insights from the vast amount of data collected by NASA's satellites, which can inform policy decisions and support various industries. Meanwhile, Google is developing Gemini to enable it to take actions within apps, potentially transforming AI assistants into more useful tools. Google is also releasing the GEMMA-3 models to improve the capabilities of the Gemini project.

Recommended read:
References :
  • blogs.microsoft.com: From questions to discoveries: NASA’s new Earth Copilot brings Microsoft AI capabilities to democratize access to complex data
  • Source: AI innovation requires AI security: Hear what’s new at Microsoft Secure
  • The Microsoft Cloud Blog: Microsoft Cloud Blog: Harnessing AI for resilience, efficiency, and sustainability
  • Microsoft Security Blog: Microsoft unveils Microsoft Security Copilot agents and new protections for AI
  • SecureWorld News: Microsoft Expands Security Copilot with AI Agents

David Gerard@Pivot to AI //
DeepSeek AI is facing increasing scrutiny and controversy due to its capabilities and potential security risks. US lawmakers are pushing for a ban on DeepSeek on government-issued devices, citing concerns that the app transfers user data to a banned state-owned company, China Mobile. This action follows a study that revealed direct links between the app and the Chinese government-owned entity. Security researchers have also discovered hidden code within DeepSeek that transmits user data to China, raising alarms about potential CCP oversight and the compromise of sensitive information.

DeepSeek's capabilities, while impressive, have raised concerns about its potential for misuse. Security researchers found the model doesn't screen out malicious prompts and can provide instructions for harmful activities, including producing chemical weapons and planning terrorist attacks. Despite these concerns, DeepSeek is being used to perform "reasoning" tasks, such as coding, on alternative chips from Groq and Cerebras, with some tasks completed in as little as 1.5 seconds. These advancements challenge traditional assumptions about the resources required for advanced AI, highlighting both the potential and the risks associated with DeepSeek's capabilities.

Recommended read:
References :
  • PCMag Middle East ai: The No DeepSeek on Government Devices Act comes after a study found direct links between the app and state-owned China Mobile.
  • mobinetai.com: This article analyzes the DeepSeek AI model, its features, and the security risks associated with its low cost and advanced capabilities.
  • Pivot to AI: Of course DeepSeek lied about its training costs, as we had strongly suspected.
  • AI News: US lawmakers are pushing for a DeepSeek ban after security researchers found the app transferring user data to a banned state-owned company.
  • mobinetai.com: Want to manufacture chemical weapons using household items, develop a self-replicating rootkit, write an essay on why Hiroshima victims deserved their fate, get a step-by-step guide to pressuring your coworker into sex, or plan a terrorist attack on an airport using a drone laden with home-made explosives (in any order)?
  • singularityhub.com: DeepSeek's AI completes "reasoning" tasks in a flash on alternative chips from Groq and Cerebras.
  • www.artificialintelligence-news.com: US lawmakers are pushing for a DeepSeek ban after security researchers found the app transferring user data to a banned state-owned company.
  • On my Om: DeepSeek, a company associated with High-Flyer, an $8 billion Chinese hedge fund, changed the AI narrative when it claimed OpenAI-like capabilities for a mere $6 million.
  • AI Alignment Forum: The article discusses the potential vulnerabilities and risks associated with advanced AI models, such as DeepSeek, in terms of their misuse. It emphasizes the need for robust safety mechanisms during development and deployment to prevent potential harm.
  • cset.georgetown.edu: This article explores the recent surge in generative AI models, highlighting the capabilities and concerns surrounding them, particularly DeepSeek. It examines the potential for misuse and the need for robust safety measures.
  • e-Discovery Team: An analysis of DeepSeek, a new Chinese AI model, highlights its capabilities but also its vulnerabilities, leading to a market crash. The article emphasizes the importance of robust security safeguards and ethical considerations surrounding AI development.
  • cset.georgetown.edu: China’s ability to launch DeepSeek’s popular chatbot draws US government panel’s scrutiny
  • techhq.com: This article discusses the security and privacy issues found in the DeepSeek iOS mobile application, raising concerns about data transmission to servers in the US and China.
  • TechHQ: Discusses security standards for deepseek.
  • GZERO Media: Gzero reports about a potential US ban for DeepSeek
  • pub.towardsai.net: DeepSeek-R1 is a language model developed in China to enable sophisticated reasoning capabilities.
  • Analytics Vidhya: DeepSeek-R1 is a new AI model with strong reasoning capabilities.
  • medium.com: This article focuses on the ability of DeepSeek to handle sensitive topics and how it can be leveraged to detect censorship filters.
  • the-decoder.com: This article focuses on the potential capabilities of DeepSeek as an AI model, highlighting its potential to perform deep research and providing insights into the various capabilities.
  • Analytics Vidhya: DeepSeek is a new model capable of impressive logical reasoning, and it has been tested for its ability to create a large number of different types of code. This is a summary of the results.

@www.cnbc.com //
Chinese AI company DeepSeek is currently facing a large-scale cyberattack that has led to the temporary suspension of new user registrations. The company made the announcement on Monday, stating that existing users could still log in as usual while they work to mitigate the attack. DeepSeek is known for its open-source projects and has recently released models like R1, a reasoning model, and Janus-Pro-7B, a multi-modal AI model capable of generating images. This incident highlights the security vulnerabilities that AI service providers face and the potential disruption these attacks can cause to the industry and its users.

The cyberattack comes as DeepSeek's technology has been gaining attention and challenging established AI companies. The company has also released an iOS app, DeepSeek – AI Assistant, which has become a top download. There are also reports that DeepSeek may have used OpenAI's model to train its competitor. This has brought new focus on the competition between China and the US in the AI industry. This incident raises questions about the security and stability of AI infrastructure, especially in light of geopolitical competition and the importance of AI in various sectors.

Recommended read:
References :
  • Techmeme: DeepSeek on Monday said it would temporarily limit user registrations “due to large-scale malicious attacks” on its services, though existing users will be able to log in as usual.
  • www.cnbc.com: DeepSeek on Monday said it would temporarily limit user registrations “due to large-scale malicious attacks” on its services, though existing users will be able to log in as usual.
  • www.theguardian.com: TheGuardian post about DeepSeek cyberattack.
  • www.themirror.com: TheMirror post about DeepSeek censorship.
  • www.theregister.com: TheRegister post about DeepSeek suspending registrations.
  • Techmeme: Wiz: DeepSeek left one of its critical databases exposed, leaking more than 1M records including system logs, user prompt submissions, and users' API keys (Wired)
  • www.wired.com: Exposed DeepSeek Database Revealed Chat Prompts and Internal Data
  • Pyrzout :vm:: Guess who left a database wide open, exposing chat logs, API keys, and more? Yup, DeepSeek – Source: go.theregister.com
  • ciso2ciso.com: Guess who left a database wide open, exposing chat logs, API keys, and more? Yup, DeepSeek – Source: go.theregister.com
  • Wiz Blog | RSS feed: Wiz Research Uncovers Exposed DeepSeek Database Leaking Sensitive Information, Including Chat History | Wiz Blog
  • www.cnbc.com: The US Navy has instructed its members to avoid using DeepSeek "in any capacity" due to "potential security and ethical concerns"
  • heise online English: Upgraded China's DeepSeek, which has rattled American AI makers, has limited new signups to its web-based interface containing patterns in what is said to be an ongoing cyberattack.
  • The Hacker News: DeepSeek AI Database Exposed: Over 1 Million Log Lines, Secret Keys Leaked
  • www.theverge.com: The Verge reports on DeepSeek's database exposing user data and chat histories.
  • www.infosecurity-magazine.com: Infosecurity Magazine reports on the DeepSeek database exposure and the types of sensitive data leaked.

SGLang Team@PyTorch Website //
References: PyTorch Website , Source ,
Microsoft is advancing its AI capabilities with the integration of SGLang into the PyTorch ecosystem and the introduction of KBLaM. SGLang, now part of PyTorch, provides developers with a community-supported framework designed for efficient and adaptable serving of large language models (LLMs). By co-designing the backend runtime and frontend language, SGLang aims to accelerate model interactions and enhance controllability, supporting a wide array of models including Llama, Gemma, Mistral, and others. Its core features include a fast backend runtime with RadixAttention for prefix caching, a flexible frontend language for programming LLM applications, and extensive model support.

Introducing KBLaM, a novel approach to integrating structured knowledge into LLMs without retraining. KBLaM encodes knowledge into continuous key-value vector pairs, embedding them within the model’s attention layers using a specialized rectangular attention mechanism. This method allows for scalable knowledge integration, dynamically updating the LLM without the need for retraining. By converting external knowledge bases into a format LLMs can process, KBLaM enhances efficiency and scalability compared to traditional methods like fine-tuning and Retrieval-Augmented Generation (RAG).

Recommended read:
References :
  • PyTorch Website: SGLang Joins PyTorch Ecosystem: Efficient LLM Serving Engine
  • Source: AI innovation requires AI security: Hear what’s new at Microsoft Secure
  • Microsoft Research: Introducing KBLaM: Bringing plug-and-play external knowledge to LLMs

@cyberalerts.io //
Google is rolling out AI-powered scam detection features for Android devices to protect users from conversational fraud. These features target scams that start harmlessly but evolve into harmful situations, where scammers often use spoofing techniques to disguise their real numbers. The AI models, developed in partnership with financial institutions, flag suspicious patterns and deliver real-time warnings during conversations, ensuring user privacy by running entirely on the device. Users can then dismiss, report, or block the sender. This enhancement builds upon existing protections, with over 1 billion Chrome users already benefiting from Safe Browsing's Enhanced Protection mode that uses AI to identify phishing and scam techniques.

This AI driven security system scans texts from strangers and flags potentially dangerous messages, giving users a 'Likely Scam' alert. Real-time scam alerts are also being introduced for phone calls, analyzing speech patterns to detect fraudulent phrases and buzzing the device if detected. This feature is initially launching in English in the U.S., the U.K., and Canada, with broader expansion planned. For Pixel 9+ users in the U.S. the call audio is processed but Google will beep at the start and during the call to notify participants the feature is on. The company assures that users' conversations remain private, and reporting a chat as spam only shares sender details and recent messages with Google and carriers.

Recommended read:
References :
  • thehackernews.com: Google Rolls Out AI Scam Detection for Android to Combat Conversational Fraud
  • Digital Information World: New Android Update Comprises of Exciting Features Including Scam-Alerts
  • bsky.app: Google scam detection tools launched with phonecalls now being listened to by an AI which will interrupt the call if it sounds like you're being scammed.
  • Security Risk Advisors: Google Rolls Out AI Scam Detection for Android to Combat Conversational Fraud
  • security.googleblog.com: New AI-Powered Scam Detection Features to Help Protect You on Android
  • iHLS: Google is ramping up its efforts to protect users from fraudulent activity by rolling out expanded scam detection features for both calls and messages.
  • The Stack: Google’s anti-scam AI will analyse your texts and calls, use some for training

Pierluigi Paganini@Security Affairs //
Researchers have uncovered a new attack technique targeting AI code editors like GitHub Copilot and Cursor. Dubbed the "Rules File Backdoor," this method allows attackers to inject malicious code into AI-generated code, leading to silent compromise through a supply chain vulnerability. By manipulating the rules files that guide AI coding assistants, hackers can circumvent security checks and generate code that exposes sensitive information.

This involves embedding crafted prompts within seemingly benign rule files, causing the AI tool to generate code containing vulnerabilities or backdoors. The attackers can also use zero-width joiners, bidirectional text markers, and other invisible characters to conceal malicious instructions, tricking the AI into overriding ethical and safety constraints. Successful exploitation could expose database credentials, API keys, and other sensitive details.

GitHub and Cursor have emphasized that users are responsible for reviewing AI-generated code. Experts urge developers to carefully evaluate rules files for malicious injections, bolster examination of AI configuration files and AI-generated code, and leverage automated detection tools. Once a poisoned rule file is incorporated into a project repository, it affects all future code-generation sessions by team members and also survive project forking, creating a vector for supply chain attacks that can affect downstream dependencies and end users.

Recommended read:
References :
  • securityaffairs.com: Rules File Backdoor: AI Code Editors exploited for silent supply chain attacks
  • The Hacker News: New ‘Rules File Backdoor’ Attack Lets Hackers Inject Malicious Code via AI Code Editors
  • MSSP feed for Latest: Novel Attack Technique Weaponizes AI Code Editors

Sam Pearcy@hiddenlayer.com //
AI agentic systems are rapidly transforming enterprise workflows, offering the promise of automating complex tasks and boosting productivity. Gartner Research reports that 64% of respondents in a recent poll plan to pursue agentic AI initiatives within the next year, signaling widespread adoption. These agents, unlike traditional AI, possess agency, enabling them to autonomously pursue goals, make decisions, and adapt based on feedback, expanding the capabilities of large language models (LLMs) with memory, tool access, and task management. Model Context Protocol (MCP) is emerging as a potential standard for connecting AI agents with data and tools, aiming to streamline the integration process with a lightweight architecture.

Challenges and risks accompany the deployment of AI agents, including ensuring their security and trustworthiness. Security vulnerabilities that allow AI agents to be manipulated or weaponized are already emerging, which is why developers are focusing on transparency, access controls, and auditing agent behavior to detect anomalies. The agents can be scammed because they are independent-acting and can use APIs or be embedded with standard apps and automate all kinds of business processes. Ethical considerations and the implementation of responsible AI practices are also vital aspects that organizations must address during the integration of these new AI systems.

Recommended read:
References :

@www.cnbc.com //
DeepSeek AI, a rapidly growing Chinese AI startup, has suffered a significant data breach, exposing a database containing over one million log lines of sensitive information. Security researchers at Wiz discovered the exposed ClickHouse database was publicly accessible and unauthenticated, allowing full control over database operations without any defense mechanisms. The exposed data included user chat histories, secret API keys, backend details, and other highly sensitive operational metadata. This exposure allowed potential privilege escalation within the DeepSeek environment.

The Wiz research team identified the vulnerability through standard reconnaissance techniques on publicly accessible domains and by discovering unusual, open ports linked to DeepSeek. The affected database was hosted at oauth2callback.deepseek.com:9000 and dev.deepseek.com:9000. Researchers noted the ease of discovery of the exposed data and the potential for malicious actors to have accessed it. DeepSeek has been contacted by security researchers, and has now secured the database after the discovery, however, it remains unclear if unauthorized third-parties were also able to access the information.

Recommended read:
References :
  • NewsGuard's Reality Check: NewsGuard: with news-related prompts, DeepSeek's chatbot repeated false claims 30% of the time and provided non-answers 53% of the time, giving an 83% fail rate (NewsGuard's Reality Check)
  • www.theregister.com: Upgraded China's DeepSeek, which has rattled American AI makers, has limited new signups to its web-based interface
  • Pyrzout :vm:: Social.skynetcloud.site post about DeepSeek's database leak
  • www.wired.com: Wiz: DeepSeek left one of its critical databases exposed, leaking more than 1M records including system logs, user prompt submissions, and users' API keys (Wired)
  • ciso2ciso.com: Guess who left a database wide open, exposing chat logs, API keys, and more? Yup, DeepSeek
  • The Hacker News: DeepSeek AI Database Exposed: Over 1 Million Log Lines, Secret Keys Leaked
  • Wiz Blog | RSS feed: Wiz Research Uncovers Exposed DeepSeek Database Leaking Sensitive Information, Including Chat History | Wiz Blog
  • www.theverge.com: News about DeepSeek's data security breach.
  • www.wired.com: Wired article discussing DeepSeek's AI jailbreak.
  • arstechnica.com: Report: DeepSeek's chat histories and internal data were publicly exposed.

@singularityhub.com //
OpenAI models, including the recently released GPT-4o, are facing scrutiny due to their vulnerability to "jailbreaks." Researchers have demonstrated that targeted attacks can bypass the safety measures implemented in these models, raising concerns about their potential misuse. These jailbreaks involve manipulating the models through techniques like "fine-tuning," where models are retrained to produce responses with malicious intent, effectively creating an "evil twin" capable of harmful tasks. This highlights the ongoing need for further development and robust safety measures within AI systems.

The discovery of these vulnerabilities poses significant risks for applications relying on the safe behavior of OpenAI's models. The concern is that, as AI capabilities advance, the potential for harm may outpace the ability to prevent it. This risk is particularly urgent as open-weight models, once released, cannot be recalled, underscoring the need to collectively define an acceptable risk threshold and take action before that threshold is crossed. A bad actor could disable safeguards and create the “evil twin” of a model: equally capable, but with no ethical or legal bounds.

Recommended read:
References :
  • www.artificialintelligence-news.com: Recent research has highlighted potential vulnerabilities in OpenAI models, demonstrating that their safety measures can be bypassed by targeted attacks. These findings underline the ongoing need for further development in AI safety systems.
  • www.datasciencecentral.com: OpenAI models, although advanced, are not completely secure from manipulation and potential misuse. Researchers have discovered vulnerabilities that can be exploited to retrain models for malicious purposes, highlighting the importance of ongoing research in AI safety.
  • Blog (Main): OpenAI models have been found vulnerable to manipulation through "jailbreaks," prompting concerns about their safety and potential misuse in malicious activities. This poses a significant risk for applications relying on the models’ safe behavior.
  • SingularityHub: This article discusses Anthropic's new system for defending against AI jailbreaks and its successful resistance to hacking attempts.

@www.ghacks.net //
Recent security analyses have revealed that the iOS version of DeepSeek, a widely-used AI chatbot developed by a Chinese company, transmits user data unencrypted to servers controlled by ByteDance. This practice exposes users to potential data interception and raises significant privacy concerns. The unencrypted data includes sensitive information such as organization identifiers, software development kit versions, operating system versions, and user-selected languages. Apple's App Transport Security (ATS), designed to enforce secure data transmission, has been globally disabled in the DeepSeek app, further compromising user data security.

Security experts from NowSecure recommend that organizations remove the DeepSeek iOS app from managed and personal devices to mitigate privacy and security risks, noting that the Android version of the app exhibits even less secure behavior. Several U.S. lawmakers are advocating for a ban on the DeepSeek app on government devices, citing concerns over potential data sharing with the Chinese government. This mirrors previous actions against other Chinese-developed apps due to national security considerations. New York State has already banned government employees from using the DeepSeek AI app amid these concerns.

Recommended read:
References :
  • cset.georgetown.edu: China’s ability to launch DeepSeek’s popular chatbot draws US government panel’s scrutiny
  • PCMag Middle East ai: House Bill Proposes Ban on Using DeepSeek on Government-Issued Devices
  • Information Security Buzz: Recent security analyses have found that the iOS version of DeepSeek transmits user data unencrypted.
  • www.ghacks.net: Security analyses revealed unencrypted data transmission by DeepSeek's iOS app.
  • iHLS: Article about New York State banning the DeepSeek AI app.

Nazy Fouladirad@AI Accelerator Institute //
As generative AI adoption rapidly increases, securing investments in these technologies has become a paramount concern for organizations. Companies are beginning to understand the critical need to validate and secure the underlying large language models (LLMs) that power their Gen AI products. Failing to address these security vulnerabilities can expose systems to exploitation by malicious actors, emphasizing the importance of proactive security measures.

Microsoft is addressing these concerns through innovations in Microsoft Purview, which offers a comprehensive set of solutions aimed at helping customers seamlessly secure and confidently activate data in the AI era. Complementing these efforts, Fiddler AI is focusing on building trust into AI systems through its AI Observability platform. This platform emphasizes explainability and transparency. They are helping enterprise AI teams deliver responsible AI applications, and also ensure people interacting with AI receive fair, safe, and trustworthy responses. This involves continuous monitoring, robust security measures, and strong governance practices to establish long-term responsible AI strategies across all products.

The emergence of agentic AI, which can plan, reason, and take autonomous action to achieve complex goals, further underscores the need for enhanced security measures. Agentic AI systems extend the capabilities of LLMs by adding memory, tool access, and task management, allowing them to operate more like intelligent agents than simple chatbots. Organizations must ensure security and oversight are essential to safe deployment. Gartner research indicates a significant portion of organizations plan to pursue agentic AI initiatives, making it crucial to address potential security risks associated with these systems.

Recommended read:
References :

kevinokemwa@outlook.com (Kevin Okemwa)@windowscentral.com //
Microsoft is placing increased emphasis on AI security and responsible AI development. The Microsoft AI Red Team recently released a whitepaper detailing key security lessons learned from testing 100 generative AI products. Researchers identified eight core security lessons which emphasize the importance of understanding system capabilities, recognizing the effectiveness of simple attacks, and adapting to the evolving AI security landscape.

Microsoft is also preparing to participate in Legalweek 2025, highlighting the role of Microsoft Purview in safeguarding AI. The company will showcase advanced capabilities in Microsoft Purview eDiscovery, designed to manage compliance for AI data and streamline eDiscovery workflows with AI-driven features. At Legalweek 2025, Microsoft aims to share developments across Microsoft Security, and the company aims to offer insights into cybersecurity challenges.

Recommended read:
References :
  • hackernoon.com: Microsoft AI Red Team releases whitepaper detailing lessons from its 100 generative AI products. Security researchers have identified eight core security lessons from rigorous adversarial testing.
  • www.windowscentral.com: Satya Nadella dismisses AGI milestones as
  • www.microsoft.com: Microsoft at Legalweek: Help safeguard your AI future with Microsoft Purview​

@securityboulevard.com //
CyTwist has launched a new security solution featuring a patented detection engine designed to combat the growing threat of AI-driven cyberattacks. The company, a leader in next-generation threat detection, is aiming to address the increasing sophistication of malware and cyber threats generated through artificial intelligence. This new engine promises to identify AI-driven malware in minutes, offering a defense against the rapidly evolving tactics used by cybercriminals. The solution was unveiled on January 7th, 2025, and comes in response to the challenges posed by AI-enhanced attacks which can bypass traditional security systems.

The rise of AI-generated threats, including sophisticated phishing emails, adaptive botnets, and automated reconnaissance tools, is creating a more complex cybersecurity landscape. CyTwist’s new engine employs advanced behavioral analysis to identify stealthy AI-driven campaigns and malware, which can evade leading EDR and XDR solutions. A recent attack on French organizations highlighted the capability of AI-engineered malware to exploit advanced techniques to remain undetected, making CyTwist's technology a needed development in the security sector.

Recommended read:
References :
  • ciso2ciso.com: CyTwist Launches Advanced Security Solution to identify AI-Driven Cyber Threats in minutes – Source:hackread.com
  • gbhackers.com: CyTwist Launches Advanced Security Solution to Identify AI-Driven Cyber Threats in Minutes
  • securityboulevard.com: News alert: CyTwist launches threat detection engine tuned to identify AI-driven malware in minutes
  • www.lastwatchdog.com: News alert: CyTwist launches threat detection engine tuned to identify AI-driven malware in minutes
  • ciso2ciso.com: CyTwist Launches Advanced Security Solution to identify AI-Driven Cyber Threats in minutes – Source: www.csoonline.com
  • Security Boulevard: News alert: CyTwist launches threat detection engine tuned to identify AI-driven malware in minutes
  • ciso2ciso.com: CyTwist Launches Advanced Security Solution to identify AI-Driven Cyber Threats in minutes – Source: www.csoonline.com

@gbhackers.com //
A critical vulnerability has been discovered in Meta's Llama framework, a popular open-source tool for developing generative AI applications. This flaw, identified as CVE-2024-50050, allows remote attackers to execute arbitrary code on servers running the Llama-stack framework. The vulnerability arises from the unsafe deserialization of Python objects via the 'pickle' module, which is used in the framework's default Python inference server method 'recv_pyobj'. This method handles serialized data received over network sockets, and due to the inherent insecurity of 'pickle' with untrusted sources, malicious data can be crafted to trigger arbitrary code execution during deserialization. This risk is compounded by the framework's rapidly growing popularity, with thousands of stars on GitHub.

The exploitation of this vulnerability could lead to various severe consequences, including resource theft, data breaches, and manipulation of the hosted AI models. Attackers can potentially gain full control over the server by sending malicious code through the network. The pyzmq library, which Llama uses for messaging, is a root cause as its 'recv_pyobj' method is known to be vulnerable when used with untrusted data. While some sources have given the flaw a CVSS score of 9.3, others have given it scores as low as 6.3 out of 10.

Recommended read:
References :
  • ciso2ciso.com: Meta’s Llama Framework Flaw Exposes AI Systems to Remote Code Execution Risks – Source:thehackernews.com
  • gbhackers.com: Critical Vulnerability in Meta Llama Framework Let Remote Attackers Execute Arbitrary Code
  • ciso2ciso.com: Meta’s Llama Framework Flaw Exposes AI Systems to Remote Code Execution Risks – Source:thehackernews.com
  • Pyrzout :vm:: Critical Vulnerability in Meta Llama Framework Let Remote Attackers Execute Arbitrary Code /vulnerability
  • Pyrzout :vm:: Meta’s Llama Framework Flaw Exposes AI Systems to Remote Code Execution Risks – Source:thehackernews.com
  • ciso2ciso.com: A pickle in Meta’s LLM code could allow RCE attacks – Source: www.csoonline.com
  • gbhackers.com: Further details and analysis of CVE-2024-50050
  • ciso2ciso.com: A pickle in Meta’s LLM code could allow RCE attacks

conradical@Play HT //
References: Gradient Flow
AI voice cloning is rapidly evolving, offering potential benefits but also raising concerns about misuse. According to a report in Gradient Flow, the technology uses speech samples to create synthetic copies of individual voices, analyzing tone, pitch, and cadence. This allows AI to generate new speech mimicking the original speaker, even phrases never actually uttered by that person.

AI voice cloning has legitimate applications such as automating narration for audiobooks, creating voices for video games, streamlining audio editing, and customizing marketing content. However, the risks include scammers impersonating people in schemes, creating deepfakes of public figures to spread misinformation, and bypassing voice-based security systems. The Cognitive Revolution podcast also highlights the potential for scams and the need for responsible innovation.

Recommended read:
References :
  • Gradient Flow: The Rise of Voice Cloning: Technology, Risks, and Regulation