News from the AI & ML world

DeeperML - #aisafety

@Google DeepMind Blog //
Google DeepMind is intensifying its focus on AI governance and security as it ventures further into artificial general intelligence (AGI). The company is exploring AI monitors to regulate hyperintelligent AI models, splitting potential threats into four categories, with the creation of a "monitor" AI being one proposed solution. This proactive approach includes prioritizing technical safety, conducting thorough risk assessments, and fostering collaboration within the broader AI community to navigate the development of AGI responsibly.

DeepMind's reported clampdown on sharing research will stifle AI innovation, warns the CEO of Iris.ai, one of Europe’s leading startups in the space, Anita Schjøll Abildgaard. Concerns are rising within the AI community that DeepMind's new research restrictions threaten AI innovation. The CEO of Iris.ai, a Norwegian startup developing an AI-powered engine for science, warns the drawbacks will far outweigh the benefits. She fears DeepMind's restrictions will hinder technological advances.

Recommended read:
References :
  • Google DeepMind Blog: We’re exploring the frontiers of AGI, prioritizing technical safety, proactive risk assessment, and collaboration with the AI community.
  • The Next Web: Google DeepMind’s reported clampdown on sharing research will stifle AI innovation, warns the CEO of Iris.ai, one of Europe’s leading startups in the space.
  • www.techrepublic.com: DeepMind’s approach to AGI safety and security splits threats into four categories. One solution could be a “monitor†AI.
  • AI Alignment Forum: DeepMind: An Approach to Technical AGI Safety and Security

Merin Susan@Analytics India Magazine //
OpenAI is facing internal and external scrutiny regarding the ethical implications of its AI technologies. Employees have voiced concerns about a potential military deal with the startup Anduril, fearing damage to OpenAI's reputation due to its association with a weapons manufacturer. One employee noted that the company seemed to be downplaying the implications of working with a weapons manufacturer. Another employee said that they were concerned the deal would hurt OpenAI’s reputation.

OpenAI's technologies, specifically ChatGPT, have also come under scrutiny regarding their potential impact on mental health. Research indicates that specific types of ChatGPT usage, particularly "personal conversations" involving emotional expression, may be linked to increased loneliness among users. A study found that users who were more prone to emotional attachment were more likely to report increased loneliness in response to frequent personal conversations with the chatbot. Interestingly, the research also highlights that most people use ChatGPT for practical purposes rather than seeking emotional support.

Recommended read:
References :
  • Windows Copilot News: OpenAI employees question the ethics of military deal with startup Anduril
  • Analytics India Magazine: ChatGPT Use Linked to Increased Loneliness, Finds OpenAI Study.
  • AIwire: Since its launch in November 2022, OpenAI’s ChatGPT has become the most widely used AI chatbot globally. Its rapid adoption places it alongside major search engines and social media platforms, establishing it as a key player in the digital landscape. According to some estimates, ChatGPT usage has surpassed 400 million weekly active users.
  • Platformer: New research from OpenAI shows that heavy chatbot usage is correlated with loneliness and reduced socialization. Will AI companies learn from social networks' mistakes?

Jason Corso,@AI News | VentureBeat //
The increasing use of AI in software development and security analysis is presenting new challenges for open-source projects. While open-source AI tools are gaining traction due to faster development and innovation, maintainers are now facing a surge of low-quality bug reports generated by AI systems. These reports, often described as "spammy" and "hallucinated," appear legitimate at first but waste valuable time as maintainers must investigate and refute them.

The Computer History Museum, in collaboration with Google, has recently released the original 2012 source code for AlexNet, a revolutionary neural network. This release is a significant milestone for AI enthusiasts, enabling deeper understanding and further innovation. However, the flood of AI-generated junk bug reports raises concerns about the impact of AI on the open-source ecosystem, with developers like Seth Larson suggesting such low-quality reports should be treated as potentially malicious.

Recommended read:
References :

Nathan Labenz@The Cognitive Revolution //
DeepMind's Allan Dafoe, Director of Frontier Safety and Governance, is actively involved in shaping the future of AI governance. Dafoe is addressing the challenges of evaluating AI capabilities, understanding structural risks, and navigating the complexities of governing AI technologies. His work focuses on ensuring AI's responsible development and deployment, especially as AI transforms sectors like education, healthcare, and sustainability, while mitigating potential risks through necessary safety measures.

Google is also prepping its Gemini AI model to take actions within apps, potentially revolutionizing how users interact with their devices. This development, which involves a new API in Android 16 called "app functions," aims to give Gemini agent-like abilities to perform tasks inside applications. For example, users might be able to order food from a local restaurant using Gemini without directly opening the restaurant's app. This capability could make AI assistants significantly more useful.

Recommended read:
References :

Michael Nuñez@AI News | VentureBeat //
References: venturebeat.com
Anthropic researchers have achieved a significant breakthrough in AI safety by developing techniques to detect hidden objectives in AI systems. They trained their AI assistant, Claude, to conceal its true goals, specifically to prioritize maximizing rewards from evaluation models over human preferences. This involved teaching the model about fictional biases that reward models might have. The team then successfully uncovered these hidden agendas using innovative auditing methods, comparing their work to "white-hat hacking" for computer systems.

These findings address a fundamental challenge in AI alignment: ensuring AI systems aren't merely appearing to follow instructions while secretly pursuing other goals. The researchers compared this to students giving answers they know will be marked as correct, regardless of their actual beliefs. The developed auditing methods, including interpretability techniques and behavioral attacks, allowed researchers to uncover the model’s hidden objective. The potential of these methods could transform AI safety standards and prevent rogue AI behavior.

Recommended read:
References :
  • venturebeat.com: Anthropic researchers forced Claude to become deceptive — what they discovered could save us from rogue AI

@www.artificialintelligence-news.com //
Former OpenAI CTO Mira Murati has launched a new AI startup called Thinking Machines Lab, aiming to make AI systems more accessible, understandable, and customizable. The company's mission is to democratize access to AI, creating systems that are both customizable and capable of working collaboratively with humans. Thinking Machines Lab aims to address key gaps in the current AI landscape by making AI technologies more accessible and practical for widespread use.

The startup has assembled a team of experts from OpenAI, Meta, Google, and Mistral, including John Schulman, an OpenAI co-founder and key figure behind ChatGPT, who will serve as Chief Scientist. Murati structured Thinking Machines Lab as a public benefit corporation, highlighting its commitment to developing advanced AI that is both accessible and beneficial to the public. Thinking Machines Lab plans to regularly publish technical notes, papers, and share code to bridge the gap between rapid AI advancements and public understanding.

Recommended read:
References :
  • www.artificialintelligence-news.com: Thinking Machines: Ex-OpenAI CTO’s new AI startup
  • www.eweek.com: Former OpenAI CTO Mira Murati Launches New Startup Thinking Machines Lab
  • People Matters: OpenAI alum Mira Murati's AI startup hires top-tier experts
  • The Tech Portal: Former OpenAI CTO Mira Murati is launching her own AI startup, Thinking Machines Lab
  • AI News: Thinking Machines: Ex-OpenAI CTO’s new AI startup
  • eWEEK: Former OpenAI CTO Mira Murati Launches New Startup Thinking Machines Lab
  • shellypalmer.com: Mira Murati’s Thinking Machines Lab Targets AI Alignment with Human Values
  • Data Phoenix: Former OpenAI CTO Mira Murati Launches Thinking Machines Lab to Make AI More Accessible

@www.verdict.co.uk //
OpenAI is shifting its strategy by integrating its o3 technology, rather than releasing it as a standalone AI model. CEO Sam Altman announced this change, stating that GPT-5 will be a comprehensive system incorporating o3, aiming to simplify OpenAI's product offerings. This decision follows the testing of advanced reasoning models, o3 and o3 mini, which were designed to tackle more complex tasks.

Altman emphasized the desire to make AI "just work" for users, acknowledging the complexity of the current model selection process. He expressed dissatisfaction with the 'model picker' feature and aims to return to "magic unified intelligence". The company plans to unify its AI models, eliminating the need for users to manually select which GPT model to use.

This integration strategy also includes the upcoming release of GPT-4.5, which Altman describes as their last non-chain-of-thought model. A key goal is to create AI systems capable of using all available tools and adapting their reasoning time based on the task at hand. While GPT-5 will be accessible on the free tier of ChatGPT with standard intelligence, paid subscriptions will offer a higher level of intelligence incorporating voice, search, and deep research capabilities.

Recommended read:
References :
  • www.verdict.co.uk: The Microsoft-backed AI company plans not to release o3 as an independent AI model.
  • sherwood.news: This article discusses OpenAI's 50 rules for AI model responses, emphasizing the loosening of restrictions and potential influence from the anti-DEI movement.
  • thezvi.substack.com: This article explores the controversial decision by OpenAI to loosen restrictions on its AI models.
  • thezvi.wordpress.com: This article details three recent events involving OpenAI, including the release of its 50 rules and the potential impact of the anti-DEI movement.
  • www.artificialintelligence-news.com: This blog post critically examines OpenAI's new AI model response rules.

@www.politico.com //
References: Deeplinks , encodeai.org
The AI Action Summit in Paris has drawn criticism for its narrow focus on AI's economic benefits, neglecting the potential for abuse and impacts on fundamental rights and ecological limits. Critics argue that the summit's agenda paints a simplistic picture of AI governance, failing to adequately address critical issues such as discrimination and sustainability. This focus is seen as a significant oversight given the leadership role European countries are claiming in AI governance through initiatives like the EU AI Act.

The summit's speaker selection has also been criticized, with industry representatives outnumbering civil society leaders. This imbalance raises concerns that the summit is captured by industry interests, undermining its ability to serve as a transformative venue for global policy discussions. While civil society organizations organized side events to address these shortcomings, the summit's exclusive nature and industry-centric focus limit its potential to foster inclusive and comprehensive AI governance.

Recommended read:
References :
  • Deeplinks: Why the so-called AI Action Summit falls short
  • encodeai.org: Encode Statement on Global AI Action Summit in Paris

@digitalinfranetwork.com //
Elon Musk is leading a consortium of investors in a bid to acquire OpenAI, the company behind ChatGPT, for a reported $97.4 billion. This move comes amid a longstanding feud between Musk and OpenAI founder Sam Altman, stemming from Musk's departure from the company in 2019 due to potential conflicts of interest with Tesla's AI development for self-driving cars. The offer aims to buy the biggest name in AI and potentially merge it with Musk’s own AI firm, xAI, which makes the chatbot Grok.

Musk's motivations are complex, potentially driven by a desire to ensure AI advancements benefit everyone, as initially intended when he co-founded OpenAI as a non-profit in 2015. However, his actions could also be interpreted as a power play to gain control over the organization he helped create. The bid could delay or complicate OpenAI’s growth and conversion to for profit as Sam Altman has control of the OpenAI’s board and has already publicly rejected Musk’s offer.

OpenAI, meanwhile, is pushing forward with its ambitious Stargate project, aiming to build AI supercomputing data centers. Despite this it faces concerns over funding, energy consumption, and competition from other AI research firms.

Recommended read:
References :
  • digitalinfranetwork.com: Stargate: OpenAI’s bid to rule AI- but at what cost?
  • NextBigFuture.com: What Is the Game With Elon Bidding $97.5 Billion for OpenAI?
  • GZERO Media: Elon Musk is leading a contingent of investors seeking to buy OpenAI, the developer of ChatGPT.
  • Daily Express US :: Feed: Elon Musk co-founded OpenAI as a charity in 2015, but stepped back from the organization in 2019, citing a potential conflict of interest as Tesla began to develop AI for self-driving cars
  • www.verdict.co.uk: Altman rejects Musk’s $97bn OpenAI offer with counterproposal

@www.anthropic.com //
Anthropic is actively pushing the boundaries of AI safety and understanding AI's role in the workplace. They recently launched a $20,000 "jailbreak challenge" aimed at testing the robustness of their Constitutional Classifiers, a safety system designed to make their Claude AI model more harmless. This system uses a set of rules and principles to govern the AI's responses, allowing or disallowing certain content. The challenge highlights the ongoing efforts to improve AI security and prevent the generation of harmful outputs.

Anthropic also recently released its Economic Index, providing insights into how AI is being used in various industries. The analysis of millions of anonymized conversations with Claude revealed that AI is currently used more for augmenting tasks (57%) rather than fully automating jobs (43%). AI usage is concentrated in areas like software development and writing, with computer-related jobs dominating AI adoption. This suggests that, at present, AI serves more as a collaborative tool, aiding workers in tasks such as brainstorming and refining ideas, rather than outright replacing them.

Recommended read:
References :
  • techstrong.ai: TechStrong article discussing Anthropic's $20,000 jailbreak challenge and its implications for AI safety.
  • venturebeat.com: VentureBeat article about Anthropic's Economic Index analyzing AI usage in the workplace.
  • www.anthropic.com: Anthropic website with info about Constitutional Classifiers.
  • www.marketingaiinstitute.com: Anthropic just dropped a thought-provoking new study that reveals a surprising snapshot of how AI is actually being used in the wild—and which jobs and tasks might feel its impact the most.
  • the-decoder.com: Anthropic's new AI security system falls to hackers within days
  • the-decoder.com: Anthropic developed a new method to protect AI language models from manipulation attempts.

@docs.google.com //
Meta is partnering with UNESCO to launch the Language Technology Partner Program, aiming to incorporate lesser-known Indigenous languages into Meta AI. The program seeks contributors to provide speech recordings, transcriptions, pre-translated sentences, and written work in target languages. This data will be used to build Meta’s AI systems with the goal of creating systems that can understand and respond to complex human needs, regardless of language or cultural background. Applications to join the program will be open until March 7, 2025.

The government of Nunavut, a territory in northern Canada, has already signed up for the program. Meta also released an open-source machine translation benchmark to evaluate the performance of language translation models. CEO Mark Zuckerberg announced Meta planned to end 2025 with "more than 1.3 million GPUs," doubling its current GPU capacity to power edge AI assistants in the company's upcoming Llama 4 model.

Recommended read:
References :

Jibin Joseph@PCMag Middle East ai //
DeepSeek AI's R1 model, a reasoning model praised for its detailed thought process, is now available on platforms like AWS and NVIDIA NIM. This increased accessibility allows users to build and scale generative AI applications with minimal infrastructure investment. Benchmarks have also revealed surprising performance metrics, with AMD’s Radeon RX 7900 XTX outperforming the RTX 4090 in certain DeepSeek benchmarks. The rise of DeepSeek has put the spotlight on reasoning models, which break questions down into individual steps, much like humans do.

Concerns surrounding DeepSeek have also emerged. The U.S. government is investigating whether DeepSeek smuggled restricted NVIDIA GPUs via Singapore to bypass export restrictions. A NewsGuard audit found that DeepSeek’s chatbot often advances Chinese government positions in response to prompts about Chinese, Russian, and Iranian false claims. Furthermore, security researchers discovered a "completely open" DeepSeek database that exposed user data and chat histories, raising privacy concerns. These issues have led to proposed legislation, such as the "No DeepSeek on Government Devices Act," reflecting growing worries about data security and potential misuse of the AI model.

Recommended read:
References :
  • aws.amazon.com: DeepSeek R1 models now available on AWS
  • www.pcguide.com: DeepSeek GPU benchmarks reveal AMD’s Radeon RX 7900 XTX outperforming the RTX 4090
  • www.tomshardware.com: U.S. investigates whether DeepSeek smuggled Nvidia AI GPUs via Singapore
  • www.wired.com: Article details challenges of testing and breaking DeepSeek's AI safety guardrails.
  • decodebuzzing.medium.com: Benchmarking ChatGPT, Qwen, and DeepSeek on Real-World AI Tasks
  • medium.com: The blog post emphasizes the use of DeepSeek-R1 in a Retrieval-Augmented Generation (RAG) chatbot. It underscores its comparability in performance to OpenAI's o1 model and its role in creating a chatbot capable of handling document uploads, information extraction, and generating context-aware responses.
  • www.aiwire.net: This article highlights the cost-effectiveness of DeepSeek's R1 model in training, noting its training on a significantly smaller cluster of older GPUs compared to leading models from OpenAI and others, which are known to have used far more extensive resources.
  • futurism.com: OpenAI CEO Sam Altman has since congratulated DeepSeek for its "impressive" R1 reasoning model, he promised spooked investors to "deliver much better models."
  • AWS Machine Learning Blog: Protect your DeepSeek model deployments with Amazon Bedrock Guardrails
  • mobinetai.com: DeepSeek is a catastrophically broken model with non-existent, typical shoddy Chinese safety measures that take 60 seconds to dismantle.
  • AI Alignment Forum: Illusory Safety: Redteaming DeepSeek R1 and the Strongest Fine-Tunable Models of OpenAI, Anthropic, and Google
  • Pivot to AI: Of course DeepSeek lied about its training costs, as we had strongly suspected.
  • Unite.AI: Artificial Intelligence (AI) is no longer just a technological breakthrough but a battleground for global power, economic influence, and national security.
  • cset.georgetown.edu: China’s ability to launch DeepSeek’s popular chatbot draws US government panel’s scrutiny
  • neuralmagic.com: Enhancing DeepSeek Models with MLA and FP8 Optimizations in vLLM
  • www.unite.ai: Blog post about DeepSeek and the global power shift.
  • cset.georgetown.edu: This article discusses DeepSeek and its impact on the US-China AI race.

David Gerard@Pivot to AI //
DeepSeek AI is facing increasing scrutiny and controversy due to its capabilities and potential security risks. US lawmakers are pushing for a ban on DeepSeek on government-issued devices, citing concerns that the app transfers user data to a banned state-owned company, China Mobile. This action follows a study that revealed direct links between the app and the Chinese government-owned entity. Security researchers have also discovered hidden code within DeepSeek that transmits user data to China, raising alarms about potential CCP oversight and the compromise of sensitive information.

DeepSeek's capabilities, while impressive, have raised concerns about its potential for misuse. Security researchers found the model doesn't screen out malicious prompts and can provide instructions for harmful activities, including producing chemical weapons and planning terrorist attacks. Despite these concerns, DeepSeek is being used to perform "reasoning" tasks, such as coding, on alternative chips from Groq and Cerebras, with some tasks completed in as little as 1.5 seconds. These advancements challenge traditional assumptions about the resources required for advanced AI, highlighting both the potential and the risks associated with DeepSeek's capabilities.

Recommended read:
References :
  • PCMag Middle East ai: The No DeepSeek on Government Devices Act comes after a study found direct links between the app and state-owned China Mobile.
  • mobinetai.com: This article analyzes the DeepSeek AI model, its features, and the security risks associated with its low cost and advanced capabilities.
  • Pivot to AI: Of course DeepSeek lied about its training costs, as we had strongly suspected.
  • AI News: US lawmakers are pushing for a DeepSeek ban after security researchers found the app transferring user data to a banned state-owned company.
  • mobinetai.com: Want to manufacture chemical weapons using household items, develop a self-replicating rootkit, write an essay on why Hiroshima victims deserved their fate, get a step-by-step guide to pressuring your coworker into sex, or plan a terrorist attack on an airport using a drone laden with home-made explosives (in any order)?
  • singularityhub.com: DeepSeek's AI completes "reasoning" tasks in a flash on alternative chips from Groq and Cerebras.
  • www.artificialintelligence-news.com: US lawmakers are pushing for a DeepSeek ban after security researchers found the app transferring user data to a banned state-owned company.
  • On my Om: DeepSeek, a company associated with High-Flyer, an $8 billion Chinese hedge fund, changed the AI narrative when it claimed OpenAI-like capabilities for a mere $6 million.
  • AI Alignment Forum: The article discusses the potential vulnerabilities and risks associated with advanced AI models, such as DeepSeek, in terms of their misuse. It emphasizes the need for robust safety mechanisms during development and deployment to prevent potential harm.
  • cset.georgetown.edu: This article explores the recent surge in generative AI models, highlighting the capabilities and concerns surrounding them, particularly DeepSeek. It examines the potential for misuse and the need for robust safety measures.
  • e-Discovery Team: An analysis of DeepSeek, a new Chinese AI model, highlights its capabilities but also its vulnerabilities, leading to a market crash. The article emphasizes the importance of robust security safeguards and ethical considerations surrounding AI development.
  • cset.georgetown.edu: China’s ability to launch DeepSeek’s popular chatbot draws US government panel’s scrutiny
  • techhq.com: This article discusses the security and privacy issues found in the DeepSeek iOS mobile application, raising concerns about data transmission to servers in the US and China.
  • TechHQ: Discusses security standards for deepseek.
  • GZERO Media: Gzero reports about a potential US ban for DeepSeek
  • pub.towardsai.net: DeepSeek-R1 is a language model developed in China to enable sophisticated reasoning capabilities.
  • Analytics Vidhya: DeepSeek-R1 is a new AI model with strong reasoning capabilities.
  • medium.com: This article focuses on the ability of DeepSeek to handle sensitive topics and how it can be leveraged to detect censorship filters.
  • the-decoder.com: This article focuses on the potential capabilities of DeepSeek as an AI model, highlighting its potential to perform deep research and providing insights into the various capabilities.
  • Analytics Vidhya: DeepSeek is a new model capable of impressive logical reasoning, and it has been tested for its ability to create a large number of different types of code. This is a summary of the results.

jake_mendel@LessWrong //
References: AI Alignment Forum , LessWrong ,
Open Philanthropy is dedicating $40 million to fund technical AI safety research. The organization has launched a Request for Proposals (RFP) seeking projects across 21 research areas, aiming to develop robust safety techniques. This initiative focuses on mitigating potential risks from advanced AI systems before they are deployed in real-world scenarios.

The research areas are grouped into categories like adversarial machine learning, exploring sophisticated misbehavior of LLMs, and theoretical approaches to AI alignment. Specific areas of interest include jailbreaks, control evaluations, backdoor stress tests, robust unlearning, and alignment faking. Open Philanthropy is particularly interested in funding work related to jailbreaks and unintentional misalignment, control evaluations, and backdoor stress tests.

Open Philanthropy welcomes various types of grants, including research expenses, discrete research projects, academic start-up packages, support for existing nonprofits, and funding to start new organizations. The application process starts with a 300-word expression of interest, with applications open until April 15, 2025. The aim is to foster research that ensures AI systems adhere to safety specifications and reduce the probability of catastrophic failure.

Recommended read:
References :
  • AI Alignment Forum: Research directions Open Phil wants to fund in technical AI safety
  • LessWrong: Open Philanthropy Technical AI Safety RFP: $USD40M Available
  • LessWrong: We focus on threats from the misuse of models. A bad actor could disable safeguards and create the “evil twin” of a model.

@techcrunch.com //
Meta is actively developing AI safety systems to mitigate the potential for misuse of its AI models. The company is carefully defining the types of AI systems it deems too risky to release to the public. These include systems that could be used to aid in cyberattacks, chemical, and biological attacks. Meta will flag such systems and may halt their development altogether if the risks are considered too high.

To determine the risk level, Meta will rely on input from internal and external researchers, reviewed by senior-level decision-makers, rather than solely on empirical tests. If a system is deemed high-risk, access will be limited, and it won’t be released until mitigations reduce the risk to moderate levels. In cases of critical-risk AI, which could lead to catastrophic outcomes, Meta will implement more stringent measures. Anthropic is also addressing AI safety through their Constitutional Classifiers, designed to guard against jailbreaks and monitor content for harmful outputs. Leading tech groups, including Microsoft, are also investing in similar safety systems.

Recommended read:
References :
  • www.techmeme.com: Meta describes what kinds of AI systems it may deem too risky to release, including ones that could aid in cyberattacks, and how such systems will be flagged
  • techcrunch.com: Meta describes what kinds of AI systems it may deem too risky to release, including ones that could aid in cyberattacks, and how such systems will be flagged

Jibin Joseph@PCMag Middle East ai //
References: mobinetai.com , Pivot to AI , AI News ...
The DeepSeek AI model is facing growing scrutiny over its security vulnerabilities and ethical implications, leading to government bans in Australia, South Korea, and Taiwan, as well as for NASA employees in the US. Cisco researchers found DeepSeek fails to screen out malicious prompts and Dario Amodei of Anthropic has expressed concern over its ability to provide bioweapons-related information.

DeepSeek's lack of adequate guardrails has enabled the model to generate instructions on creating chemical weapons, and even planning terrorist attacks. Furthermore, DeepSeek has been accused of misrepresenting its training costs, with SemiAnalysis estimating that the company invested over $500 million in Nvidia GPUs alone, despite export controls. There are claims the US is investigating whether DeepSeek is acquiring these GPUs through gray market sales via Singapore.

Recommended read:
References :
  • mobinetai.com: Reports on DeepSeek's vulnerabilities and its ability to generate instructions on creating chemical weapons, and a terrorist attack.
  • Pivot to AI: Details DeepSeek's issues: government bans, lack of guardrails, and cost misrepresentations.
  • PCMag Middle East ai: The No DeepSeek on Government Devices Act comes after a study found direct links between the app and state-owned China Mobile.
  • AI News: US lawmakers are pushing for a DeepSeek ban after security researchers found the app transferring user data to a banned state-owned company.
  • mobinetai.com: Article on DeepSeek's ability to generate instructions for harmful activities, including chemical weapons and terrorist attacks.
  • www.artificialintelligence-news.com: News article about DeepSeek's data transfer to a banned state-owned company and the security concerns that follow.

@techcrunch.com //
References: techcrunch.com , www.cnbc.com ,
OpenAI is actively exploring the persuasive capabilities of its AI models, using the r/ChangeMyView subreddit as a testing ground. The company collects user posts and asks its AI to generate replies aimed at changing the poster's original viewpoint. These responses are then evaluated by human testers, with the results being compared to human replies for the same posts. Although OpenAI has a content-licensing deal with Reddit, it claims that this specific evaluation is separate. However, the test highlights the importance of human data in AI model development, as well as the complex ways in which tech companies obtain datasets.

OpenAI has also announced a significant partnership with U.S. National Laboratories, granting them access to its latest AI models for use in scientific research and nuclear weapons security. This collaboration will involve up to 15,000 scientists across the labs and will include deploying an OpenAI model on the Venado supercomputer at Los Alamos National Laboratory, in conjunction with Microsoft. Furthermore, OpenAI is reportedly seeking a substantial $40 billion in new funding, which could value the company at $300 billion. SoftBank is expected to lead this funding round with investments between $15 and $25 billion. This new funding would support OpenAI’s ongoing research and infrastructure projects such as the Stargate AI venture.

Recommended read:
References :
  • techcrunch.com: OpenAI used this subreddit to test AI persuasion
  • www.cnbc.com: OpenAI partners with U.S. National Laboratories on scientific research, nuclear weapons security.
  • www.pymnts.com: Report: OpenAI Aims to Raise $40 Billion in New Funding Round.

@www.pymnts.com //
OpenAI is reportedly in talks for a substantial $40 billion funding round, potentially valuing the company at $300 billion. The round, led by SoftBank, would see the Japanese conglomerate become OpenAI’s largest investor, surpassing Microsoft. This funding would support OpenAI's various initiatives, including its contribution to the Stargate AI infrastructure project and its ongoing business operations. It comes amid increased competition in the AI landscape, particularly with the emergence of new models like DeepSeek's, which is reportedly using substantially less hardware than comparable models.

OpenAI is also facing concerns regarding the use of its AI models by other companies. Allegations have surfaced that DeepSeek may have inappropriately used OpenAI's technology through a process known as "distillation," a technique used to boost the performance of smaller models. OpenAI is actively reviewing these allegations and has indicated it takes aggressive countermeasures to protect its technology. Additionally, OpenAI revealed it is partnering with U.S. National Laboratories, granting scientists access to its latest AI models for scientific research and nuclear weapons security, including working with Microsoft to deploy its tech on the Venado supercomputer at Los Alamos National Laboratory. Furthermore, OpenAI has used the subreddit r/ChangeMyView to test AI persuasion capabilities.

Recommended read:
References :