News from the AI & ML world

DeeperML - #opensourceai

Chris McKay@Maginative //
OpenAI has secured a massive $40 billion funding round, led by SoftBank, catapulting its valuation to an unprecedented $300 billion. This landmark investment makes OpenAI the world's second-most valuable private company alongside TikTok parent ByteDance Ltd, trailing only Elon Musk's SpaceX Corp. This deal marks one of the largest capital infusions in the tech industry and signifies a major milestone for the company, underscoring the escalating significance of AI.

The fresh infusion of capital is expected to fuel several key initiatives at OpenAI. The funding will support expanded research and development, and upgrades to computational infrastructure. This includes the upcoming release of a new open-weight language model with enhanced reasoning capabilities. OpenAI said the funding round would allow the company to “push the frontiers of AI research even further” and “pave the way” towards AGI, or artificial general intelligence.

Recommended read:
References :
  • Fello AI: OpenAI Secures Historic $40 Billion Funding Round
  • AI News | VentureBeat: In a move that surprised the tech industry Monday, OpenAI said it has secured a monumental $40 billion funding round led by SoftBank, catapulting its valuation to an unprecedented $300 billion -- making it the largest private equity investment on record.
  • InnovationAus.com: OpenAI has closed a significant $40 billion funding round, led by SoftBank Group, pushing its valuation to $300 billion.
  • Maginative: OpenAI Secures Record $40 Billion in Funding, Reaching $300 Billion Valuation
  • www.theguardian.com: OpenAI said it had raised $40bn in a funding round that valued the ChatGPT maker at $300bn – the biggest capital-raising session ever for a startup.
  • The Verge: OpenAI just raised another $40 billion round led by SoftBank
  • SiliconANGLE: OpenAI bags $40B in funding, increasing its post-money valuation to $300B. The bumper funding round was led by SoftBank Group Corp. and saw participation from existing backers of OpenAI, including Microsoft Corp., Coatue Management, Thrive Capital and Altimeter Capital.
  • techxplore.com: OpenAI says it raised $40 bn at valuation of $300 bn
  • THE DECODER: OpenAI nears completion of multi-billion dollar funding round
  • Kyle Wiggers ?: OpenAI raises $40B at $300B post-money valuation
  • THE DECODER: Softbank leads OpenAI's $40 billion funding round
  • Verdict: OpenAI has secured a $40 billion funding round, marking the biggest capital raising ever for a startup, with a $300 billion valuation. The deal is led by SoftBank and backed by leading investors.
  • Crunchbase News: OpenAI secured $40 billion in funding in a record-breaking round led by SoftBank, valuing the company at $300 billion.
  • bsky.app: OpenAI has raised $40 billion at a $300 billion valuation.
  • Pivot to AI: OpenAI has secured its $40 billion in fresh funding from SoftBank — probably. The FT lists $30 billion from SoftBank and $10 billion more from Microsoft and various venture capital funds. [OpenAI; FT, archive] That $30 billion from SoftBank may not exist yet. SoftBank is currently looking for someone to loan them $16 billion to […]
  • TechInformed: OpenAI has raised more than $40 billion in a fundraise with Japanese telco SoftBank and other investors, valuing the ChatGPT company at more than $300bn.… The post appeared first on .
  • www.techrepublic.com: OpenAI Secures $40B in Historic Funding Round — But There’s a $10B Catch
  • venturebeat.com: OpenAI announced plans to release its first “open-weight” language model since 2019, marking a dramatic strategic shift for the company that built its business on proprietary AI systems.
  • CyberInsider: OpenSNP to Shut Down and Delete All User-Submitted DNA Data

Michael Nuñez@venturebeat.com //
OpenAI, the company behind ChatGPT, has announced a significant strategic shift by planning to release its first open-weight AI model since 2019. This move comes amidst mounting economic pressures from competitors like DeepSeek and Meta, whose open-source models are increasingly gaining traction. CEO Sam Altman revealed the plans on X, stating that the new model will have reasoning capabilities and allow developers to run it on their own hardware, departing from OpenAI's cloud-based subscription model.

This decision marks a notable change for OpenAI, which has historically defended closed, proprietary models. The company is now looking to gather developer feedback to make the new model as useful as possible, planning events in San Francisco, Europe and Asia-Pacific. As models improve, startups and developers increasingly want more tunable latency, and want to use on-prem deplouments requiring full data control, according to OpenAI.

The shift comes alongside a monumental $40 billion funding round led by SoftBank, which has catapulted OpenAI's valuation to $300 billion. SoftBank will initially invest $10 billion, with the remaining $30 billion contingent on OpenAI transitioning to a for-profit structure by the end of the year. This funding will help OpenAI continue building AI systems that drive scientific discovery, enable personalized education, enhance human creativity, and pave the way toward artificial general intelligence. The release of the open-weight model is expected to help OpenAI compete with the growing number of efficient open-source alternatives and counter the criticisms that have come from remaining a closed model.

Recommended read:
References :
  • Data Science at Home: Is DeepSeek the next big thing in AI? Can OpenAI keep up? And how do we truly understand these massive LLMs?
  • venturebeat.com: OpenAI to release open-source model as AI economics force strategic shift
  • WIRED: Sam Altman Says OpenAI Will Release an ‘Open Weight’ AI Model This Summer
  • Fello AI: OpenAI Secures Historic $40 Billion Funding Round
  • www.theguardian.com: OpenAI said it had raised $40bn in a funding round that valued the ChatGPT maker at $300bn.
  • SiliconANGLE: OpenAI to launch its first ‘open-weights’ model since 2019
  • techxplore.com: OpenAI says it raised $40 bn at valuation of $300 bn
  • SiliconANGLE: OpenAI bags $40B in funding, increasing its post-money valuation to $300B
  • techxplore.com: OpenAI says it raised $40 bn at valuation of $300 bn
  • www.tomsguide.com: OpenAI is planning on launching its first open-weight model in years
  • THE DECODER: OpenAI plans to release open-weight reasoning LLM without usage restrictions
  • www.it-daily.net: OpenAI raises 40 billion dollars from investors
  • bsky.app: OpenAI has raised $40 billion at a $300 billion valuation. For context, Boeing has a $128 billion market cap, Disney has a $178 billion market cap, and Chevron has a $295 billion market cap. So, OpenAI has been valued at something like Boeing plus Disney, or just some $5 billion more than Chevron.
  • THE DECODER: SoftBank and OpenAI announced a major partnership on Monday that includes billions in annual spending and a new joint venture focused on the Japanese market.
  • The Tech Portal: OpenAI has closed a record-breaking $40 billion private funding round, marking the…
  • www.techrepublic.com: Developers Wanted: OpenAI Seeks Feedback About Open Model That Will Be Revealed ‘In the Coming Months’
  • bdtechtalks.com: Understanding OpenAI’s pivot to releasing open source models
  • techstrong.ai: OpenAI has secured up to $40 billion in a record new funding round led by SoftBank Group that would give the artificial intelligence (AI) pioneer a whopping $300 billion valuation as it ramps up AI research, infrastructure and tools.

Ryan Daws@AI News //
DeepSeek V3-0324 has emerged as a leading AI model, topping benchmarks for non-reasoning AI in an open-source breakthrough. This milestone signifies a significant advancement in the field, as it marks the first time an open weights model has achieved the top position among non-reasoning models. The model's performance surpasses proprietary counterparts and edges it closer to proprietary reasoning models, highlighting the growing viability of open-source solutions for latency-sensitive applications. DeepSeek V3-0324 represents a new era for open-source AI, offering a powerful and adaptable tool for developers and enterprises.

DeepSeek-V3 now runs at 20 tokens per second on Apple’s Mac Studio, presenting a challenge to OpenAI’s cloud-dependent business model. The 685-billion-parameter model, DeepSeek-V3-0324, is freely available for commercial use under the MIT license. This achievement, coupled with its cost efficiency and performance, signals a shift in the AI sector, where open-source frameworks increasingly compete with closed systems. Early testers report significant improvements over previous versions, positioning DeepSeek's new model above Claude Sonnet 3.5 from Anthropic.

Recommended read:
References :
  • Analytics India Magazine: The model outperformed all other non-reasoning models across several benchmarks but trailed behind DeepSeek-R1, OpenAI’s o1, o3-mini, and other reasoning models.
  • venturebeat.com: DeepSeek-V3 now runs at 20 tokens per second on Mac Studio, and that’s a nightmare for OpenAI
  • AI News: DeepSeek V3-0324 tops non-reasoning AI models in open-source first
  • Analytics Vidhya: DeepSeek V3-0324: Generated 700 Lines of Code without Breaking
  • Analytics Vidhya: DeepSeek V3-0324 vs Claude 3.7: Which is the Better Coder?
  • Cloud Security Alliance: Markets reacted dramatically, with Nvidia alone losing nearly $600 billion in value in a single day, part of a broader...
  • GZERO Media: Just a few short months ago, Silicon Valley seemed to have the artificial intelligence industry in a chokehold.
  • MarkTechPost: DeepSeek AI Unveils DeepSeek-V3-0324: Blazing Fast Performance on Mac Studio, Heating Up the Competition with OpenAI
  • SiliconANGLE: DeepSeek today released an improved version of its DeepSeek-V3 large language model under a new open-source license.
  • techstrong.ai: DeepSeek Ups Ante (Again) in Duel with OpenAI, Anthropic
  • www.zdnet.com: DeepSeek V3 model gets a major upgrade
  • www.techradar.com: DeepSeek’s new AI is smarter, faster, cheaper, and a real rival to OpenAI's models
  • Composio: Deepseek v3 0324: Finally, the Sonnet 3.5 at Home
  • AI News: DeepSeek disruption: Chinese AI innovation narrows global technology divide

Ryan Daws@AI News //
DeepSeek V3-0324, the latest large language model from Chinese AI startup DeepSeek, is making waves in the artificial intelligence industry. The model, quietly released with an MIT license for commercial use, has quickly become the highest-scoring non-reasoning model on the Artificial Analysis Intelligence Index. This marks a significant milestone for open-source AI, surpassing proprietary counterparts like Google’s Gemini 2.0 Pro, Anthropic’s Claude 3.7 Sonnet, and Meta’s Llama 3.3 70B.

DeepSeek V3-0324's efficiency is particularly notable. Early reports indicate that it can run directly on consumer-grade hardware, specifically Apple’s Mac Studio with an M3 Ultra chip, achieving speeds of over 20 tokens per second. This capability is a major departure from the typical data center requirements associated with state-of-the-art AI. The updated version demonstrates substantial improvements in reasoning and benchmark performance, as well as enhanced Chinese writing proficiency and optimized translation quality.

Recommended read:
References :
  • venturebeat.com: DeepSeek-V3 now runs at 20 tokens per second on Mac Studio, and that’s a nightmare for OpenAI
  • AI News: DeepSeek V3-0324 tops non-reasoning AI models in open-source first
  • Analytics Vidhya: DeepSeek V3-0324: Generated 700 Lines of Code without Breaking
  • Analytics India Magazine: The model outperformed all other non-reasoning models across several benchmarks but trailed behind DeepSeek-R1, OpenAI’s o1, o3-mini, and other reasoning models.
  • Cloud Security Alliance: DeepSeek: Behind the Hype and Headlines
  • techstrong.ai: DeepSeek Ups Ante (Again) in Duel with OpenAI, Anthropic
  • www.techradar.com: Deepseek’s new AI is smarter, faster, cheaper, and a real rival to OpenAI's models
  • Analytics Vidhya: DeepSeek V3-0324 vs Claude 3.7: Which is the Better Coder?
  • MarkTechPost: DeepSeek AI Unveils DeepSeek-V3-0324: Blazing Fast Performance on Mac Studio, Heating Up the Competition with OpenAI
  • www.zdnet.com: It's called V3-0324, but the real question is: Is it foreshadowing the upcoming launch of R2?
  • SiliconANGLE: DeepSeek today released an improved version of its DeepSeek-V3 large language model under a new open-source license.
  • Composio: Deepseek v3 o324, a new checkpoint, has been released by Deepseek in silence, with no marketing or hype, just a tweet and The post appeared first on .
  • Composio: Deepseek v3-0324 vs. Claude 3.7 Sonnet

Ryan Daws@AI News //
References: SiliconANGLE , venturebeat.com , AI News ...
DeepSeek, a Chinese AI company, has released DeepSeek V3-0324, an updated AI model that demonstrates impressive performance. The model is now running at 20 tokens per second on a Mac Studio. This model is said to contain 685 billion parameters and its cost-effectiveness challenges the dominance of American AI models, signaling that China continues to innovate in AI despite chip restrictions. Reports from early testers show improvements over previous versions and the model tops non-reasoning AI models in open-source first.

This new model runs on consumer-grade hardware, specifically Apple's Mac Studio with the M3 Ultra chip, diverging from the typical data center requirements for AI. It is freely available for commercial use under the MIT license. According to AI researcher Awni Hannun, the model runs at over 20 tokens per second on a 512GB M3 Ultra. The company has made no formal announcement, just an empty README file and the model weights themselves. This stands in contrast to the carefully orchestrated product launches by Western AI companies.

Recommended read:
References :
  • SiliconANGLE: DeepSeek today released an improved version of its DeepSeek-V3 large language model under a new open-source license.
  • venturebeat.com: DeepSeek-V3 now runs at 20 tokens per second on Mac Studio, and that’s a nightmare for OpenAI
  • AI News: Chinese AI innovation is reshaping the global technology landscape, challenging assumptions about Western dominance in advanced computing. Recent developments from companies like DeepSeek illustrate how quickly China has adapted to and overcome international restrictions through creative approaches to AI development.
  • AI News: DeepSeek V3-0324 tops non-reasoning AI models in open-source first
  • MarkTechPost: DeepSeek AI Unveils DeepSeek-V3-0324: Blazing Fast Performance on Mac Studio, Heating Up the Competition with OpenAI
  • Cloud Security Alliance: Cloud Security Alliance: DeepSeek: Behind the Hype and Headlines
  • Quinta?s weblog: DeepSeek-V3 now runs at 20 tokens per second on Mac Studio, and that’s a nightmare for OpenAI
  • Composio: Deepseek v3-0324 vs. Claude 3.7 Sonnet

Dashveenjit Kaur@AI News //
References: venturebeat.com , AI News , Sify ...
Chinese AI startup DeepSeek is shaking up the global technology landscape with its latest large language model, DeepSeek-V3-0324. This new model has been lauded for matching the performance of American AI models, while boasting significantly lower development costs. According to Lee Kai-fu, CEO of Chinese startup 01.AI, the gap between Chinese and American AI capabilities has narrowed dramatically, with China even ahead in some specific areas.

DeepSeek-V3-0324 features enhanced reasoning capabilities and improved performance in multiple benchmarks, particularly in mathematics. The model scored 59.4 on the American Invitational Mathematics Examination (AIME), a significant improvement over its predecessor. Häme University lecturer Kuittinen Petri noted DeepSeek's achievements were realized with just a fraction of the resources available to competitors like OpenAI. This breakthrough has been attributed to DeepSeek’s focus on algorithmic efficiency and novel approaches to model architecture, allowing them to overcome restrictions on access to the latest silicon.

This disruption is not going unnoticed, when DeepSeek launched its R1 model in January, America’s Nasdaq plunged 3.1%, while the S&P 500 fell 1.5%. While DeepSeek claimed a $5.6 million training cost, this represented only the marginal cost of the final training run. SemiAnalysis estimates DeepSeek's actual hardware investment at closer to $1.6 billion, with hundreds of millions in operating costs. The developments present opportunities and challenges for the.

Recommended read:
References :
  • venturebeat.com: DeepSeek-V3 now runs at 20 tokens per second on Mac Studio, and that’s a nightmare for OpenAI
  • AI News: DeepSeek disruption: Chinese AI innovation narrows global technology divide
  • GZERO Media: How DeepSeek changed China’s AI ambitions
  • Sify: DeepSeek’s AI Revolution: Creating an Entire AI Ecosystem
  • Nordic APIs: ChatGPT vs. DeepSeek: A Side-by-Side Comparison
  • Composio: Deepseek v3-0324 vs. Claude 3.7 Sonnet

Ryan Daws@AI News //
References: AI News , BigDATAwire , NVIDIA Newsroom ...
NVIDIA has launched Dynamo, an open-source inference software, designed to accelerate and scale reasoning models within AI factories. Dynamo succeeds the NVIDIA Triton Inference Server, representing a new generation of AI inference software specifically engineered to maximize token revenue generation for AI factories deploying reasoning AI models. The software orchestrates and accelerates inference communication across thousands of GPUs, utilizing disaggregated serving.

Dynamo optimizes AI factories by dynamically managing GPU resources in real-time to adapt to request volumes. Dynamo’s intelligent inference optimizations have shown to boost the number of tokens generated by over 30 times per GPU and has demonstrated the ability to double the performance and revenue of AI factories serving Llama models on NVIDIA’s current Hopper platform.

Recommended read:
References :
  • AI News: NVIDIA Dynamo: Scaling AI inference with open-source efficiency
  • BigDATAwire: At its GTC event in San Jose today, Nvidia unveiled updates to its AI infrastructure portfolio, including its next-generation datacenter GPU, the NVIDIA Blackwell Ultra.
  • AIwire: Nvidia’s DGX AI Systems Are Faster and Smarter Than Ever
  • NVIDIA Newsroom: NVIDIA Blackwell Powers Real-Time AI for Entertainment Workflows
  • MarkTechPost: Details the Open Sourcing of Dynamo

Alex Knapp,@Alex Knapp //
References: Meta , Alex Knapp ,
Meta's open-source large language model (LLM), Llama, has achieved a significant milestone, surpassing one billion downloads since its release in 2023. This achievement underscores the growing influence of Llama in the AI community, attracting both researchers and enterprises seeking to integrate it into various applications. The model's popularity has surged, with companies like Spotify, AT&T, and DoorDash adopting Llama-based models for production environments.

Meta views open sourcing AI models as crucial, with each download of Llama moving closer to this goal. However, Llama's widespread use hasn't been without its challenges, including copyright lawsuits alleging training on copyrighted books without permission. The company plans to introduce multimodal models and improved reasoning capabilities. Additionally, Meta has been working to incorporate innovations from competing models to enhance Llama's performance.

Recommended read:
References :
  • Meta: Celebrating 1 Billion Downloads of Llama
  • Alex Knapp: The Prompt: Meta’s Open Source LLM Llama Has Been Downloaded Over One Billion Times
  • Maginative: Meta’s Llama AI Model Hits 1 Billion Downloads

Jason Corso,@AI News | VentureBeat //
The open-source AI landscape is currently facing challenges related to transparency, maintainability, and evaluation. Selective transparency is raising concerns, as truly open-source AI should allow for inspection, experimentation, and understanding of all contributing elements. In tandem, open-source maintainers report being overwhelmed by a surge in junk bug reports generated by AI systems. These reports, often low-quality and hallucinated, require time and effort to refute, increasing the workload for maintainers.

Efforts are underway to improve the red-teaming of AI systems to enhance understanding and governance. A recent workshop highlighted challenges and offered recommendations for better AI evaluations. While the policy landscape has shifted towards prioritizing AI innovation, evaluations like red-teaming remain critical for identifying safety and security risks. This involves emulating attacker tactics to "break" AI models and identifying unwanted outputs.

Recommended read:
References :

Jason Corso,@AI News | VentureBeat //
The increasing use of AI in software development and security analysis is presenting new challenges for open-source projects. While open-source AI tools are gaining traction due to faster development and innovation, maintainers are now facing a surge of low-quality bug reports generated by AI systems. These reports, often described as "spammy" and "hallucinated," appear legitimate at first but waste valuable time as maintainers must investigate and refute them.

The Computer History Museum, in collaboration with Google, has recently released the original 2012 source code for AlexNet, a revolutionary neural network. This release is a significant milestone for AI enthusiasts, enabling deeper understanding and further innovation. However, the flood of AI-generated junk bug reports raises concerns about the impact of AI on the open-source ecosystem, with developers like Seth Larson suggesting such low-quality reports should be treated as potentially malicious.

Recommended read:
References :

Muhammad Zulhusni@AI News //
References: AI News , GZERO Media
Several major US artificial intelligence companies have expressed fears that America is losing its edge in AI development. In submissions to the US government in March 2025, these companies warned that Chinese AI models, like DeepSeek R1, are becoming increasingly sophisticated and competitive. The submissions, prompted by a request for input on an AI Action Plan, underscore the growing challenge posed by China in terms of technological capabilities and pricing within the AI sector.

China's growing AI capabilities are exemplified by DeepSeek R1, a state-supported model that has garnered attention from US developers. OpenAI noted that DeepSeek demonstrates a narrowing technological gap between the US and China, expressing concerns about the model's potential to influence global AI development, particularly given its "state-subsidized, state-controlled, and freely available" nature. Competition from China also includes Ernie X1 and Ernie 4.5, released by Baidu, which are designed to compete with Western systems.

Recommended read:
References :
  • AI News: Is America falling behind in the AI race?
  • GZERO Media: How DeepSeek changed China’s AI ambitions

Harsh@Composio //
The integration of Artificial Intelligence (AI) into coding and software development is rapidly evolving, sparking both excitement and ethical considerations. GitHub's COO, Kyle Daigle, recently discussed the impact of AI-assisted coding, highlighting tools like GitHub Copilot and the potential of ambient AI to seamlessly integrate into developer workflows. The discussion included licensing concerns and the importance of developers understanding and navigating the ethical complexities that arise with AI-driven development. This comes as developers are exploring AI Agents SDK, a framework that simplifies the creation of multi-agent systems.

The appeal of AI in development is further underscored by the comparison of Agents SDK with alternatives like LangGraph, Autogen, and CrewAI. Each framework offers unique strengths, with Agents SDK focusing on simplicity and production readiness, while LangGraph excels in complex workflows. However, amidst the enthusiasm, ethical considerations are surfacing, most notably the controversy surrounding OpenAI employees questioning the company's military deal with startup Anduril, raising concerns about the potential misuse of AI and its impact on OpenAI's reputation. This highlights the ongoing debate about the ethical boundaries of AI development and deployment, particularly in sensitive areas like defense and healthcare.

Recommended read:
References :
  • Composio: OpenAI Agents SDK vs LangGraph vs Autogen vs CrewAI
  • Windows Copilot News: OpenAI employees question the ethics of military deal with startup Anduril

Ryan Daws@AI News //
References: AI News
The open-source AI movement is gaining momentum, with several significant developments highlighting its growing influence. Hugging Face is actively advocating for an open-source approach in the US government's upcoming AI Action Plan, emphasizing that innovation thrives with diverse contributors and accessible infrastructure. They propose focusing on strengthening open-source AI ecosystems, promoting efficient AI adoption, and establishing robust security standards.

The All Things Open AI conference saw unexpected success, reflecting the increasing interest in the field. Attendance exceeded expectations, indicating the strong demand for collaborative learning and knowledge sharing within the open-source AI community. This event, a partnership between All Things Open and The Artificially Intelligent Enterprise, featured training sessions and presentations, drawing a large crowd of participants.

In a landmark event for AI history, the Computer History Museum, in collaboration with Google, has released the original source code for AlexNet, the groundbreaking neural network that revolutionized AI in 2012. This opens up new avenues for research and understanding of the foundations of modern AI, enabling developers and researchers to delve into the intricacies of AlexNet's architecture and algorithms. This is considered a monumental moment for AI enthusiasts.

Recommended read:
References :
  • AI News: Hugging Face calls for open-source focus in the AI Action Plan