Ashutosh Singh@The Tech Portal
//
Apple is enhancing its AI capabilities, known as Apple Intelligence, by employing synthetic data and differential privacy to prioritize user privacy. The company aims to improve features like Personal Context and Onscreen Awareness, set to debut in the fall, without collecting or copying personal content from iPhones or Macs. By generating synthetic text and images that mimic user behavior, Apple can gather usage data and refine its AI models while adhering to its strict privacy policies.
Apple's approach involves creating artificial data that closely matches real user input to enhance Apple Intelligence features. This method addresses the limitations of training AI models solely on synthetic data, which may not always accurately reflect actual user interactions. When users opt into Apple's Device Analytics program, the AI models will compare these synthetic messages against a small sample of a user’s content stored locally on the device. The device then identifies which of the synthetic messages most closely matches its user sample, and sends information about the selected match back to Apple, with no actual user data leaving the device.
To further protect user privacy, Apple utilizes differential privacy techniques. This involves adding randomized data to broader datasets to prevent individual identification. For example, when analyzing Genmoji prompts, Apple polls participating devices to determine the popularity of specific prompt fragments. Each device responds with a noisy signal, ensuring that only widely-used terms become visible to Apple, and no individual response can be traced back to a user or device. Apple plans to extend these methods to other Apple Intelligence features, including Image Playground, Image Wand, Memories Creation, and Writing Tools. This technique allows Apple to improve its models for longer-form text generation tasks without collecting real user content.
Recommended read:
References :
- www.artificialintelligence-news.com: Apple leans on synthetic data to upgrade AI privately
- The Tech Portal: Apple to use synthetic data that matches user data to enhance Apple Intelligence features
- www.it-daily.net: Apple AI stresses privacy with synthetic and anonymised data
- www.macworld.com: How will Apple improve its AI while protecting your privacy?
- www.techradar.com: Apple has a plan for improving Apple Intelligence, but it needs your help – and your data
- machinelearning.apple.com: Understanding Aggregate Trends for Apple Intelligence Using Differential Privacy
- AI News: Apple AI stresses privacy with synthetic and anonymised data
- THE DECODER: Apple will use your emails to improve AI features without ever seeing them
- Computerworld: Apple’s big plan for better AI is you
- Maginative: Apple Unveils Clever Workaround to Improve AI Without Collecting Your Data
- thetechbasic.com: Apple intends to improve its AI products, Siri and Genmoji, by developing better detection capabilities without accessing personal communication content. Apple released a method that functions with artificial data and privacy mechanisms.
- www.verdict.co.uk: Apple to begin on-device data analysis to enhance AI
- 9to5mac.com: Apple details on-device Apple Intelligence training system using user data
- The Tech Basic: Apple intends to improve its AI products, Siri and Genmoji, by developing better detection capabilities without accessing personal communication content.
- Digital Information World: Apple Silently Shifting Gears on AI by Analyzing User Data Through Recent Snippets of Real World Data
- PCMag Middle East ai: With an upcoming OS update, Apple will compare synthetic AI training data with real customer data to improve Apple Intelligence—but only if you opt in.
- www.zdnet.com: How Apple plans to train its AI on your data without sacrificing your privacy
- www.eweek.com: Apple recently outlined several methods it plans to use to improve Apple Intelligence while maintaining user privacy.
- eWEEK: Apple Reveals How It Plans to Train AI – Without Sacrificing Users’ Privacy
- analyticsindiamag.com: New Training Methods to Save Apple Intelligence?
@www.thecanadianpressnews.ca
//
Meta is resuming its AI training program using public content shared by adult users in the European Union. This decision follows earlier delays due to regulatory concerns and aims to improve the understanding of European cultures, languages, and history within Meta's AI models. The data utilized will include public posts and comments from platforms like Facebook and Instagram, helping the AI to better reflect the nuances and complexities of European communities. Meta believes this is crucial for developing AI that is not only available to Europeans but is specifically tailored for them.
Meta will begin notifying EU users this week through in-app notifications and email, explaining the types of data they plan to use and how it will enhance AI functionality and the overall user experience. These notifications will include a direct link to an objection form, allowing users to easily opt out of having their data used for AI training purposes. Meta emphasizes that they will honor all objection forms, both those previously received and any new submissions. This approach aims to balance AI development with individual privacy rights under the stringent data privacy rules in the EU.
The move comes after Meta had to previously shelve its European AI rollout plans following concerns raised about the privacy implications of its AI tools. Meta also faces ongoing legal challenges related to the use of copyright-protected material in its large language model development. The company maintains that access to EU user data is essential for localizing its AI tools, enabling them to understand everything from dialects and colloquialisms to hyper-local knowledge and unique cultural expressions like humor and sarcasm. Without this data, Meta argues, the region risks being left behind in AI development, particularly as AI models become more advanced and multi-modal.
Recommended read:
References :
- cyberinsider.com: Meta has announced it will soon begin using public data from adult users in the European Union — including posts, comments, and AI interactions — to train its generative AI models, raising concerns about the boundaries of consent and user awareness across its major platforms.
- discuss.privacyguides.net: Meta to start training its AI models on public content in the EU after Est. reading time: 3 minutes If you are an EU resident with an Instagram or Facebook account, you should know that Meta will start training its AI models on your posted content.
- Malwarebytes: Meta users in Europe will have their public posts swept up and ingested for AI training, the company announced this week.
- : Meta says it will start using publicly available content from European users to train its artificial intelligence models, resuming work put on hold last year after activists raised concerns about data privacy.
- bsky.app: Meta announced today that it will soon start training its artificial intelligence models using content shared by European adult users on its Facebook and Instagram social media platforms. https://www.bleepingcomputer.com/news/technology/meta-to-resume-ai-training-on-content-shared-by-europeans/
- BleepingComputer: Meta to resume AI training on content shared by Europeans
- oodaloop.com: Meta says it will resume AI training with public content from European users
- BleepingComputer: Meta announced today that it will soon start training its artificial intelligence models using content shared by European adult users on its Facebook and Instagram social media platforms.
- techxplore.com: Social media company Meta said Monday that it will start using publicly available content from European users to train its artificial intelligence models, resuming work put on hold last year after activists raised concerns about data privacy.
- finance.yahoo.com: Meta says it will resume AI training with public content from European users
- www.theverge.com: The Verge news article on Meta's use of EU user data for AI training.
- The Hacker News: Meta Resumes E.U. AI Training Using Public User Data After Regulator Approval
- www.socialmediatoday.com: Social Media Today article on Meta training AI on EU user data.
- Meta: Today, we’re announcing our plans to train AI at Meta using public content —like public posts and comments— shared by adults on our products in the EU.
- Synced: Meta’s Novel Architectures Spark Debate on the Future of Large Language Models
- securityaffairs.com: Meta will use public EU user data to train its AI models
- about.fb.com: Today, we’re announcing our plans to train AI at Meta using public content —like public posts and comments— shared by adults on our products in the EU. People’s interactions with Meta AI – like questions and queries – will also be used to train and improve our models.
- www.bitdegree.org: Meta Cleared to Train AI with Public Posts in the EU
- MEDIANAMA: Meta to begin using EU users’ data to train AI models
- www.medianama.com: Meta to begin using EU users’ data to train AI models
- The Register - Software: Meta to feed Europe's public posts into AI brains again
- www.artificialintelligence-news.com: Meta will train AI models using EU user data
- AI News: Meta will train AI models using EU user data
- techxmedia.com: Meta announced it will use public posts and comments from adult EU users to train its AI models, ensuring compliance with EU regulations.
- Digital Information World: Despite all the controversy that arose, tech giant Meta is now preparing to train its AI systems on data belonging to Facebook and Instagram users in the EU.
- TechCrunch: Meta will start training its AI models on public content in the EU
Ryan Daws@AI News
//
OpenAI is set to release its first open-weight language model since 2019, marking a strategic shift for the company. This move comes amidst growing competition in the AI landscape, with rivals like DeepSeek and Meta already offering open-source alternatives. Sam Altman, OpenAI's CEO, announced the upcoming model will feature reasoning capabilities and allow developers to run it on their own hardware, departing from OpenAI's traditional cloud-based approach.
This decision follows OpenAI securing a $40 billion funding round, although reports suggest a potential breakdown of $30 billion from SoftBank and $10 billion from Microsoft and venture capital funds. Despite the fresh funding, OpenAI also faces scrutiny over its training data. A recent study by the AI Disclosures Project suggests that OpenAI's GPT-4o model demonstrates "strong recognition" of copyrighted data, potentially accessed without consent. This raises ethical questions about the sources used to train OpenAI's large language models.
Recommended read:
References :
- Fello AI: OpenAI Secures Historic $40 Billion Funding Round
- AI News | VentureBeat: $40B into the furnace: As OpenAI adds a million users an hour, the race for enterprise AI dominance hits a new gear
- InnovationAus.com: OpenAI has closed a significant $40 billion funding round, led by SoftBank Group, pushing its valuation to $300 billion.
- Maginative: OpenAI Secures Record $40 Billion in Funding, Reaching $300 Billion Valuation
- www.theguardian.com: OpenAI said it had raised $40bn in a funding round that valued the ChatGPT maker at $300bn – the biggest capital-raising session ever for a startup.
- The Verge: OpenAI just raised another $40 billion round led by SoftBank
- SiliconANGLE: OpenAI bags $40B in funding, increasing its post-money valuation to $300B
- techxplore.com: OpenAI says it raised $40 bn at valuation of $300 bn
- THE DECODER: OpenAI nears completion of multi-billion dollar funding round
- Kyle Wiggers ?: OpenAI raises $40B at $300B post-money valuation
- THE DECODER: Softbank leads OpenAI's $40 billion funding round
- Verdict: OpenAI has secured a $40 billion funding round, marking the biggest capital raising ever for a startup, with a $300 billion valuation. The deal is led by SoftBank and backed by leading investors.
- Crunchbase News: OpenAI secured $40 billion in funding in a record-breaking round led by SoftBank, valuing the company at $300 billion.
- bsky.app: OpenAI has raised $40 billion at a $300 billion valuation. For context, Boeing has a $128 billion market cap, Disney has a $178 billion market cap, and Chevron has a $295 billion market cap.
- Pivot to AI: OpenAI signs its $40 billion deal with SoftBank! Or maybe $30 billion, probably
- TechInformed: OpenAI has raised more than $40 billion in a fundraise with Japanese telco SoftBank and other investors, valuing the ChatGPT company at more than $300bn.… The post appeared first on .
- CyberInsider: OpenSNP to Shut Down and Delete All User-Submitted DNA Data
- www.techrepublic.com: OpenAI Secures $40B in Historic Funding Round — But There’s a $10B Catch
- techstrong.ai: OpenAI has secured up to $40 billion in a record new funding round led by SoftBank Group that would give the artificial intelligence (AI) pioneer a whopping $300 billion valuation as it ramps up AI research, infrastructure and tools.
- SiliconANGLE: OpenAI to launch its first ‘open-weights’ model since 2019
- venturebeat.com: OpenAI to release open-source model as AI economics force strategic shift
- AI News: Study claims OpenAI trains AI models on copyrighted data
- Charlie Fink: OpenAI raises $40 billion, Runway’s $380 million raise and its stunning Gen-4 AI model, Anthropic warns AI may lie, plus vibe filmmaking with DeepMind.
- thezvi.wordpress.com: Greetings from Costa Rica! The image fun continues. We Are Going to Need A Bigger Compute Budget Fun is being had by all, now that OpenAI has dropped its rule about not mimicking existing art styles.
@www.cnbc.com
//
DeepSeek AI, a rapidly growing Chinese AI startup, has suffered a significant data breach, exposing a database containing over one million log lines of sensitive information. Security researchers at Wiz discovered the exposed ClickHouse database was publicly accessible and unauthenticated, allowing full control over database operations without any defense mechanisms. The exposed data included user chat histories, secret API keys, backend details, and other highly sensitive operational metadata. This exposure allowed potential privilege escalation within the DeepSeek environment.
The Wiz research team identified the vulnerability through standard reconnaissance techniques on publicly accessible domains and by discovering unusual, open ports linked to DeepSeek. The affected database was hosted at oauth2callback.deepseek.com:9000 and dev.deepseek.com:9000. Researchers noted the ease of discovery of the exposed data and the potential for malicious actors to have accessed it. DeepSeek has been contacted by security researchers, and has now secured the database after the discovery, however, it remains unclear if unauthorized third-parties were also able to access the information.
Recommended read:
References :
- NewsGuard's Reality Check: NewsGuard: with news-related prompts, DeepSeek's chatbot repeated false claims 30% of the time and provided non-answers 53% of the time, giving an 83% fail rate (NewsGuard's Reality Check)
- www.theregister.com: Upgraded China's DeepSeek, which has rattled American AI makers, has limited new signups to its web-based interface
- Pyrzout :vm:: Social.skynetcloud.site post about DeepSeek's database leak
- www.wired.com: Wiz: DeepSeek left one of its critical databases exposed, leaking more than 1M records including system logs, user prompt submissions, and users' API keys (Wired)
- ciso2ciso.com: Guess who left a database wide open, exposing chat logs, API keys, and more? Yup, DeepSeek
- The Hacker News: DeepSeek AI Database Exposed: Over 1 Million Log Lines, Secret Keys Leaked
- Wiz Blog | RSS feed: Wiz Research Uncovers Exposed DeepSeek Database Leaking Sensitive Information, Including Chat History | Wiz Blog
- www.theverge.com: News about DeepSeek's data security breach.
- www.wired.com: Wired article discussing DeepSeek's AI jailbreak.
- arstechnica.com: Report: DeepSeek's chat histories and internal data were publicly exposed.
@www.cnbc.com
//
Chinese AI company DeepSeek is currently facing a large-scale cyberattack that has led to the temporary suspension of new user registrations. The company made the announcement on Monday, stating that existing users could still log in as usual while they work to mitigate the attack. DeepSeek is known for its open-source projects and has recently released models like R1, a reasoning model, and Janus-Pro-7B, a multi-modal AI model capable of generating images. This incident highlights the security vulnerabilities that AI service providers face and the potential disruption these attacks can cause to the industry and its users.
The cyberattack comes as DeepSeek's technology has been gaining attention and challenging established AI companies. The company has also released an iOS app, DeepSeek – AI Assistant, which has become a top download. There are also reports that DeepSeek may have used OpenAI's model to train its competitor. This has brought new focus on the competition between China and the US in the AI industry. This incident raises questions about the security and stability of AI infrastructure, especially in light of geopolitical competition and the importance of AI in various sectors.
Recommended read:
References :
- Techmeme: DeepSeek on Monday said it would temporarily limit user registrations “due to large-scale malicious attacks” on its services, though existing users will be able to log in as usual.
- www.cnbc.com: DeepSeek on Monday said it would temporarily limit user registrations “due to large-scale malicious attacks” on its services, though existing users will be able to log in as usual.
- www.theguardian.com: TheGuardian post about DeepSeek cyberattack.
- www.themirror.com: TheMirror post about DeepSeek censorship.
- www.theregister.com: TheRegister post about DeepSeek suspending registrations.
- Techmeme: Wiz: DeepSeek left one of its critical databases exposed, leaking more than 1M records including system logs, user prompt submissions, and users' API keys (Wired)
- www.wired.com: Exposed DeepSeek Database Revealed Chat Prompts and Internal Data
- Pyrzout :vm:: Guess who left a database wide open, exposing chat logs, API keys, and more? Yup, DeepSeek – Source: go.theregister.com
- ciso2ciso.com: Guess who left a database wide open, exposing chat logs, API keys, and more? Yup, DeepSeek – Source: go.theregister.com
- Wiz Blog | RSS feed: Wiz Research Uncovers Exposed DeepSeek Database Leaking Sensitive Information, Including Chat History | Wiz Blog
- www.cnbc.com: The US Navy has instructed its members to avoid using DeepSeek "in any capacity" due to "potential security and ethical concerns"
- heise online English: Upgraded China's DeepSeek, which has rattled American AI makers, has limited new signups to its web-based interface containing patterns in what is said to be an ongoing cyberattack.
- The Hacker News: DeepSeek AI Database Exposed: Over 1 Million Log Lines, Secret Keys Leaked
- www.theverge.com: The Verge reports on DeepSeek's database exposing user data and chat histories.
- www.infosecurity-magazine.com: Infosecurity Magazine reports on the DeepSeek database exposure and the types of sensitive data leaked.
|
|