Pierluigi Paganini@securityaffairs.com
//
OpenAI is facing scrutiny over its ChatGPT user logs due to a recent court order mandating the indefinite retention of all chat data, including deleted conversations. This directive stems from a lawsuit filed by The New York Times and other news organizations, who allege that ChatGPT has been used to generate copyrighted news articles. The plaintiffs believe that even deleted chats could contain evidence of infringing outputs. OpenAI, while complying with the order, is appealing the decision, citing concerns about user privacy and potential conflicts with data privacy regulations like the EU's GDPR. The company emphasizes that this retention policy does not affect ChatGPT Enterprise or ChatGPT Edu customers, nor users with a Zero Data Retention agreement.
Sam Altman, CEO of OpenAI, has advocated for what he terms "AI privilege," suggesting that interactions with AI should be afforded the same privacy protections as communications with professionals like lawyers or doctors. This stance comes as OpenAI faces criticism for not disclosing to users that deleted and temporary chat logs were being preserved since mid-May in response to the court order. Altman argues that retaining user chats compromises their privacy, which OpenAI considers a core principle. He fears that this legal precedent could lead to a future where all AI conversations are recorded and accessible, potentially chilling free expression and innovation. In addition to privacy concerns, OpenAI has identified and addressed malicious campaigns leveraging ChatGPT for nefarious purposes. These activities include the creation of fake IT worker resumes, the dissemination of misinformation, and assistance in cyber operations. OpenAI has banned accounts linked to ten such campaigns, including those potentially associated with North Korean IT worker schemes, Beijing-backed cyber operatives, and Russian malware distributors. These malicious actors utilized ChatGPT to craft application materials, auto-generate resumes, and even develop multi-stage malware. OpenAI is actively working to combat these abuses and safeguard its platform from being exploited for malicious activities. Recommended read:
References :
Carl Franzen@AI News | VentureBeat
//
ElevenLabs has launched Conversational AI 2.0, a significant upgrade to its platform designed for building advanced voice agents for enterprise use. The new system allows agents to handle both speech and text simultaneously, enabling more fluid and natural interactions. This update introduces features aimed at creating more intelligent and secure conversations, making it suitable for applications like customer support, call centers, and outbound sales and marketing. According to Jozef Marko from ElevenLabs, Conversational AI 2.0 sets a new standard for voice-driven experiences.
One key highlight of Conversational AI 2.0 is its advanced turn-taking model. This technology analyzes conversational cues in real-time, such as hesitations and filler words like "um" and "ah", to determine when the agent should speak or listen. This eliminates awkward pauses and interruptions, creating a more natural flow. The platform also features integrated language detection, enabling seamless multilingual discussions without manual configuration. This allows the agent to recognize the language spoken by the user and respond accordingly, catering to global enterprises and fostering more inclusive experiences. In related news, Anthropic is rolling out voice mode for its Claude apps, utilizing ElevenLabs for speech generation. While currently only available in English, this feature allows users to engage in spoken conversations with Claude, enhancing accessibility and convenience. The voice conversations count toward regular usage limits based on subscription plans, with varying limits for free and paid users. This integration marks a significant step in making AI more conversational and user-friendly, leveraging ElevenLabs' technology to power its speech capabilities. Recommended read:
References :
Megan Crouse@techrepublic.com
//
OpenAI has unveiled a suite of advancements, including enhanced audio models and a significantly more expensive AI reasoning model called o1 Pro. The new audio models, including gpt-4o-transcribe and gpt-4o-mini-transcribe, offer improved transcription capabilities compared to Whisper, although they are susceptible to prompt injection attacks due to their foundation on language models. Users can access these models via the Realtime API, enabling real-time transcription from microphone input using a standalone Python script.
OpenAI's o1 Pro comes with a steep price tag of $150 per million input tokens and $600 per million output tokens. This makes it ten times more expensive than the standard o1 model and twice as costly as GPT-4.5. While OpenAI claims o1 Pro "thinks harder" and delivers superior responses for complex reasoning tasks, early benchmarks suggest only incremental improvements. Access to o1 Pro is currently limited to developers who have spent at least $5 on OpenAI's API services, targeting users building AI agents and automation tools. Recommended read:
References :
Keshav Kumaresan@DagsHub Blog
//
References:
DagsHub Blog
, The Cognitive Revolution
,
AI is making waves in unexpected areas. A recent study has found that AI-generated memes are, on average, funnier and more shareable than those created solely by humans. Researchers from KTH Royal Institute of Technology, LMU Munich, and TU Darmstadt, discovered that memes crafted entirely by OpenAI's GPT-4 scored higher in humor, creativity, and shareability. However, human-created memes still hold the crown for the absolute funniest individual examples, showcasing the unique personal touch humans bring to humor.
The Cognitive Revolution podcast recently featured Andreessen Horowitz partners Olivia Moore and Anish Acharya discussing the rapid advancements in voice AI. The discussion explored how the latest improvements are enabling more natural voice interactions across various platforms. Businesses are already utilizing voice AI for tasks ranging from complex negotiations to after-hours customer support. Recommended read:
References :
Chris McKay@Maginative
//
OpenAI has recently unveiled new audio models based on GPT-4o, significantly enhancing its text-to-speech and speech-to-text capabilities. These new tools are intended to give AI agents a voice, enabling a range of applications, with demonstrations including the ability for an AI to read emails in character. The announcement includes the introduction of new transcription models, specifically gpt-4o-transcribe and gpt-4o-mini-transcribe, which are designed to outperform the existing Whisper model.
The text-to-speech and speech-to-text tools are based on GPT-4o. While these models show promise, some experts have noted potential vulnerabilities. Like other large language model (LLM)-driven multi-modal models, they appear susceptible to prompt-injection-adjacent issues, stemming from the mixing of instructions and data within the same token stream. OpenAI hinted it may take a similar path with video. Recommended read:
References :
|
BenchmarksBlogsResearch Tools |