- AI Emergence
- Posts
- đź‘€ Did Claude Read Reddit Without Permission?
đź‘€ Did Claude Read Reddit Without Permission?
Along with: ChatGPT Gets Serious for Work, DeepSeek Gets Smarter
Hi there đź‘‹
Reddit just sued Anthropic - claiming Claude was trained on data scraped from over 100,000 Reddit pages without permission.
While OpenAI and Google paid for Reddit’s data, Anthropic (allegedly) took the backdoor. Now it’s in hot water over user privacy and data rights.
Reddit sues Anthropic for allegedly not paying for training data | TechCrunch
— TechCrunch (@TechCrunch)
6:38 PM • Jun 4, 2025
But that doesn’t stop new AIs getting shipped!
OpenAI dropped new tools that turn ChatGPT into a serious work assistant. DeepSeek rolled out a model that thinks deeper, hallucinates less, and challenges the best in math and logic. And Bing? It quietly lets you generate short Sora-style videos for free.
Let’s get into this week’s AI news.
What would be the format? Every week, we will break the newsletter into the following sections:
The Input - All about recent developments in AI
The Algorithm - Resources for learning
The Output - Our reflection
Table of Contents
OpenAI just rolled out a serious upgrade for ChatGPT Team users, turning the app into something that feels less like a chatbot and more like a workplace command center.
What’s New:
Built-in Connectors: ChatGPT can now plug directly into Google Drive, Gmail, Outlook, OneDrive, SharePoint, Dropbox, and GitHub. That means you can ask it to read documents, answer questions from your actual files, write emails, or summarize team updates, all without copy-pasting. Access controls stay in place, so it only sees what you can.
Record Mode (Mac-only for now): You can record up to 2 hours of meetings, voice memos, or brainstorm sessions. It transcribes everything, summarizes key points, and even suggests follow-ups, all saved as a canvas you can revisit and build on later.
Custom Integrations (MCP): For companies with their own systems, OpenAI now supports the Model Context Protocol - letting devs build custom connectors that pull data from proprietary tools, databases, or internal APIs.
Prebuilt & custom connectors, record mode, and some more updates to ChatGPT for business.
— OpenAI (@OpenAI)
4:55 PM • Jun 4, 2025
Chinese startup DeepSeek just rolled out an upgraded version of its flagship model - DeepSeek-R1-0528 and it’s aiming straight at OpenAI and Google. The new model thinks deeper, hallucinates less, and delivers big gains across math, coding, and logic tasks.
What’s New:
Smarter math: On the AIME 2025 benchmark, accuracy jumped from 70% to 87.5% mostly by thinking longer and harder (literally, averaging 23K tokens per problem).
Cleaner output: Fewer hallucinations, support for JSON, and better front-end control baked in.
Competitive benchmarks: R1-0528 is now in the same league as o3 and Gemini 2.5 Pro, especially in multi-step reasoning tasks.
Cost disruptor: Analyst Mary Meeker flagged DeepSeek as a low-cost challenger that could shake up the AI market.
It’s another reminder that the frontier isn’t just about intelligence, it’s about accessibility, performance, and who can deliver the most value at scale.
đź”— Try it now: chat.deepseek.com
🚀 DeepSeek-R1-0528 is here!
🔹 Improved benchmark performance
🔹 Enhanced front-end capabilities
🔹 Reduced hallucinations
🔹 Supports JSON output & function calling✅ Try it now: chat.deepseek.com
🔌 No change to API usage — docs here: api-docs.deepseek.com/guides/reasoni…
🔗— DeepSeek (@deepseek_ai)
12:11 PM • May 29, 2025
Microsoft just rolled out a new feature in the Bing app called Video Creator, and it’s exactly what it sounds like - type a prompt, get a 5-second AI-generated video. It's powered by OpenAI's Sora and aimed at quick, shareable vertical content.
What’s New:
Free video generation: Just open the Bing app (iOS/Android), type a prompt, and get a short 9:16 clip.
Two speeds: Choose standard (free, slower) or fast (10 free fast gens, then 100 Microsoft Rewards points per video).
Queue up to 3 at a time: And your videos stay available for 90 days.
Perfect for social: The clips are designed for TikTok/Instagram-style formats, with horizontal support coming soon.
It’s not Veo-level cinematic yet, but for quick creative tests and casual content, it’s a solid starting point and free’s hard to argue with. Try Bing Video Creator here. (source)
OpenAI also rolled out two handy updates and they quietly make ChatGPT a lot more useful.
First, Codex (the dev assistant inside ChatGPT) can now access the internet while working. That means it can install packages, fetch real-time data, and run external tests — all while you control what it can or can’t access. Now available for Plus users as well.
Second, memory is now coming to free users. This lets ChatGPT remember useful info across chats — like your name, preferences, or past projects so you don’t have to start from scratch every time. It’s on by default, and you can turn it off or reset it anytime.
Codex is rolling out to ChatGPT Plus users today.
Get started at chatgpt.com/codex
— OpenAI (@OpenAI)
5:44 PM • Jun 3, 2025
Tencent Hunyuan and Tencent Music just launched HunyuanVideo-Avatar, a model that animates static images into lifelike, expressive video avatars-complete with speech or singing, synced to your audio input.
What’s New:
Photo + Audio In, Dynamic Avatar Out: Auto-detects emotion and context to generate expressive video with natural lip-sync and head motion.
Supports Diverse Use Cases: Multi-style, multi-species, and multi-character support-ideal for short videos, ads, e-commerce, and more.
Battle-Tested: Already deployed across Tencent Music’s ecosystem.
Now Open-Sourced: Single-character mode is live on Hunyuan - supports audio up to 14s. Multi-character mode coming soon.
Check out the project page, dive into the GitHub repo, or read the tech report. Whether you’re building for entertainment or product demos, HunyuanVideo-Avatar brings a new level of realism to AI-driven avatars.
🚀 Introducing HunyuanVideo-Avatar, a model jointly developed by Tencent Hunyuan and Tencent Music, bringing photos to life.
✅ Upload a photo + audio — auto-detect scene context & emotion, then generate lifelike speech/singing with dynamic visuals.
âś… Supports multi-style,
— Hunyuan (@TencentHunyuan)
3:58 AM • May 28, 2025
Resemble AI just dropped Chatterbox, a fully open-source voice cloning model (MIT licensed) that covers text-to-speech, voice conversion, and speech-to-speech-and it's outperforming the big names.
What’s New:
Outperforms ElevenLabs: In blind tests, 63.75% of listeners preferred Chatterbox’s output.
TTS + Voice Conversion + Speech Mimicry: Clone voices, generate expressive speech, or mirror an uploaded voice-same tone, accent, emotion.
Fine-Grained Control: Adjust speaking style, intensity (angry, soft, dramatic), tone, or pacing directly via prompt.
Built for Scale & Creativity: Chatterbox is powered by a 0.5B LLaMA backbone and trained on over 500,000 hours of meticulously cleaned data. It delivers stable, high-quality audio with built-in watermarking for safety and attribution. Unlike many other models, it doesn’t require prior voice samples to generate expressive, human-like speech-making it both powerful and highly accessible. (source)
🚨BREAKING: A new open-source voice model just dropped and it’s better than ElevenLabs.
This open-source voice AI is ultra-expressive, high quality, and fully free with no paywalls, no limits.
Here’s how it works (with real examples):👇
— Hasan Toor ✪ (@hasantoxr)
7:21 AM • Jun 3, 2025
Black Forest Labs just dropped FLUX.1 Kontext, a new family of generative flow models that lets you create, edit, and remix images using text, images, or both-all with blazing speed and no extra finetuning.
What’s New:
Prompt with Images + Text: FLUX.1 Kontext handles both-generate, refine, and remix with full multimodal control.
Iterative, Step-by-Step Edits: Make changes across multiple turns without losing quality, style, or character details.
Local Edits, Global Control: Update a tree, not the whole forest. Selectively tweak image elements without affecting the rest.
Character + Style Consistency: Keep your subjects recognizable and your art style intact across scenes.
Fast + Sharp Results: Generates up to 8x faster than GPT-Image, with strong prompt following and photorealism.
Two Models, Two Power Levels
FLUX.1 Kontext [pro]: Combines in-context edits, local tweaks, and full text-to-image generation in one fast, flexible model. Handles multi-turn edits with minimal drift.
FLUX.1 Kontext [max]: Cranks up prompt accuracy, typography, and edit quality-without sacrificing speed.
Available On: KreaAI, Freepik, Lightricks, OpenArt, LeonardoAI, and infra partners like FAL, Replicate, TogetherAI, and more.
No integration needed. Head to playground.bfl.ai to test the FLUX models in real time. Perfect for prototyping, demos, and showing stakeholders what this tech can really do. (source)
Perplexity just launched Labs, a powerful new workspace for Pro users that doesn’t just answer questions-it builds full deliverables. Think of it as your on-demand research and dev team. While Search gives you instant facts and Research dives deeper, Labs takes it a step further by generating reports, spreadsheets, dashboards, and even simple web apps using tools like deep web browsing, code execution, and charting.
Labs can turn a vague idea into a polished output in about 10 minutes-no dev tools or manual wrangling required. You can even build interactive mini apps inside your project and keep everything organized in the Assets tab. Whether you're planning a campaign, analyzing data, or prototyping a product, Labs gets it done. Now live for Pro users on Web, iOS, and Android (Mac & Windows coming soon) at perplexity.com. (source)
Today we're launching Perplexity Labs.
Labs is for your more complex tasks. It's is like having an entire team at your disposal.
Build anything from analytical reports and presentations to dynamic dashboards. Now available for all Pro users.
— Perplexity (@perplexity_ai)
5:27 PM • May 29, 2025
ElevenLabs just dropped Conversational AI 2.0, a major upgrade to its voice agent platform. Designed for real-time, natural interaction, this release brings in smarter turn-taking, multi-language support, and enterprise-ready features-just five months after v1.
What’s New:
Turn-Taking That Feels Human: Understands pauses, interruptions, and filler cues-knows when to talk or listen for smoother conversation flow.
Built-in RAG: Pulls accurate, up-to-date info from your knowledge base in real-time-low latency, high privacy.
Multimodal + Multi-Persona Support: Agents now speak, text, or both-plus switch between different personas as needed.
Language Auto-Detection: Understands and replies in whatever language is spoken-no setup required.
Batch Calling at Scale: Run mass outbound voice campaigns for alerts, surveys, or personalized outreach.
Enterprise-Grade Infra: HIPAA-compliant, secure, scalable, and with optional EU data residency for regulated industries.
With real-time intelligence, richer expressiveness, and broad channel support, Conversational AI 2.0 sets a new bar for voice agents in customer support, sales, healthcare, and beyond. Try it out here, or explore the docs to start building your own assistant.
A modern spin on Jürgen Schmidhuber’s Gödel Machine, this new research explores AI that rewrites its own code to improve over time-not by proof, but by evolution. By combining self-editing with open-ended search and foundation models, the Darwin Gödel Machine hints at a path toward truly self-improving, lifelong-learning AI. A must-read if you're into meta-learning or the future of agent design.
E-Tattoo That Tracks Mental Strain at Work: A new wireless forehead e-tattoo can detect and predict mental workload in real time-offering a lightweight, low-cost alternative to bulky EEG headsets. Designed for high-stakes jobs like air traffic control and surgery, this tech could revolutionize how we monitor and manage mental fatigue in the workplace. A fascinating step toward balancing performance in the brain’s “Goldilocks zone.”
This week, we’re adding two foundational courses designed to kickstart your journey into data science and databases - ideal for anyone curious about data roles or ready to dive into SQL:
Foundations of Data Science – Explore the core skills behind data science, including analysis, visualization, and machine learning. This course walks you through real-world use cases, career paths (analyst vs. engineer vs. scientist), and the tools that power today’s data-driven innovations.
Essential: SQL Skills for Data Beginners – Go from zero to SQL-ready with hands-on lessons in writing queries, managing databases, and analyzing data using SELECT, JOIN, and GROUP BY. Whether you're building tables or exploring data, this course helps you develop real SQL confidence, fast.
While the AI war continues, the fight over what fuels these models is heating up too.
The fight between Anthropic and Reddit could be interesting as there are millions of creators who feel the same way about AI models stealing their work.
What side of the fight are you on?
Reply