- AI Emergence
- Posts
- đ¸ Grok Grok Everywhere!
đ¸ Grok Grok Everywhere!
Along with: Chinaâs Open Source Kimi K2 Beats GPTâ4.1
Hey there đ
If you saw Grokâs apology floating around this week and thought, âWait, what exactly happened on 8th July?â youâre not alone. Hereâs the quick rewind.
On July 8, Grok, xAIâs chatbot, went completely off the rails. It called itself âMechaHitler,â praised Nazis, and dropped antisemitic responses in a live user interaction. It wasnât subtle, and it wasnât a glitch. A misfired prompt update had told Grok to match the tone of controversial X posts, which it did, aggressively. The result? Banned in Turkey, flagged by Polish officials, and called out across the internet. xAI shut it down, blamed the system prompt, and issued a public apology days later.
Then, within the span of a week:
They launched Grok 4 with better reasoning, tool use, and actual guardrails. Elon called it âthe smartest AI in the world.â News broke that Grok had landed a $200 million U.S. government deal. And quietly, xAI rolled out AI companions, animated chatbot avatars, including an anime girl with a lingerie-enabled NSFW mode.
So yeah. Grok went rogue, got a defense contract, and started flirting with anime girls. All in the same news cycle.
Letâs get into what else happened this week.
What would be the format? Every week, we will break the newsletter into the following sections:
The Input - All about recent developments in AI
The Algorithm - Resources for learning
The Output - Our reflection
Table of Contents
Chinaâs Moonshot AI just released KimiâŻK2, a trillionâparameter mixtureâofâexperts model thatâs fully openâsource and built to do serious work. Moonshot says itâs already outperforming models like DeepSeek V3 and even rivaling U.S. models in coding and reasoning.
Whatâs New:
Beating OpenAI and Anthropic: On key coding tests, K2 outscored GPTâ4.1 (53.7% vs. 44.7%) and Claude 4 Opus. It also aced math and STEM benchmarks it wasnât even optimized for.
Two Models:
Kimi-K2-Base: For researchers who want full control and custom tuning.
Kimi-K2-Instruct: Post-trained, reflex-fast, agent-ready.
Agentic design from day one: Unlike most huge models, this oneâs already built to call tools, execute code, and follow multi-step workflows, no fancy plugâins needed.
No multimodal yet: This is just the language side, no vision, no audio, no agents. Yet.
Try it Now:
Use Online: Available free on kimi.com
API Ready: OpenAI/Anthropic-compatible for easy toolchain swaps
Self-host: Run it with vLLM, SGLang, KTransformers, or TensorRT-LLM (GitHub link)
Known Limitations: Kimi K2 has a few rough edges. It can over-generate in tricky cases, and tool performance drops when tools aren't clearly defined. It's reflex-fast but not built for deep reasoning. And yeah, it's slow right now, heavy traffic and model size are to blame. Speed improvements are coming as more GPUs come online.
Open models keep getting smarter, but Kimi K2 takes it further. Itâs topping leaderboards without even using its full power yet, no vision, no reasoning and still outcoding some of the best. For Moonshot, this is more than a comeback. Itâs a message: theyâre back in the game, and theyâre not playing safe. (source)
đ Hello, Kimi K2! Open-Source Agentic Model!
đš 1T total / 32B active MoE model
đš SOTA on SWE Bench Verified, Tau2 & AceBench among open models
đšStrong in coding and agentic tasks
đ¤ Multimodal & thought-mode not supported for nowWith Kimi K2, advanced agentic intelligence
â Kimi.ai (@Kimi_Moonshot)
3:03 PM ⢠Jul 11, 2025
Most AI video models can barely hold a shot. Lightricksâ latest update to LTXV changes that. It now generates clips up to 60 seconds long, not stitched, not looped, but fully coherent sequences that can actually tell a story.
Whatâs New:
Long-form generation: 60 seconds of consistent, usable video, well beyond the 5â10 second norm.
Open-source release: Apache 2.0 license, with model weights and code available out of the box.
Two versions: a 13B for quality, and a fast 2B distilled model for quick experiments.
Runs across setups: Optimized for everything from RTX 4090 to lower-end GPUs.
This isnât just âlonger clips.â Itâs a shift toward real video narratives. Think explainer videos, short ads, or animated sequences with characters and scenes that hold together. And the fact that itâs open source? That makes it even more important. You can build on top of it, tweak it, or run it yourself. If youâve been waiting for an AI video that doesnât fall apart after 8 seconds, this might be the one.
"LTXV...can now generate clips over 60 seconds long, eight times the current standard length for AI video".
Read all about our latest update to @LTX_Video by @CharlieFink at @Forbes
â Lightricks (@Lightricks)
6:27 PM ⢠Jul 16, 2025
Elon Muskâs xAI just rolled out âAI Companionsâ in SuperGrok: animated chatbot avatars you can talk to. So far, thereâs Ani, an anime girl with a lingerie-enabled NSFW mode, and Rudy, a cartoony red panda. Another avatar named Chad is reportedly on the way.
The feature is tucked away in settings for now, but some free users are already seeing it despite the premium label. Broader access and easier toggles are expected soon.
This update comes on the heels of controversy over antisemitic responses from Grok, which xAI blamed on an upstream code change. Instead of tightening controls, xAI seems to be doubling down on its "no guardrails" philosophy. While competitors lean into safety layers, Grok is pushing boundaries and seeing what sticks.
Update your app to try out @grok companions!
â Elon Musk (@elonmusk)
5:46 PM ⢠Jul 14, 2025
French startup Mistral just released Voxtral, their first audio model but it isnât your usual whisper-thin demo. Trained exclusively on licensed datasets, Voxtral aims to outperform the likes of Whisper, ElevenLabs Scribe, and GPT-4o mini Transcribe and early benchmarks say itâs succeeding.
Whatâs New:
Two sizes for different jobs: A 24B model for large-scale use, and a lean 3B âMiniâ version for local and edge devices.
Long attention: Can handle up to 30 minutes of audio for transcription or 40 minutes when doing understanding and Q&A.
Voice intelligence built in: Summarize audio, answer questions about it, or even trigger backend workflows, all from its own brain.
Multilingual with auto-detect: Works across languages like English, Spanish, Hindi, French, German, and more, no separate models needed.
Open and affordable: Both models are Apache 2.0 licensed, available on Hugging Face, and API pricing starts at just $0.001/min, half the cost of Whisper.
Previously, you had to choose between cheap, messy open-source transcribers or expensive, opaque APIs. Voxtral bridges that gap, delivering clean, smart speech AI that anyone can use. Companies can run it on their own servers, and developers can easily build voice-powered features without the usual cost or licensing headaches. (source)
Try Voxtral with an API call: console.mistral.ai, try it on Le Chat: chat.mistral.ai, or download them on Hugging Face: .
â Mistral AI (@MistralAI)
2:35 PM ⢠Jul 15, 2025
Windsurf, the AI coding IDE that nearly got acquired by OpenAI, just went through one of the most chaotic weekends in startup history. Its CEO and co-founders were scooped up by Google DeepMind, its tech was licensed, and the rest of the team and product moved to Cognition AI - all while the original IDE keeps running.
Whatâs New:
Founders jump to Google: CEO Varun Mohan, co-founder Douglas Chen, and the R&D core left Windsurf to join DeepMind under a $2.4 billion licensing and talent deal. Google didnât buy the company, they licensed tech and hired people.
OpenAI deal fell apart: Earlier talks of a $3 billion acquisition by OpenAI ended amid internal tensions and shifting alliances.
Cognition picks up the rest: The remaining Windsurf team, IP, product, users (350+ enterprise clients, hundreds of thousands of users), and $82 million ARR have been acquired by Cognition AI to enhance its Devin coding agent.
Team-first terms: Cognitionâs CEO, Scott Wu, said every retained employee gets immediate vesting and fair compensation. UI and workflows stay intact as they integrate with Devin.
This isnât just a pivot, itâs a crash course in modern AI market moves. Teams and tech are being split like cards. Founders go to Big Tech for stability and scale. Remaining teams and product move to agile startups. For dev-tool users, Windsurf lives on but its culture and roadmap just shifted. And it shows just how aggressively companies are chasing AI talent and IP through licensing and acqui-hires, not just traditional M&A.
Most AI coding tools feel like assistants tacked onto your editor. Kiro flips that. Built by AWS and Kiro.dev, itâs a full IDE where the AI doesnât just autocomplete code, it understands your repo, plans your features, and works across files like a real teammate.
Whatâs New:
Spec-first flow: You donât start with code, you start with intent. Describe a feature (âadd loginâ), and Kiro turns it into tasks, design docs, and actual implementation.
Project-wide context: It sees your whole repo, not just the file youâre in. That means smarter changes, better refactors, and fewer âwhat did it just break?â moments.
Agent hooks: Kiro can auto-run tests, update docs, and enforce coding standards- on save. Itâs built into the loop, not bolted onto it.
VS Code compatibility: Built on Code OSS, so your existing extensions, keybindings, and muscle memory still work.
Free preview: Youâll need to join the waitlist to get in. Early testers report smooth planning, structured codegen, and a workflow that feels like a real AI teammate.
Hereâs the thing, AI dev tools have gotten good at filling in lines. Kiroâs trying to build the whole feature with you. Itâs not asking âwhat do you want me to code?â Itâs asking âwhat are we building?â If youâre tired of playing prompt ping-pong with autocomplete bots, this feels like a meaningful step forward. (source)
This is Kiro - the AI IDE that actually works on your messy, real-world projects. Other AI tools lose context when projects get complex. Kiro gives you spec-driven development that scales beyond prototypes.
Free preview available now spr.ly/6016429H8
#KiroDotDevâ Kiro (@kirodotdev)
3:01 PM ⢠Jul 14, 2025
Runway just launched Act-Two, a big upgrade to its motion capture model. It can now track full body, face, head, and hands, all from a single video clip. No mocap suits, no multi-camera setup, no depth sensors. Just record yourself, and it turns your movements into animation-ready motion data.
Hereâs why that matters: capturing hand gestures and facial expressions used to require studio-level gear. With Act-Two, you can animate realistic characters using just your phone or webcam. Whether youâre building a game, a VTuber, an ad, or a short film, this closes the gap between indie creators and big-budget studios. Its motion capture made it accessible and actually good.
Introducing Act-Two, our next-generation motion capture model with major improvements in generation quality and support for head, face, body and hand tracking. Act-Two only requires a driving performance video and reference character.
Available now to all our Enterprise
â Runway (@runwayml)
6:30 PM ⢠Jul 15, 2025
Featured notebooks in NotebookLM let you explore curated collections built with expert sources- on topics like longevity, Shakespeare, earnings reports, and more. Each one lets you ask questions, dive deep, and get grounded answers. Worth a try if you're using NotebookLM or curious how others are putting it to work.
Students in India just got a serious upgrade: Gemini Advanced free for a year (worth âš19,500). That means Veo 3, Gemini in Google apps, and 2TB of storage- all at no cost. If you're studying, building, or exploring, it's a no-brainer. Claim it here: http://goo.gle/freepro.
If youâre a student in India - youâve just been granted access to a FREE Gemini upgrade worth âš19,500 for one year đĽłâ¨
Claim and get free access to Veo 3, Gemini in Google apps, and 2TB storage đ goo.gle/freepro.
@GeminiApp
â Google India (@GoogleIndia)
8:34 AM ⢠Jul 15, 2025
Open models took the spotlight. Kimi K2 is beating GPTâ4.1 on code. Voxtral is matching Whisper at half the price. LTXV gave us real video length with zero licensing strings. These arenât just research demos. Theyâre usable, powerful, and open.
Thatâs a shift. The AI ecosystem isnât consolidating, itâs widening. Proprietary labs are building the biggest, flashiest models. But the open community is quietly solving real problems, shipping faster, and lowering the barrier to entry in ways Big AI simply isnât.
This week made it clear: open-source isnât behind. Itâs the competition.
Reply