I think we have it: our open source Claude Code = GLM-5.1 + Pi (https://pi.dev/) - Built a Three.js racing game to eval and it's extremely impressive. Thoughts:
- One-shot car physics with real drift mechanics (this is hard)
- My fav part: Awesome at self iterating (with no vision!) created 20+ Bun.WebView debugging tools to drive the car programmatically and read game state. Proved a winding bug with vector math without ever seeing the screen
- 531-line racing AI in a single write: 4 personalities, curvature map, racing lines, tactical drifting. Built telemetry tools to compare player vs AI speed curves and data-tuned parameters
- All assets from scratch: 3D models, procedural textures, sky shader, engine sounds, spatial AI audio!
- Can do hard math: proved road normals pointed DOWN via vector cross products, computed track curvature normalized by arc length to tune AI cornering speed
You are going to hear about this model a lot in the next months - open source let's go - and thanks z-ai🚀🚀
PASD isn’t recent, but still delivers strong results — worth restoring rather than replacing.
Getting it to run again wasn’t a simple dependency issue. It relied on parts of diffusers that no longer exist, while moving to Gradio 6 forced a much newer HF stack — and I couldn’t modify the original source directly.
Recreating the old environment wasn’t practical. So I patched the downloaded code at runtime before import and made it compatible with today’s stack.
That ended up being the only approach that held without forking or freezing everything to outdated versions.
If you’ve used it before (or are curious), feel free to give it another try.
My TIGER app is now fully working again, with fixes and full compatibility with Gradio 6 🚀
It lets you: - 🎙️ Separate multiple speakers from an audio file - 🎬 Extract each speaker directly from a video - 🎧 Split audio into dialog, music, and sound effects (DnR) - 🎥 Apply DnR separation directly on videos
All powered by lightweight TIGER models for fast and efficient speech separation.
I’ve fixed the Space and brought it back to life: - ✅ Working again after being broken for a while - ✅ Updated to Gradio 6 - ✅ Compatible with ZeroGPU - ✅ Output videos now preserve original resolution and FPS
I also added advanced controls so you can experiment more (tracking, seed, motion, sketch).
I improved the public demo for TADA — a generative framework for speech modeling via text–acoustic dual alignment.
TADA models speech as a joint sequence of text tokens and acoustic tokens, using a transformer backbone to keep text and audio synchronized during generation.
The original demo already exposed these mechanisms, but the workflow made the pipeline hard to understand.
This updated demo makes the process clearer:
• load the model • prepare a reference voice (optionally with transcript or Whisper auto-transcription) • generate speech conditioned on that reference
It also adds multilingual support.
Presets are included for a few languages, but the model supports more: