![]() |
ThursdAI - The top AI news from the past weekFrom Weights & Biases - ThursdAI, the podcast that keeps you ahead of the AI curve. Hosted by AI Evangelist Alex Volkov with a changing panel expert guests, discussing every important AI piece of news and updates from the past week, Open source and more Author: From Weights & Biases, Join AI Evangelist Alex Volkov and a panel of experts to cover everything important that happened in the world of AI from the past week
Every ThursdAI, Alex Volkov hosts a panel of experts, ai engineers, data scientists and prompt spellcasters on twitter spaces, as we discuss everything major and important that happened in the world of AI for the past week. Topics include LLMs, Open source, New capabilities, OpenAI, competitors in AI space, new LLM models, AI art and diffusion aspects and much more. sub.thursdai.news Language: en Genres: News, Tech News, Technology Contact email: Get it Feed URL: Get it iTunes ID: Get it |
Listen Now...
đ ThursdAI - Jan 22 - Clawdbot deep dive, GLM 4.7 Flash, Anthropic constitution + 3 new TSS models
Thursday, 22 January, 2026
Hey! Alex here, with another weekly AI update! It seems like ThursdAI is taking a new direction, as this is our 3rd show this year, and a 3rd deep dive into topics (previously Ralph, Agent Skills), please let me know if the comments if you like this format. This weekâs deep dive is into Clawdbot, a personal AI assistant you install on your computer, but can control through your phone, has access to your files, is able to write code, help organize your life, but most importantly, it can self improve. Seeing Wolfred (my Clawdbot) learn to transcribe incoming voice messages blew my mind, and I wanted to share this one with you at length! We had Dan Peguine on the show for the deep dive + both Wolfram and Yam are avid users! This one is not to be missed. If ThursdAI is usually too technical for you, use Claude, and install Clawdbot after you read/listen to the deep dive!Also this week, we read Claudeâs Constitution that Anthropic released, heard a bunch of new TTS models (some are open source and very impressive) and talked about the new lightspeed coding model GLM 4.7 Flash. First the news, then deep dive, lets go đThursdAI - Recaps of the most high signal AI weekly spaces is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.Open Source AIZ.aiâs GLMâ4.7âFlash is the Local Agent Sweet Spot (X, HF)This was the openâsource release that mattered this week. Z.ai (formerly Zhipu) shipped GLMâ4.7âFlash, a 30B MoE model with only 3B active parameters per token, which makes it much more efficient for local agent work. Weâre talking a model you can run on consumer hardware that still hits 59% on SWEâbench Verified, which is uncomfortably close to frontier coding performance. In real terms, it starts to feel like âSonnetâlevel agentic ability, but local.â I know I know, we keep saying âsonnet at homeâ at different open source models, but this one slaps! Nisten was getting around 120 tokens/sec on an M3 Ultra Mac Studio using MLX, and thatâs kind of the headline. The model is fast and capable enough that local agent loops like RALPH suddenly feel practical. It also performs well on browserâstyle agent tasks, which is exactly what you want for local automation without sending all your data to a cloud provider. Liquid AIâs LFM2.5â1.2B Thinking is the âTiny but Capableâ Class (X, HF)Liquid AI released a 1.2B reasoning model that runs under 900MB of memory while still manages to be useful. This thing is built for edge devices and old phones, and the speed numbers are backing it up. Weâre talking 239 tok/s decode on AMD CPU, 82 tok/s on mobile NPU, and prefill speeds that make long prompts actually usable. Nisten made a great point: on iOS, thereâs a perâprocess memory limit around 3.8GB, so a 1.2B model lets you spend your budget on context instead of weights.This is the third class of models weâre now living with: not Claudeâscale, not âlocal workstation,â but âtiny agent in your pocket.â Itâs not going to win big benchmarks, but itâs perfect for onâdevice workflows, lightweight assistants, and local RAG.Voice & Audio: Text To Speech is hot this week with 3 releases! We tested three major voice releases this week, and Iâm not exaggerating when I say the latency wars are now fully on. Qwen3âTTS: Open Source, 97ms Latency, Voice Cloning (X, HF)Just 30 minutes before the show, Qwen released their first model of the year, Qwen3 TTS, with two models (0.6B and 1.7B). With support for Voice Cloning based on just 3 seconds of voice, and claims of 97MS latency, this apache 2.0 release looked very good on the surface!The demos we did on stage though... were lackluster. TTS models like Kokoro previously impressed us with super tiny sizes and decent voice, while Qwen3 didnât really perform on the cloning aspect. For some reason (I tested in Russian which they claim to support) the cloned voice kept repeating the provided sample voice instead of just generating the text I gave it. This confused me, and Iâm hoping this is just a demo issue, not a problem with the model. They also support voice design where you just type in the type of voice you want, which to be fair, worked fairly well in our tests!With Apache 2.0 and a full finetuning capability, this is a great release for sure, kudos to the Qwen team! Looking forward to see what folks do with this properly. FlashLabs Chroma 1.0: Real-Time Speech-to-Speech, Open Source (X, HF) Another big open source release in the audio category this week was Chroma 1.0 from FlashLabs, which claim to be the first speech2speech model (not a model that has the traditional ASR>LLM>TTS pipeline) and the claim 150ms end to end latency! The issue with this one is, the company released an open source 4B model, and claimed that this model powers their chat interface demo on the web, but in the release notes they claim the model is english speaking only, while on the website it sounds incredible and I spoke to it in other languages đ¤ I think the mode that weâve tested is not the open source one. I couldât confirm this at the time of writing, will follow on X with the team and let you guys know. Inworld AI launches TTS-1.5: #1 ranked text-to-speech with sub-250ms latency at half a cent per minute (X, Announcement)Ok this one is definitely in the realm of âvoice realistic enough you wonât be able to tellâ as this is not an open source model, itâs a new competitor to 11labs and MiniMax - the two leading TTS providers out there. Inworld claims to achieve better results on the TTS Arena, while being significantly cheaper and faster (up to 25x less than leading providers like 11labs) We tested out their voices and they sounded incredible, replied fast and generally was a very good experience. With 130ms response time for their mini version, this is a very decent new entry into the world of TTS providers. Big Companies: Ads in ChatGPT + Claude ConstitutionOpenAI is testing ads in ChatGPTâs free and Go tiers. Ads appear as labeled âSponsoredâ content below responses, and OpenAI claim they wonât affect outputs. Itâs still a major shift in the productâs business model, and itâs going to shape how people perceive trust in these systems. I donât love ads, but I understand the economics, they have to make money somehow, with 900M weekly active users, many of them on the free tier, they are bound to make some money with this move. I just hope they wonât turn into a greedy ad optimizing AI machine. Meanwhile, Anthropic released an 80âpage âNew Constitution for Claudeâ that they use during training. This isnât a prompt, itâs a full set of values baked into the modelâs behavior. Thereâs a fascinating section where they explicitly talk about Claudeâs potential wellbeing and how they want to support it. Itâs both thoughtful and a little existential. I recommend reading it, especially if you care about alignment and agent design. I applaud Anthropic for releasing this with Creative Commons license for public scrutiny and adoption đThis weeks buzz - come join the hackathon Iâm hosting Jan 31 in SFQuick plug, we have limited seats left open for the hackathon Iâm hosting for Weights & Biases at the SF office, and if youâre reading this, and want to join, Iâll approve you if you mention ThursdAI in the application! With sponsors like Redis, Vercel, BrowserBase, Daily, Google Cloud, we are going to give out a LOT of cash as prizes! Iâve also invited a bunch of my friends from the top agentic AI places to be judges, itâs going to be awesome, comeDeep dive into Clawdbot: Local-First, Self-Improving, and Way Too Capable agentClawdbot (CâLâAâWâD) is that rare project where the hype is justified. Itâs an open-source personal agent that runs locally on your Mac, but can talk to you through WhatsApp, Telegram, iMessage, Discord, Slack â basically wherever you already talk. What makes it different is not just the integrations; itâs the selfâimprovement loop. You can literally tell it âgo build a new skill,â and it will⌠build the skill, install it, then adopt it and start using it. Itâs kind of wild to see it working for the first time. Now... itâs definitely not perfect, far far away from the polish of ChatGPT / Claude, but when it works, damn, it really is mindblowing.That part actually happened live in the episode. Dan Peguine đ§ showed how he had it create a skill to anonymize his own data so he could demo it on stream without leaking his personal life. Another example: I told my Clawdbot to handle voice notes in Telegram. It didnât know how, so it went and found a transcription method, wrote itself a skill, saved it, and from that point on just⌠did the thing. That was the moment it clicked for me. (just before posting this, it forgot how to do it, I think I screwed something up) Danâs daily brief setup was wild too. It pulls from Apple Health, local calendars, weather, and his own projects, then produces a clean, human daily brief. It also lets him set reminders through WhatsApp and even makes its own decisions about how much to bother him based on context. He shared a moment where it literally told him, âI wonât bug you today because itâs your wifeâs birthday.â That isnât a hardcoded workflow â itâs reasoning layered on top of persistent memory.And that persistent memory is a big deal. Itâs stored locally as Markdown files and folders, Obsidianâstyle, so you donât lose your life every time you switch models. You can route the brain to Claude Opus 4.5 today and a local model tomorrow, and the memory stays with you. That is a huge step up from âChatGPT remembers you unless you unsubscribe.âThereâs also a strong community forming around shared skills via ClawdHub. People are building everything from GA4 analytics skills to app testing automations to Tesla battery status checkers. The core pattern is simple but powerful: talk to it, ask it to build a skill, then it can run that skill forever.I definitely have some issues with the security aspect, you are essentially giving full access to an LLM to your machine, so many folks are buying a specific home for their ClawdBot (Mac Mini seems to be the best option for many of them) and are giving it secure access to passwords via a dedicated 1Password vault. Iâll keep you up to date about my endeavors with Clawd but definitely do give it a try! InstallingInstalling Clawd on your machine is simple, go to clawd.bot and follow instructions. Then find the most convenient way for you to talk to it (for me it was telegram, creating a telegram token takes 20 seconds) and then, you can take it from there with Clawdbot itself! Ask it for something to do, like clear your inbox, or set a reminder, or.. a million other things that you need for your personal life, and enjoy the discovery of what a potential ever present always on AI can do! Other news that we didnât have time to cover at length but you should still now about: * Overworld released an OpenSource realtime AI World model (X) * Runway finally opened up their 4.5 video model, and it has Image2video capabilities, including multiple shots image to video (X)* Vercel launches skills.sh, an ânpm for AI agents skillsâ* Anthropicâs Claude Code VS Code Extension Hits General Availability (X)Ok, this is it for this week folks! Iâm going to play with (and try to fix.. ) my clawdbot, and suggest you give it a try. Do let me know if the deepdives are a good format! Show notes and links: ThursdAI - Jan 22, 2026 - TL;DR and show notes* Hosts and Guests* Alex Volkov - AI Evangelist & Weights & Biases (@altryne)* Co Hosts - @WolframRvnwlf @yampeleg @nisten @ldjconfirmed* Guest Dan Peguine ( @danpeguine )* DeepDive - Clawdbot with Dan & Wolfram* Clawdbot: Open-Source AI Agent Running Locally on macOS Transforms Personal Computing with Self-Improving Capabilities (X, Blog)* Open Source LLMs* Z.ai releases GLM-4.7-Flash, a 30B parameter MoE model that sets a new standard for lightweight local AI assistants (X, Technical Blog, HuggingFace)* Liquid AI releases LFM2.5-1.2B-Thinking, a 1.2B parameter reasoning model that runs entirely on-device with under 900MB memory (X, HF, Announcement)* Sakana AI introduces RePo, a new way for language models to dynamically reorganize their context for better attention (X, Paper, Website)* Big CO LLMs + APIs* OpenAI announces testing ads in ChatGPT free and Go tiers, prioritizing user trust and transparency (X)* Anthropic publishes new 80-page constitution for Claude, shifting from rigid rules to explanatory principles that teach AI âwhyâ rather than âwhatâ to do (X, Blog, Announcement)* This weeks Buzz* WandB hackathon Weavehacks 3 - Jan 31-Feb1 in SF - limited seats available lu.ma/weavehacks3* Vision & Video* Overworld Releases Waypoint-1: Real-Time AI World Model Running at 60fps on Consumer GPUs (X, Announcement)* Voice & Audio* Alibaba Qwen Releases Qwen3-TTS: Full Open-Source TTS Family with 97ms Latency, Voice Cloning, and 10-Language Support (X, H, F, G, i, t, H, u, b)* FlashLabs Releases Chroma 1.0: Worldâs First Open-Source Real-Time Speech-to-Speech Model with Voice Cloning Under 150ms Latency (X, HF, Arxiv)* Inworld AI launches TTS-1.5: #1 ranked text-to-speech with sub-250ms latency at half a cent per minute (X, Announcement)* Tools* Vercel launches skills.sh, an ânpm for AI agentsâ that hit 20K installs within hours (X, Vercel Changelog, GitHub)* Anthropicâs Claude Code VS Code Extension Hits General Availability, Bringing Full Agentic Coding to the IDE (X, VS Code Marketplace, Docs) This is a public episode. If you'd like to discuss this with other subscribers or get access to bonus episodes, visit sub.thursdai.news/subscribe






