Tactical, practitioner-focused content for people building AI products — model selection, infrastructure, prompt engineering, and what actually works in production.
“does model selection really matter? I mean, are your clients like picking, oh, okay, I'm going to use this Lama model of this size or Quinn or do they make those decisions? Or is this something that's kind of like handled automatically by the autonomous intelligence system? So model selection matters, but it's also exhausting. So funny thing. There are two interesting phenomena that's so unique to AI. One is the model depreciation is very fast. As you can observe, every couple of weeks, there's …”
Ridealong summary
Model selection in AI is not just important; it's exhausting due to rapid changes in technology. With new models and hardware launching almost weekly, developers struggle to choose the best fit for their evolving use cases. This dual challenge of model and hardware depreciation creates a significant headache for application developers trying to keep up.
Super Data Science: ML & AI Podcast with Jon Krohn·971: 90% of The World’s Data is Private; Lin Qiao’s Fireworks AI is Unlocking It·Mar 03, 2026
“Jen, you just came back from Mexico. You were talking about your dogs. There's a lot of new paperwork you have to have. And so there's this story going viral of this woman that went to an airport with her dog. And then I'll let the video play. It'll explain it. We're back with a dog left alone at the airport in Las Vegas. Video shows the woman leaving the Golden Doodle Mini Poodle at a JetBlue ticket counter earlier this month. She didn't have the right documents to travel with it as a service …”
Ridealong summary
A woman recently abandoned her dog, a Golden Doodle Mini Poodle named Jet Blue, at a JetBlue ticket counter in Las Vegas due to lack of proper travel documents. The dog was rescued and adopted by the officer who found him, but the incident raises questions about irresponsible pet ownership and societal trends, such as performative hydration with trendy cups. This story highlights the emotional impact of such neglect and the importance of responsible pet care.
“And one of the extremely controversial things we did was we felt that we should focus on fine-tuning existing models because we thought fine-tuning was important. Some other folks were doing work contemporaneously with that. So Jason Ussinsky did some really great research. I think it was during his PhD on how to fine-tune models and how good they can be and some other folks in the computer vision world. We were, you know, amongst the first. There was a bunch of us kind of really investing in …”
Ridealong summary
Fine-tuning existing models can drastically improve performance, especially when using discriminative learning rates. Researchers discovered that training only the last few layers significantly speeds up the process, challenging previous assumptions about learning rates. This approach, combined with the necessity of fine-tuning batch normalization layers, has transformed how we think about transfer learning.
Machine Learning Street Talk (MLST)·"Vibe Coding is a Slot Machine" - Jeremy Howard·Mar 03, 2026
“holding multiple possibilities in parallel backtracking when needed and converging on solutions that satisfy all rules simultaneously These are precisely the skills needed for countless real challenges in medicine law operations planning and many other spaces domains where you're balancing competing constraints under uncertainty. A system that can reason through these spaces natively, rather than forcing everything into a text-based chain of thought, could eventually do more than summarize …”
Ridealong summary
A groundbreaking architecture called BDH achieves a staggering 97.4% success rate on Sudoku puzzles, while traditional transformer models fail at 0%. This shift highlights BDH's potential to revolutionize AI capabilities beyond mere information summarization, enabling strategic reasoning in complex fields like medicine and law. With evidence mounting against transformers, the future of AI could be on the brink of a significant transformation.
Super Data Science: ML & AI Podcast with Jon Krohn·A Post-Transformer Architecture Crushes Sudoku (Transformers Solve ~0%)·Mar 27, 2026
“The biggest moat is going to be which companies understand something that's super hard for other people to understand. And if your answer to that is, I don't know, then you maybe could get vibe coded away. Block was one of the first to make a pretty drastic decision in cutting 40% of the workforce. What led up to that decision? There's been this correlation between the number of folks at a company and the output from the company for decades and decades. I think that basically broke. And what …”
Ridealong summary
Block dramatically cut 40% of its workforce to test a radical new approach: restructuring around small teams and AI agents. This shift has shown that a few engineers can now achieve productivity levels previously thought impossible, reshaping the company's future. With tools like BuilderBot, Block is redefining what it means to build software efficiently.
The a16z Show·What Happens When a Public Company Goes All In on AI·Apr 01, 2026
“But you've mentioned a few times the communication. I know you're a big voice user, so I'd love to hear what your voice setup looks like or if there's anything special that you've learned that think people should follow your example on. And then you mentioned Slack too. And I'm curious about even just such practical details. Is it like one to one chats with five open calls? Or is it one channel where you like tag which one you want to assign to things and, but they can all see what's going on. …”
Ridealong summary
Using AI agents effectively requires a strategic communication setup. By leveraging platforms like Slack, the speaker has created a unique channel system that enhances team coordination and agent interaction. This innovative approach not only simplifies communication but also ensures that tasks are managed efficiently through context-aware prompts.
"The Cognitive Revolution" | AI Builders, Researchers, and Live Player Analysis·Try this at Home: Jesse Genet on OpenClaw Agents for Homeschool & How to Live Your Best AI Life·Mar 08, 2026
“um let's let's shift gears now a little bit chris to talking about inference specifically uh you know we talked earlier in this episode about how inference is most of what's happening at scales with ai models training is an important part but inference is most of what we're doing with gpus these days and so you talk in the book about multi-node inference and disaggregated architectures. What the heck are those? Yeah. Welcome to another chapter that was supposed to be a single chapter or a topic …”
Ridealong summary
AI inference is more complex than it seems, involving intricate processes that ensure quick responses from models like ChatGPT. While the concept of a simple forward pass sounds straightforward, it actually requires sophisticated caching strategies to avoid repetitive calculations. This is why understanding multi-node inference and KV cache is crucial for enhancing AI performance.
Super Data Science: ML & AI Podcast with Jon Krohn·973: AI Systems Performance Engineering, with Chris Fregly·Mar 10, 2026
“They basically said, hey, what if we gave it more experts? So we're going to use more memory capacity, but we keep the amount of activated experts the same. We increase the expert sparsity. So we have fewer experts. The ratio of experts activated to number of experts is smaller. And we decrease the number of attention heads. And kind of for context, what we had been seeing was you make models sparser instead. So no one was really touching heads. You're just having. Well, they implicitly made it …”
Ridealong summary
Increasing expert sparsity in model training can enhance efficiency without sacrificing performance. By adjusting the number of activated experts and attention heads, researchers are discovering new ways to optimize models for specific tasks. This approach highlights the importance of hardware and model co-design in achieving better outcomes.
Latent Space: The AI Engineer Podcast·NVIDIA's AI Engineers: Agent Inference at Planetary Scale and "Speed of Light" — Nader Khalil (Brev), Kyle Kranen (Dynamo)·Mar 10, 2026
“Here's my pitch from a computer science point of view. Pretty rare if people ask me this question. That is, if you're working at a vertically integrated company that have an end product for, let's say, for chatbots, for a system, you are working on the vertical slice of the problem. In Infract, you will be working on a abstraction of horizontal layer. And this is similar to operating system, databases, and different kinds of abstraction that people have built over the years. Operating system, …”
Ridealong summary
The future of machine learning lies in creating a universal inference layer that abstracts the complexities of accelerated computing devices. This innovative approach parallels the evolution of operating systems and databases, enabling developers to build more efficient systems for AI applications. At Infract, we are pioneering this essential layer, pushing the boundaries of inference technology.
The a16z Show·Inferact: Building the Infrastructure That Runs Modern AI·Jan 22, 2026
“This is great news. To form a joint venture that will distribute enterprise products across the firm's portfolio companies and beyond. the proposed deal as a free money valuation. AI is coming to Fogo de Chao. Bain Capital owns Fogo de Chao, the Brazilian steakhouse. Maybe now they can take Apple Pay. Oh yeah, we got cooked on that. Maybe they deliberately don't. I wonder if Apple Pay is expensive for them and this is actually a cost consideration. Fiji Simo said, this news came out a little …”
Ridealong summary
OpenAI is launching a joint venture to distribute enterprise AI products, a move fueled by a skyrocketing demand from over a million businesses. This initiative, which includes embedding engineers into companies, aims to help enterprises effectively deploy AI solutions. With API usage surging and a dedicated deployment arm in the works, OpenAI is sprinting to meet the urgent needs of the market.
TBPN·OpenAI Ends Side Quests, SF Housing Market is Back, Kalshi’s $1B Prize | Diet TBPN·Mar 17, 2026
“If you fuse too aggressively you make the kernel too complex and that blows your memory register your memory budget and ironically it makes it slower So the AI has to look at your whole graph and plan of your computations to actually execute this properly And anyway there like a whole bunch of challenges associated with this The the the art of knowing how to let data just flow through your system and, and how to spin up, uh, spin up kernels that do what you want efficiently is just really, …”
Ridealong summary
A groundbreaking approach uses reinforcement learning to optimize CUDA kernels, potentially surpassing traditional methods. By automatically generating 6,000 training problems from PyTorch, the AI learns to write, compile, and improve CUDA code efficiently. This innovative method could change how developers tackle kernel optimization challenges in AI research.
Last Week in AI·#237 - Nemotron 3 Super, xAI reborn, Anthropic Lawsuit, Research!!!·Mar 16, 2026
“Take me through DepreciationGate. How did you process that, and where do we stand now with the fear that GPUs will depreciate precipitously? And H100s will be worthless in 6 to 12 months. It's totally not a problem right now. Like CoreWeave has talked about these things are lasting 5 to 6 years. And they're getting like almost 90, 95% of the pricing. So it could potentially be a problem if this is a bubble. I don't think it's a bubble. But if it's a bubble 2, 3 years from now and there's a …”
Ridealong summary
Currently, fears of GPUs depreciating drastically in value are unfounded, as demand for AI compute is outpacing supply. Companies like CoreWeave report that even older GPUs are still in high demand, with rental prices remaining strong. However, if a compute glut occurs in the next few years, it could change the landscape dramatically.
TBPN·The Lawyer Who Beat Meta and Google, Revisiting The Jetsons, Japan Twitter | Tae Kim, Logan Bartlett, Sam Stephenson, Ben Broca, Brett Adcock, Andrei Serban·Mar 30, 2026
“there enough uh victory giant is there enough pcb this is like one of the largest suppliers of pcbs to nvidia and they're a chinese company all the all the pcbs come from china sort of from them um or many of them and anyways they're like do you have enough pcb capacity great oh hey uh memory vendors who has all the memory capacity okay nvidia does great um so when you look at sort of in the same way you know who who is agi pilled enough to buy compute and long timelines at levels that seem …”
Ridealong summary
NVIDIA is outpacing competitors like Google and Amazon in the AI chip market due to its aggressive data center expansion and better supply chain management. While Google struggles to deploy enough TPUs, NVIDIA is capitalizing on the growing demand for GPUs, positioning itself as a leader in accelerated computing. This dynamic highlights the critical role of semiconductor supply chains in the race for AI dominance.
Dwarkesh Podcast·Dylan Patel — Deep Dive on the 3 Big Bottlenecks to Scaling AI Compute·Mar 13, 2026
“And anyway, a lot to say from the first hand experience of using those models, but for the sake of speed, we should probably move on. And next model release coming from Google, also fast model, Gemini 3.1 Flash Lite, getting an improvement in both cost and speed. They say this is 2.5x faster time to first token. And first token time, by the way, when you're using fast models, typically you want them for shorter tasks. At least often that's the case. You want to have a quick output with a short …”
Ridealong summary
Google’s Gemini 3.1 Flash Lite model boasts a staggering 2.5x faster time to first token, making it a game-changer for quick tasks. With a 45% increase in overall output speed, it delivers 360 tokens per second, creating an almost instantaneous experience. This leap in speed not only enhances user perception but also significantly improves operational efficiency, making it a vital update for interactive products.
Last Week in AI·#236 - GPT 5.4, Gemini 3.1 Flash Lite, Supply Chain Risk·Mar 12, 2026
“Speaking of AI stuff, I noticed you all have both an MCP and a llms.txt, which is just the docs in a text format. What's the MCP doing for you all? It's just doc search. I believe it's using the Cloudflare AI gateway and then Autorag pointed at the doc site itself. So just an easier way, you know, I think if you just like pointed at that server and tell it to install Varlock and do the onboarding, it'll do a pretty good job. But I probably need to retest that because it's been a few weeks. …”
Ridealong summary
The MCP enhances document search by leveraging the Cloudflare AI gateway, making it easier to access and manage documentation. This tool simplifies onboarding and server installations, proving invaluable for tech startups like the one discussed at Ben Vinegar's event. The integration with GitHub actions also streamlines environment variable management, a common pain point for developers.
Syntax - Tasty Web Development Treats·985: Stop putting secrets in .env·Mar 09, 2026
“And they are making physical changes to the booster to help with that exact recovery process, right? Yes. Booster 19 has new hardware. They installed three grid fins that are 50% larger than the previous generation. 50% larger? That's a massive aerodynamic change. It is. And they are mounted lower on the hull and equipped with specific lifting pins. Those lifting pins are the contact points for the mechanical arms. Correct. The entire system is being optimized for rapid capture and turnaround. …”
Ridealong summary
Booster 19 features 50% larger grid fins designed to enhance its recovery process, aiming for a safer long-term reusability. However, engineers will only attempt land catches after two successful ocean landings, prioritizing safety over speed. This careful approach minimizes risks while optimizing the entire system for rapid turnaround.
“So while the PRD is generating, can we take a look at that Cloud MD file? Yep. Let me pull it up for you now. And that trick, guys, that he's doing there to open preview, make sure you do that because it's a pain to look at markdown files with all of those pound symbols. Annoying. Cursor, it's actually on their forums. You'll see me on the forums whinging about it every month. don't know why they can't fix it but there used to be a setting that by default it would open up in preview anyway but …”
Ridealong summary
To enhance your Product Requirements Document (PRD), leveraging AI tools like Claude can streamline the process and improve efficiency. By using a structured Claude MD file and session start hooks, product managers can ensure adherence to best practices while maintaining version control through GitHub. This approach not only tidies up documentation but also maximizes the potential of AI in product management.
The Growth Podcast·This CPO Uses Claude Code to Run his Entire Work Life | Dave Killeen, Field CPO @ Pendo·Mar 11, 2026
“Things were a bit crass back then. Also, speaking of crass, this editorial at the beginning from John sure hits a little different in this modern era of hardware being like wildly unaffordable. And like, here he is going like, hey, make games that require this brand new $500 graphics card. Like, stop supporting the entire rest of the market. Well, but it's, hey... I get it. I mean... Like this was before crisis, right? We didn't know how bad it would be if you released a game that wouldn't run …”
Ridealong summary
The GeForce 3 graphics card revolutionized game development when it launched in 2001, paving the way for advanced technologies like ray tracing. At the time, its $500 price tag was a gamble that many developers took, not knowing how it would affect the industry's future. This rapid evolution in graphics technology led to a compressed timeline of innovations that shaped the gaming landscape we know today.
Brad & Will Made a Tech Pod.·329: A Plaid Decade·Mar 08, 2026
“Or that could be like a cloud skill or something. It could be anything. Yeah. Anything that's writing down steps. Skills can be tools, though, too. I'm less familiar with how Cloud does the skills. But then the last piece is memory right So and memory right now is in the form of files So whenever you talk to your club bot or whatever you calling these things is just it writing down these markdown files and it reading them So the analogy people use is that movie Memento where the guy wakes up …”
Ridealong summary
AI agents are struggling with memory retention, akin to Alzheimer's patients missing chunks of their past. As these models rely on text files for memory, they face limitations in processing long sequences, leading to mistakes and lost information. Exploring alternatives like graph databases could be the key to enhancing AI memory capabilities.
TFTC: A Bitcoin Podcast·#726: Mapping The Mind Of The Machine with Brian Murray & Paul Itoi·Mar 14, 2026
“I do think the cursors and the UI makes a huge difference. Because if I use cloud code, I can theoretically do this too, but it's just generally a bunch of code that I can't see. But being able to see this happen is like a game time changer. It's actually really hard. I mean, sure, a lot of these platforms that are now going into parallel mode, you know? Yeah. They're like assigning different roles to different agents and so on. But what I wanted really to do is to have parallelism with same …”
Ridealong summary
Parallelism in AI design agents can drastically speed up processes by allowing multiple agents to work simultaneously on the same task without conflict. This approach not only enhances efficiency but also provides visibility into each agent's contributions, making the workflow easier to manage. Imagine instructing AI agents to build something while you grab coffee, all while knowing they have a clear plan in front of you.
Behind the Craft·I Watched 6 AI Agents Design an App In Real Time And It Blew My Mind | Tom Krcha·Mar 08, 2026
“Wait for AMD to launch the RDNA 5 96 gigabyte cards mid-2027. Pre-order 3,000 cards. Hopefully we can negotiate for $2,500 each. Build $520,000 tiny boxes with six of the cards in each box. Run all the Chinese LLMs. make $600,000 per month revenue selling tokens on OpenRouter. Market depth is there. This is 1% of OpenRouter. Improvements to TinyGrad yield revenue improvements. Due to how power is priced in Oregon, it's only like $50,000 for the electric bill before the four megawatts, before …”
Ridealong summary
Investing in AMD's upcoming RDNA 5 cards could revolutionize AI token sales, generating up to $600,000 monthly. With low electricity costs and high demand for tokens, this venture promises rapid returns and scalability, marking a pivotal moment in AI technology akin to the internet's rise. The landscape is shifting, and those who recognize the opportunity now stand to benefit immensely.
“And then I have one more question, which is, do you have some go-to like test prompts? I love the astronaut as a prompt because there's a lot of ways you could generate an astronaut. not do you have some like go-to's that you run through when you're doing a mood board or does it really depend on your client and what you're working on yeah I would say I love doing like ethereal female model for some reason I do like I do that a lot yeah this is like a crazy silver one that I was doing I tend to …”
Ridealong summary
Using specific prompts like 'ethereal female model' and 'astronaut' can elevate your creative projects significantly. These targeted ideas help generate unique visuals, especially with popular subjects like cats, which have abundant training data. Discover how narrowing your focus can enhance your mood boards and overall creativity.
How I AI·Mastering Midjourney: How to create consistent, beautiful brand imagery without complex prompts | Jamey Gannon·Mar 09, 2026
“And we're learning to live with it. And we're learning to accept it. As our treasury is being, your bank account is being robbed. warning a the world's largest robbery is happening right now it is the warning flare so big it should be able to be seen from space if hospice fraud can flourish in the shadows if taxpayer money can fund legal warfare against you with your money if billions can move through programs with no accountability then the deficit we see on paper is only part of the story. …”
Ridealong summary
The most dangerous line in the federal budget isn't the interest payments; it's the growing cost of ignoring financial fraud. As taxpayer money is misused and programs operate without accountability, we face a deficit of moral courage and attention. Unless we confront these issues soon, the consequences will be dire.
The Glenn Beck Program·Is Iran Preparing to Attack California?! | Guests: Sen. Mike Lee & Bill O'Reilly | 3/12/26·Mar 12, 2026
“It's always interesting when not only is the leading product in a category changing, We're just figuring out what the categories are that we need things for MCP or skills or stuff like that. And it's all very fast moving. And that just feels to me like a very interesting time where clearly a new way of engineering is shaking out. And 2026 is clearly not the final word. Absolutely. And in fact, I'm growing more. skeptical of mcp as like a meaningful part of the the future not it's fine as a …”
Ridealong summary
Multi-agent systems are evolving, and the traditional MCP protocol may not be the key to future AI development. As we rethink how context is shared among agents, simpler solutions like OpenClaw, which utilizes markdown files, could prove more effective. This shift suggests a need for agents to possess deeper contextual understanding to enhance user experiences.
Cheeky Pint·Bret Taylor of Sierra on AI agents, outcome-based pricing, and the OpenAI board·Mar 10, 2026
“But it was really awkward. You know, I think folks like you and me back in the day used to probably dismiss Microsoft. I mean, I certainly did. I didn't I didn't foresee them becoming as powerful and strong as they are today. But it was good leadership, good technology. But I don't think the market knows like who is Siebel Systems and who is Microsoft in this like landscape of software companies. Probably no one knows what Siebel Systems was. That was the company that Salesforce beat to become …”
Ridealong summary
In the battle for AI supremacy, startups may have the upper hand over incumbents like Microsoft. While established companies are often slow to adapt to new technologies, agile newcomers can leverage their innovative capabilities to outpace them. This dynamic creates a critical window where the best of breed competitors can scale before the giants catch up.
Uncapped with Jack Altman·Uncapped #42 | Bret Taylor from Sierra·Feb 19, 2026
“more and figuring out how we can make Boston the official hub for ethical AI. Well said. You both inspire me to get on an airplane. But thankfully, to our audience and to me, we can have these conversations. Please come back and consider yourself in the Babe Tribe, Building AI Boston. We're celebrating Women's History Month, and you are certainly a woman making history. Thank you so much for being here. Thank you so much for having me. This was so enjoyable and I look forward to chatting again. …”
Ridealong summary
Boston is poised to become the official hub for ethical AI, driven by passionate advocates in the field. In a heartfelt closing, the hosts invite their audience to join the 'Babe Tribe' as they celebrate Women's History Month and recognize the contributions of women making history in tech. This conversation is just the beginning, with a promise of future discussions to continue shaping the landscape of AI.
Building AI Boston·The Ethical AI Puzzle with guest Cansu Canca on Building AI Boston·Mar 11, 2026
“That's the decentralized blockchain powered cloud storage platform. It's sort of like a decentralized AWS or Google Cloud. So they're going to they're going to come on now and show us around. You know, Tau is very similar in a distributed fashion. Mog and Dubs, welcome. It's an inspiring project because I remember having Bitcoin on this very project, this very podcast in I think 2010 or 2009, 2010, 2011, right in that time period. I believe it was 2011. Inspired by 2011. Yeah. Super inspired by …”
Ridealong summary
BitTensor is revolutionizing decentralized computing by transforming cryptographic processes into productive work for the AI revolution. Unlike traditional systems, it utilizes distributed compute power to provide better solutions than centralized providers like OpenAI. This innovation is drawing early believers and is seen as essential for the growing demand for computational infrastructure.
This Week in Startups·One Genius Rule That Made This Coffee Brand Famous | EP 2262·Mar 14, 2026
“Last thing I'll mention, I've got a PyKVM hooked up to that machine for actual remote admin if I really need it when I'm out of the house and need to get into the bios of that machine or something. I also have a serial console running out of that machine. I love a serial console. I had to go right in there and look up the, what is it? The UART, I guess. Does it have a serial port on the motherboard or just a USB dongle? It's got a header. It's got the, my motherboard has a com port header that …”
Ridealong summary
With a PyKVM setup, you can remotely access and manage your server's BIOS and terminal, even when you're away from home. This innovative system allows for persistent log storage of your server's console output, giving you weeks or months of data at your fingertips. It's a game-changer for anyone needing reliable remote management solutions.
Brad & Will Made a Tech Pod.·325: renderDEEZ128·Feb 08, 2026
“but I have them all in Slack because I thought Slack would be better for collaboration because it's like a human collaboration tool. But to be perfectly frank, I believe now, after spending more than a week with five agents, that there are no one communication channel that is native to OpenClaw, meaning what you're talking about, Telegram, Slack, iMessage, Signal, is actually very good for agent-to-agent collaboration because all of these tools have been made for humans to use and agents are …”
Ridealong summary
To ensure security and proper organization, I use multiple Mac Minis for different AI agents, partitioning their tasks completely. For example, one agent handles sensitive financial information while another manages scheduling, preventing any accidental data leaks. This setup, though complex, is my way of addressing the security concerns surrounding AI collaboration tools like OpenClaw.
How I AI·5 OpenClaw agents run my home, finances, and code | Jesse Genet·Feb 25, 2026
“kind of gets to a better than junior person in that specific job that just works for you 24 7 and that's what we've been doing. So we have this DevOps agent that lives in a signal chat with us, and I can ask the bot in real human language, hey, how's everything looking? Are there any problems with our cluster? And then it'll say, yeah, there are like two people whose agents have crashed, and they haven't come back for like five hours. Maybe you should look at them, and just gives me further …”
Ridealong summary
OpenClaw represents a transformative leap in AI technology, akin to the impact of ChatGPT, and is expected to become ubiquitous within a year due to its immense power and potential.
OpenClaw represents a transformative leap in AI technology, akin to the impact of ChatGPT, and will become ubiquitous within a year due to its immense power and potential.
AI agents are revolutionizing industries by providing 24/7 monitoring and analysis at a fraction of the cost of human labor.
TFTC: A Bitcoin Podcast·#721: The Human Skills That Matter Now That AI Can Code with Calle·Feb 28, 2026
“Where do you think are other opportunities for companies to be started here in 2026? Where do you think people should be looking for entrepreneurial opportunities or just technical challenges that haven't been properly addressed? More labs, I think, is still interesting. Can we do more on model architecture is always interesting. You think we have not fully explored model architecture space? Yeah, I mean, the Frontier Labs have done a pretty good job of exploring it. But I think, I mean, as the …”
Ridealong summary
In 2026, the future of AI could hinge on breaking traditional model architecture constraints. By separating training and serving models, we can optimize performance and resource allocation. This innovative approach may unlock new opportunities in AI development and entrepreneurship.
Cheeky Pint·Reiner Pope of MatX on accelerating AI with transformer-optimized chips·Feb 26, 2026
“This is like a few Sentences back but actually Anthropic has blocked US companies first before the Chinese companies has blocked both OpenAI and XAI from from using the models and I think maybe Explicitly accused XAI of distilling stuff. I don't know but definitely not like in a full blog post like this So this one is like definitely the most high profile case And yeah, and like I do think like it is actually pretty hard to distinguish from like hey I'm just running my internal benchmark man. …”
Ridealong summary
Companies like Anthropic are blocking U.S. firms from using their AI models, raising concerns about potential distillation practices. Distillation, which involves extracting knowledge from one model to enhance another, can appear similar to legitimate benchmarking, but the volume and distribution of generated responses can reveal suspicious activity. This raises ethical questions about data privacy and transparency in AI development.
Latent Space: The AI Engineer Podcast·[LIVE] Anthropic Distillation & How Models Cheat (SWE-Bench Dead) | Nathan Lambert & Sebastian Raschka·Feb 26, 2026
“on a basic level, that's not a problem but I think the challenge compared to previous, let's say problems in the field is that you need a whole team to do it. Like you need an expert in GPU infrastructure. You need the person, the researcher who implements the core architecture. You need to run experiments. It's like, it's not usually something someone can do by themselves or on a weekend. I mean, understanding, yes but doing this whole thing is a lot of work. Looking for example at, I mean, …”
Ridealong summary
Training large language models (LLMs) requires a dedicated team and extensive resources, not just individual effort. For instance, the LLAMA model's training involved thousands of GPUs, frequent hardware failures, and the need for robust systems to prevent losing progress. This complexity highlights why LLM development is a massive undertaking beyond a weekend project.
Digital Disruption with Geoff Nielson·LLMs in 2026: What’s Real, What’s Hype, and What’s Coming Next·Feb 23, 2026
“Apparently, it's even better at computer use, and it's even better at interface stuff. So let's try it. And this is another little fun thing. I think this will work. I use WhisperFlow to talk to it. So let's try this. I love WhisperFlow. It's great. All right. So can you please build a 3D chess game and then test that chess game to make sure it works correctly? All right. And if you don't have WhisperFlow, it looks like you can also type into that microphone there for people who don't have that …”
Ridealong summary
Using GPT-4, I challenged the AI to create a fully functional 3D chess game and automate testing for it. This showcases the power of AI in coding, including features like scaffolding a React app and implementing automated tests. The potential for AI to streamline development processes is truly revolutionary.
The Neuron: AI Explained·GPT 5.4 LIVE Test & Learn to Code in 2026: What's Essential vs. What AI Handles Now·Mar 06, 2026
“Kneecapping model distillation will be far harder than restricting the shipments of physical goods like GPUs. In many ways, it seems like fully restricting distillation through distributed access methods seems almost impossible, and restricting GPU sales would be far more impactful. Anthropic and the AI industry should choose their battles. When API endpoints are available for the best models, other entities will use them to train variants of said model. This is a natural evolution of AI …”
Ridealong summary
Restricting access to AI models is becoming a critical issue, as distillation methods make it nearly impossible to control. Companies like Anthropic face tough choices: either limit their models to first-party products or risk losing market share to cheaper alternatives. The future of AI model access hinges on how these companies navigate this competitive landscape.
Interconnects·How much does distillation really matter for Chinese LLMs?·Feb 24, 2026
“Trump, so that's why he's walking that back. Meanwhile, Microsoft says it will keep Anthropic's AI tools embedded in its client products after its lawyers concluded the DoD's designation is only for defense projects, quoting CNBC. Microsoft is the first major company to say it will keep working with Anthropic after the Pentagon's actions. Some defense technology companies have told employees to stop using Anthropic's cloud models and migrate to alternatives. Our lawyers have studied the …”
Ridealong summary
Microsoft defies Pentagon restrictions by continuing to integrate Anthropic's AI tools in its products, ensuring they remain available for non-defense clients. This decision comes as many defense tech firms halt use of Anthropic's models, highlighting Microsoft's commitment to innovation and collaboration in AI development. OpenAI's recent launch of GPT-5.4 adds another layer to the competitive landscape, making this a pivotal moment in AI technology.
Tech Brew Ride Home·Silicon Valley Circling The Wagons Around Anthropic?·Mar 06, 2026
“This is like 2010. I think after that, I started writing a book on deep learning. And I think at a certain point, got called out to Silicon Valley to hang out with a lot of the early researchers in this space. We were all hanging out in Greg's apartment for a little while, and that group became OpenAI. In that time, I think there was a belief that in general, unsupervised learning and reinforcement learning was going to give rise to general intelligence and general reasoning models, but no one …”
Ridealong summary
Transformers are transforming healthcare AI, but early models struggled with accuracy. In the early days, the founders of a healthcare startup faced immense challenges integrating these technologies, leading to a breakthrough in risk prediction. Now, as the models improve, the potential for AI in healthcare is becoming clearer, despite initial setbacks.
The a16z Show·Ambience CEO Nikhil Buduma on AI in Clinical Workflows·Mar 04, 2026
“Yeah. Before we get into the landscape of what everybody's building, competing protocols, not only with traditional payment networks like Visa, MasterCard, stable coins, but even within Bitcoin, that's something I've observed over the last month is all these development kits and different ways of equipping your agent with the ability to send and receive Bitcoin. before that, what led you to, what was the tipping point for you to believe this? Was it observing what's happening, or did you build …”
Ridealong summary
Agentic payments are evolving rapidly, with new protocols emerging to challenge traditional networks like Visa and MasterCard. Recent advancements in AI models have made it possible for agents to effectively compare products and make purchasing decisions, paving the way for a new era of digital transactions. As companies gear up to implement these changes, the landscape of payments is about to get a significant makeover.
TFTC: A Bitcoin Podcast·#723: The Battle for the Agentic Economy with Matt Corallo·Mar 08, 2026
“They also have to disclose whether the data is copyrighted or in the public domain, whether it was purchased or licensed. Back up. Yeah. What exactly qualifies as a high-level summary? because listing out trillions of individual data points, categorizing them by copyright status, and tracking down the original owners sounds like the exact opposite of a summary. Yeah. And just to ground that for you listening, a token isn't a whole word, right? It's more like a syllable or a fragment of a word. …”
Ridealong summary
AI developers are now required to disclose every detail of their training data, turning the process into a meticulous audit instead of a quick scrape of the internet. This shift not only complicates data collection but also risks revealing competitive advantages, allowing rivals to reverse engineer successful models. The implications could fundamentally reshape the landscape of AI development.
Elon Musk Podcast·Elon Musk fights California over First Amendment·Mar 09, 2026
“And then you have the so-called leaders of this industry. Sam Altman recently said in a public discussion, people talk about how much energy it takes to train an AI model, but it also takes a lot of energy to train a human. It takes like 20 years of life and all the food you eat during that time before you get smart. SpaceX engineer Moratz and Coylan responded to this, I build AI for a living. I believe in what we're building. But this kind of rhetoric makes my work harder and more dangerous. …”
Ridealong summary
Sam Altman's comparison of human development to AI training has sparked backlash from industry insiders. As anger grows over job losses and fears of AI, the real challenge is fostering optimism and collaboration between human and AI intelligence for a better future. This discussion highlights the need for responsible leadership in shaping public sentiment around AI technology.
The AI Daily Brief: Artificial Intelligence News and Analysis·The Rise of the Anti-AI Movement·Feb 24, 2026
“Are you doing anything in this loop to do deterministic things? For example, linting is very deterministic, as you will very well know. Have you thought of marrying some of these ideas? Or are you using, for example, are you using linters on the code base where you found no need for it? Yeah, absolutely, absolutely. Yeah, we have type checkers, we have linters, we run the build. Claude is actually so good at writing lint rules. So actually what I do now, I used to tally stuff up in the …”
Ridealong summary
Automating code quality checks with linting can drastically improve your development process. By leveraging tools like Claude, developers can create custom lint rules directly in pull requests, ensuring consistent code quality. This approach not only streamlines collaboration but also enhances determinism in coding practices.
The Pragmatic Engineer·Building Claude Code with Boris Cherny·Mar 04, 2026
“As soon as you can record the signals, we've very quickly figured out what we, We can talk about neural representations, what they are. Second site's instructive. So in the retina, there's three layers of cells that matter. There's 150 million rods and cones. This connects to 100 million bipolar cells, bipolar because they've got two ends. And that connects the rods and cones to 1.5 million optic nerve cells, called the retinal ganglion cells. Ganglion is like a fancy word for like reaches a …”
Ridealong summary
A groundbreaking discovery shows that stimulating specific bipolar cells in the retina allows images to be perceived in the mind's eye, even after significant neural compression. This revelation was part of a larger vision to create a retinal prosthesis, leveraging existing technology to restore sight. The founders of the company aimed to revolutionize healthcare with neural engineering, culminating in this pivotal moment around 2021.
Y Combinator Startup Podcast·How To Build The Future: Max Hodak·Mar 09, 2026
“Bala, I wanted to say a big thank you for joining us today. Maybe jumping right into it, you know, can you tell me a little bit about your outlook for technology and AI in the coming year and the years ahead? And maybe specifically how it relates to the work you're doing and, you know, engineering as a field. yeah thank you thanks for having me jeff i'm great to be here i would say the time we are in it's definitely the most exciting time from a technology evolution like we have seen a lot of …”
Ridealong summary
AI is transforming engineering, allowing tasks that once took hours to be completed in just 30 minutes. This rapid evolution is creating an exciting environment for engineers and leaders alike, who must adapt to the changing landscape and harness the potential of AI. The challenge lies in balancing enthusiasm with strategic creation to ensure impactful innovations.
Digital Disruption with Geoff Nielson·Will AI Replace Software Engineers? Here’s What Lyft’s Engineering Director Says·Mar 09, 2026
“And there's so many brilliant minds right now working in the AI space that if these practices are considered important, then, you know, we can figure out ways to try to scale this and change how these technologies are being developed and preserve more data rights in the process. Yeah, because I get that, that whole idea of, well, it's out there. My images are out there. Everything is out there anyway. But, you know, we can all make a bit of progress towards that. And if everybody does a little …”
Ridealong summary
Sony is leveraging AI technology, specifically Phoebe, to improve fairness assessments in their computer vision products. By testing Phoebe across various business units, they aim to identify and mitigate biases before product launch, promoting industry-wide standards for data rights and ethical AI development. This initiative highlights the importance of accountability in AI, even amid the challenges of resource constraints.
Me, Myself, and AI·An Industry Benchmark for Data Fairness: Sony’s Alice Xiang·Mar 10, 2026
“gave both their original. But since they didn't have digital signatures, they used a different algorithm. And they had the property that to tolerate one faulty process, you needed four processes. Whereas if you used digital signatures, you only needed three processes. So the original paper contained both algorithms. And so I was one of the authors. The other algorithm without digital signatures is more complicated. And the general one for n processes was really a work of genius. It was almost …”
Ridealong summary
Using digital signatures can reduce the number of processes needed to tolerate faults from four to three. This discovery, made by researchers including Marshall Pease, highlights the brilliance of Byzantine fault tolerance algorithms, showcasing how a simpler proof can be derived for complex systems. The implications are significant, especially for critical systems like multi-computer setups in aviation.
The Peterman Pod·Turing Award Winner On Thinking Clearly, Paxos vs Raft, Working With Dijkstra | Leslie Lamport·Feb 23, 2026
“That's really cool. That's awesome. So, you go ahead, Grant. I want you to ask this one instead. I know where you're at. Okay, so I was just playing around right before this call with the Canva app in ChatGPT, and I was low-key blown away by it and just how functional it is. So I want you to help us step people through what actually happens when someone adds the Canva app in ChatGPT and Canva helped me create a pitch deck or mine was, what did I say? I said, can you make me a presentation about …”
Ridealong summary
Canva's integration with ChatGPT can generate multiple editable presentations in seconds, like one about why dogs rule. This seamless functionality is powered by their MCP server, allowing for consistent experiences across various AI platforms. The shared technology simplifies development and enhances user experience across tools like Microsoft Copilot and Clawed.
The Neuron: AI Explained·24 Billion AI Uses Later: What Canva Learned About the Future of Design·Mar 10, 2026
“Now some news from Moonpay. So Moonpay launches agents giving AI systems wallets and on-chain cash flow. Folks, AI agents are going to be a big part of commerce and assisting people to do different things. And they're going to be able to transact. And guess what? They're not going to use cash. They're not going to use Venmo. They're going to use stable coins. Crypto payments and stable coin infrastructure company Moonpay has introduced a new software layer that gives artificial intelligence …”
Ridealong summary
Moonpay has launched a groundbreaking tool that allows AI systems to create wallets and manage digital assets independently, without human intervention. This innovation could transform commerce by enabling AI to transact using stable coins, paving the way for a future where autonomous AI and robotics are integrated into our daily lives. As technology converges, this represents a significant shift in the token economy and investment opportunities.
Thinking Crypto News & Interviews·BIG CRYPTO NEWS! META FACEBOOK STABLECOIN, TRUMP SBF PARDON, SBI RIPPLE ASIA XRP LEDGER!·Feb 25, 2026
“Anthropic has debuted code review for Claude Code, which uses agents to check pull requests for bugs and says a typical code review costs $15 to $25 in token usage, quoting ZDNet. A pull request is initiated when a programmer wants to check in some new or changed code to a code repository. Rather than just merging it into the main track, a PR tells repo supervisors that there's something new ready to be reviewed. Sometimes the code is very carefully checked over before being merged into the …”
Ridealong summary
Anthropic's new Claude Code Review system dramatically improves code quality by catching nearly three times more coding errors before they cause damage. This AI-driven tool provides substantive feedback on 54% of pull requests, up from just 16%, allowing engineers to produce code at an unprecedented rate while reducing the risk of catastrophic bugs. With the pressure on human reviewers intensifying, this innovation could change how software development is approached.
Tech Brew Ride Home·Meta Plumps For Bot Social Networks·Mar 10, 2026
“There are other metrics that are specific to diffusion models, like how many kind of like denoising steps do you need to do? So essentially how fast they are. That's another thing that we need to track and we need to optimize because there's always kind of like a straight off between quality and speed. And so everything becomes a little bit more complicated because there's an extra knob that you can kind of play with. And so there are a few other things that are maybe diffusion specific, but …”
Ridealong summary
Assessing the quality of machine learning models in production is critical, and it revolves around balancing speed, quality, and cost. While speed and cost are straightforward to measure, quality evaluation requires nuanced approaches and qualitative assessments. To effectively implement models in business workflows, it's essential to track relevant metrics while understanding that benchmarks need careful interpretation.
The Neuron: AI Explained·Diffusion for Text: Why Mercury Could Make LLMs 10x Faster·Feb 24, 2026
“For sure. Speaking of balancing, our next topic transitions into how AI is impacting cybersecurity. Right. Let's see how we're steering AI into fighting cybercriminals. Cyberattacks are ramping up particularly from places like China We can just think of cybersecurity as a tech issue can we Exactly It a multi challenge Think of it like a complex safety net woven together with technology, policy, and people. So, if I'm reading this right, we need not only robust tech, but also well-trained …”
Ridealong summary
Organizations must invest in both technology and training to combat increasingly sophisticated cyber threats. As AI vulnerabilities grow, empowering individuals is crucial to creating a culture of vigilance against cybercriminals. This proactive approach is essential for building a safer digital landscape.
Embracing Digital This Week·March 2, 2026 | Embracing Digital Transformation: AI, Cybersecurity & Future Tr...·Mar 02, 2026
“These initial technical decisions that then dictate the lives of hundreds of professional engineers in the future, and that Scala choice was one of them, and they live with the faults of that now. What were the consequential, it could be good or bad, initial conditions of the Stripe Big Bang that you guys still live with right now? I think that metaphor is ... Maybe it's a little bit of survivorship bias where the actual statement is the early decisions we made that we never changed are …”
Ridealong summary
Stripe's choice of Ruby and MongoDB laid the groundwork for their industry-leading API availability of 99.99986%. These foundational decisions shaped their infrastructure and reliability, demonstrating the long-term impact of early technical choices. As Stripe evolved, they also faced debates over potential migrations, showcasing the challenges of adapting technology while maintaining stability.
The a16z Show·Patrick Collison on Stripe’s Early Choices, Smalltalk, and What Comes After Coding·Feb 20, 2026
“Brandon, as you look ahead, how do you see Zero Trust evolving over the next few years, especially with the growing importance of things like identity, data security, and the introduction of automation essentially everywhere in the security stack? yeah i think one of the biggest things and this is already happening today so we're moving from that perimeter approach like where you just rely on whether you're outside the network or you're in the network and you're moving more towards a …”
Ridealong summary
In the evolving landscape of cybersecurity, static policies are becoming obsolete in favor of dynamic, data-centric approaches. Organizations are now leveraging AI and machine learning to adapt security measures based on real-time data and user roles, ensuring better protection against emerging threats. This shift is crucial as businesses operate across diverse environments, from SaaS applications to on-premise systems.
Threat Vector by Palo Alto Networks·Zero Trust Without the Hype·Mar 05, 2026
“Ben, a couple of questions from the audience about what you're building. Seems that the consumers watching the show are very curious about when all the stuff we're talking about is going to impact their daily life. So first question, a little bit generic, but roll with me here. When will we be able to order coffees from AVs, just general commerce on the road? Super excited about this and trying to help build it. So talk to the people. The first thing I'll say is autonomous commerce and Autolane …”
Ridealong summary
Autonomous vehicles are set to revolutionize how we order coffee and other goods, with delivery systems already being tested in major cities like Miami and Los Angeles. Companies like DoorDash and Waymo are piloting these innovative solutions, indicating that this technology will impact our daily lives sooner than we think. As multimodal delivery methods, including drones and sidewalk bots, emerge, the landscape of commerce is about to change dramatically.
This Week in Startups·The Global Expansion of Self-Driving Vehicles·Mar 11, 2026
“Of course, you had other advice for for builders, and we kept interrupting you because we have so many questions. I would say so, OK, so maybe two pieces of advice that are kind of weird because it's like about building for the model. So one is don't build for the model of today. Build for the model of six months from now. This is like sort of weird, right? Because like you can't find PMF if the product doesn't work. But actually, this is the thing that you should do, because otherwise what …”
Ridealong summary
To stay competitive, always build for the AI model of six months from now, not the current one. This approach prevents being leapfrogged by competitors and ensures your product remains relevant as models rapidly evolve. At Anthropic, they constantly rewrite their code, adapting to advancements in AI capabilities rather than getting bogged down by temporary solutions.
Y Combinator Startup Podcast·Inside Claude Code With Its Creator Boris Cherny·Feb 17, 2026
“like there aren't like 30 to 50 crypto related benchmarks or training environments that people are shipping. In some sense, it's actually agency inducing for us because they'll just defer to the crypto industry to just figure out what's valuable for them. But I think it's fundamentally a social issue. And it's sort of tied to all of these dynamics around like, you know, you see someone who gets extremely wealthy, who you don't think should get extremely wealthy. Like maybe it was like there's a …”
Ridealong summary
AI models are mastering crypto coding faster than expected, despite initial challenges with languages like Solidity. This is due to the verifiable nature of blockchain technology, allowing models to learn the 'language of crypto' with less direct training. This dynamic reflects broader social perceptions of the crypto industry, where skepticism can lead to unexpected innovation.
“You know, what is the challenge? What is the issue? What is the core mission? You know, I'm trying to continue making it happen with the technology and with the technical solution. I guess I'm lucky that I'm in the valley, in Silicon Valley. And, you know, I find companies like Hayden, you know, a startup. And I think I have, as I said, I'm lucky that I have this wonderful board and a great executive team that they believe in technology and they believe in trials and POCs and pilots to really, …”
Ridealong summary
AI is transforming public transit by addressing major challenges and enhancing operational efficiency. In Silicon Valley, innovative startups like Hayden are collaborating with public agencies to implement technology that can significantly improve transit systems. This approach encourages bold experimentation and sharing of successful methodologies across the industry, ultimately benefiting millions of commuters.
NVIDIA AI Podcast·Safer, Faster Public Transportation: AC Transit’s AI-Powered Upgrade with Hayden AI - Ep 290·Feb 18, 2026
“with a couple thoughts on what they shouldn't do? And from your perspective in this space, if you were talking to a client or customer, what would you say don't do in relation to AI agents and those types of things? I think I would say, my take would be don't do two things, right? I think one is don't try to take an approach of that you block everything because I think we are way beyond that, right? I think the cost of missing out is very high. And if we couldn't stop people going to the …”
Ridealong summary
Blocking everything in AI security is a costly mistake. Instead, prioritize visibility and apply basic controls before advancing to complex solutions. This approach acknowledges the fundamental shifts in technology and prevents outdated methods from hindering progress.
DrZeroTrust·Beyond Perimeter Defenses: DLP, CASB, and the AI Agent Revolution·Mar 03, 2026
“And the fifth agent took all of these research packages and built a narrative structure. The final agent wrote a full script, 4,600 words, technically 28 minutes, technically in my voice. The total token cost for that, well, I'm a bit generous about this. I give the agents really large token budgets to work with. They never come close. So I told it this was consequential. I was going to be reading a lot of this out to this incredibly important audience, this group of people who matter so much …”
Ridealong summary
In an experiment, I orchestrated AI agents to write a 4,600-word script in just 40 minutes, drastically reducing the time I usually spend on script preparation. By setting objectives and constraints, I delegated tasks to these agents, which resulted in a final product costing only a fraction of what I expected. This innovative approach showcases the potential of AI in content creation, allowing me to focus on more enjoyable activities while still achieving high-quality results.
Azeem Azhar's Exponential View·Showing you my AI chief of staff (OpenClaw practical guide)·Mar 05, 2026
“Does that concern you at all? Or is this just, you know, it's just money from a different place? I think it's just money from a different place, honestly. And it really does make sense, right? When you think about the way that weaponry or even view it more defensive than weapons, whether that be things like interceptors, et cetera, these are all technology plays, right? This isn't, you know, training a bunch of guys to be able to go storm a beach, right? That's a physical thing, you know, put …”
Ridealong summary
The defense sector is poised for significant growth as Silicon Valley increasingly funds military technology, including autonomous weaponry. With ongoing conflicts, like the strikes in Iran, there's a pressing need to replenish stockpiles, suggesting a lucrative future for companies like Anduril. This shift illustrates how tech innovation is reshaping modern warfare and defense strategies.
Prof G Markets·$500M Bet On The Iran Strike — Before It Happened·Mar 05, 2026
“for creative individuals. I mean, Xbox is a hardware company. It's also a live streaming company. It's also a studio where you just have studio heads that go and greenlight projects. It's not all directly related. Handle says, Larry Ellison using Oracle on a nice Sunday morning. Let's go. Good job. That's correct. I don't know. He probably does store a lot of data in Oracle. Who knows? Yeah, Aaron says the CEO necessarily is not in the product daily. Who knows? We'll wait to see. It already …”
Ridealong summary
Recent advancements in AI technology are transforming gaming hardware, with systems like Cerebrus achieving unprecedented speeds of 16,000 tokens per second. This leap in performance is set to enhance user experiences in gaming, making interactions faster and more seamless. As AI continues to evolve, it promises to redefine how we engage with games, pushing the boundaries of what's possible.
TBPN·CitriniPocalypse, Dot Com Lore, Gene-Edited Polo Horses | Alap Shah, Will Brown, Michelle Lee, Mike Annunziata·Feb 23, 2026
“you use it to come up with what to build? Honestly, the simplest thing is like open quad code or co-work and point it at a Slack thread. You know, like for us, we have this channel that's all the internal feedback about quad code. Since we first released it, even in like 2024 internally, it's just been this firehose of feedback. And it's the best. And like in the early days, what I would do is anytime that someone sends feedback, I would just go in and I would fix every single thing as fast as …”
Ridealong summary
With the introduction of Quad Code, productivity per engineer has skyrocketed by 200%, a staggering achievement in the tech world. This leap is attributed to rapid feedback cycles and AI's ability to streamline coding processes, transforming how teams approach product development. The results are so remarkable that they dwarf previous gains seen in large organizations like Meta.
Lenny's Podcast: Product | Career | Growth·Head of Claude Code: What happens after coding is solved | Boris Cherny·Feb 19, 2026
“What's going to happen to central buyers and platform teams and IT teams if agents are making the decision? It's very clear that coding is pretty much dead, but engineering is very much not. Every time you have a technical epoch, you have to redo everything, and we forget that every time. I don't think people even have a common definition of a bubble. If AI demand is real and accelerating, why does everything still feel constrained? Why does a technology that's clearly delivering value also …”
Ridealong summary
Despite AI's rapid growth, infrastructure bottlenecks are hindering its scalability, raising questions about decision-making in tech. As agents begin to write code and provision systems, the visibility of decision-making layers fades, complicating how businesses adapt. This segment reveals why understanding infrastructure's role is crucial in the AI revolution.
The a16z Show·Martin Casado on the Demand Forces Behind AI·Jan 21, 2026
“And I'm just going to put a note here. Use my OpenAI key from my shared Chrome extension config. So I've had some... I don't want to keep pasting my OpenAI keys and stuff like that. So I end up pulling out some shared config to share across all my Chrome extensions. That way, I don't need to rinse and repeat that step over and over again. So I save this file now. And now whenever I close this, it's going to replace my prompt in Cloud Code with that completed prompt. Oh, interesting. Yeah, so …”
Ridealong summary
Using a symbolic link for your API keys can save you time and headaches. Instead of updating each Chrome extension individually when your key changes, a single update in one config file propagates everywhere. This approach streamlines your development process and enhances security by keeping sensitive information centralized.
How I AI·How this visually impaired engineer uses Claude Code to make his life more accessible | Joe McCormick·Feb 16, 2026
“to deploy AI agents across their enterprise network. Which is honestly a massive signal hiding in plain sight. You see these corporate handshakes all the time, and you might just brush it off as another chatbot integration. But that completely misses the point of what is actually happening. We were looking at a fundamental rewiring of a $283 billion IT industry. So what happens when the business of outsourcing human labor gets replaced by orchestrating software? That is the exact tension we are …”
Ridealong summary
The shift from human labor to AI agents in India's outsourcing sector represents both a futuristic opportunity and an existential threat to traditional jobs.
The shift from human outsourcing to AI agent orchestration is an existential threat to India's traditional outsourcing model.
Elon Musk Podcast·AI UPDATE: Infosys Replaces Human Labor With Anthropic·Feb 21, 2026
“the world's largest cloud security platform. That should get your attention. We talk a lot about AI in all of our shows, of course, the potential rewards of AI in your business. I think it's too great to ignore. No business can afford not to at least explore AI, but the risks are there, too. I mean, there's the issue of loss of sensitive data, even attacks against enterprise-managed AI. Of course, the bad guys love it. Generative AI increases opportunities for these threat actors. It helps them …”
Ridealong summary
Last year, 1.3 million social security numbers were leaked to AI applications, highlighting the urgent risks of AI in cybersecurity. Chad Pallett, acting CISO at BioIVT, turned to Zscaler, cutting cyber premiums by 50% and doubling coverage, showcasing a successful strategy to mitigate these dangers. With Zscaler's Zero Trust Plus AI, businesses can safely adopt generative AI while enhancing productivity and compliance.
“and you'll see how little data we're actually sending over. Even though he's updating 2,500 divs per frame, like if you go to wherever it's, yeah, whichever one's the one that's update, there you go, yeah. So if you look here, and look at how much is being sent versus how many is actually, like try to, like this is just a different paradigm for how you, and the thing is, again, not everybody has to care about these low-level things, but the thing is, is that once you do this, the idea of CRUD …”
Ridealong summary
Imagine updating your web app in real-time without the hassle of traditional CRUD operations. By leveraging an open data stream and database query subscriptions, developers can simplify updates and enhance user experience. This approach not only streamlines data handling but also keeps users informed without misleading them during interactions.
Talk Python To Me·#537: Datastar: Modern web dev, simplified·Feb 21, 2026
“Yeah, I think physics is the thread. So having spent a lot of time in theoretical physics, I think there is first very fundamental and exciting questions, like things that haven't actually been figured out in quantum gravity. So there's really the frontier. There's also a lot of mathematical tools that you can use, right? For instance, in particle physics, but also in general relativity, symmetry plays an enormously important role. And this goes all the way to gauge symmetries as well. And so …”
Ridealong summary
The relationship between diffusion models and stochastic thermodynamics reveals groundbreaking insights in AI. By applying concepts from theoretical physics, researchers are uncovering deep connections that could revolutionize generative AI. This journey led to the publication of a new book, bridging these complex fields and showcasing the power of interdisciplinary thinking.
Latent Space: The AI Engineer Podcast·🔬Nature as a Computer: Prof. Max Welling, CuspAI on AI x Materials Science·Feb 25, 2026
“double back and realize that it was wrong and correct course. Another thing you can do, and this is where it gets, I think, really interesting, although notably those interventions drive a lot of the reduction. But another thing you could do is use the presence of activations which were classified as hallucinations as a signal for reinforcement learning to try to get the model to not go into this state in the first place by just basically punishing it for getting into the hallucination state at …”
Ridealong summary
AI models can be trained to avoid generating false information, known as hallucinations, by implementing reinforcement learning strategies. By using a 'probe' to identify hallucinations, developers can correct the model’s behavior, ensuring it prioritizes factual accuracy. This innovative approach reshapes the AI's capabilities, enhancing its reliability in generating content.
"The Cognitive Revolution" | AI Builders, Researchers, and Live Player Analysis·Don't Fight Backprop: Goodfire's Vision for Intentional Design, w/ Dan Balsam & Tom McGrath·Mar 05, 2026
“And then what is the end state? And therefore, you can make a lot of inferences about what the actual total shape is. Yeah, that's right. It's almost like, you know, you have this big, like, three dimensional valley, you know, where you're sort of trying to find like these like low energy states. There's so much to search through, that's almost overwhelming. But these hints, they sort of maybe put you in an area of the space that's already like, kind of close to the solution, maybe not quite …”
Ridealong summary
AlphaFold3 revolutionizes protein structure prediction by utilizing evolutionary data to create a pairwise distance matrix. This innovative approach allows the model to navigate complex protein folding landscapes effectively. By leveraging insights from evolutionary information, AlphaFold3 optimizes the search for low-energy states in protein structures.
Latent Space: The AI Engineer Podcast·🔬Beyond AlphaFold: How Boltz is Open-Sourcing the Future of Drug Discovery·Feb 12, 2026
“then this whole conversation that I'm having with people right now in this quarter is about open source. I wonder if you can talk about the company's strategy, why the company decided to go and publish open weights of the models. What's the benefits? What's the cons to that? So for our team, like for the researchers team, we always wanted to go open source because, you know, open source communities is fantastic. I learned that from day one when I joined the team is open source community is …”
Ridealong summary
Open source models can enhance research but may hurt business by reducing API usage. While researchers thrive in collaborative open source communities, they face challenges in engineering discipline compared to closed APIs. This dynamic creates a tension between innovation and commercial viability in tech companies.
"The Cognitive Revolution" | AI Builders, Researchers, and Live Player Analysis·Intelligence with Everyone: RL @ MiniMax, with Olive Song, from AIE NYC & Inference by Turing Post·Feb 22, 2026
“there's already been so much of this. We're going to talk a little bit more about Quen, Coder, and GLM maybe on the open source and project section. But we got to close out this and move on to business. So real quick, also worth mentioning, Cursor did announce Composer 1.5. So Cursor for context is the coding ID. It used to be sort of like the exciting thing in 2024 with Intelligent Autocomplete. Now they are trying to be a Cloud Code, a Codex competitor, and they claim that Composer 1.5 is a …”
Ridealong summary
Cursor's Composer 1.5 is positioned as a more capable coding agent, aiming to rival Cloud Code and Codex. Enhanced through reinforcement learning, it could significantly impact organizations that utilize it, despite mixed excitement compared to other tools. Meanwhile, XAI's Grok Imagine API enters the competitive space of text and image to video conversion, but faces challenges against established players like Google and OpenAI.
Last Week in AI·#235 - Opus 4.6, GPT-5.3-codex, Seedance 2.0, GLM-5·Feb 16, 2026
“little bit about kind of where you're at and what you're doing. You're very familiar with the show. You've been coming on the Twit Network for a long time, I think are very well known to the audience. But just in case there's a few people who don't know, why don't you talk a little bit about what you do at Blackbird AI? And also, I'll just say for those who don't know, Dan is a longtime journalist. He worked for me at publications that I've worked for multiple times. We go back a long way. And …”
Ridealong summary
Blackbird AI is revolutionizing the battle against misinformation and disinformation in today's digital landscape. With a strong background in journalism, Dan, an investigative reporter, is at the forefront of this crucial fight, leveraging AI and cybersecurity to uncover the truth. This discussion highlights the evolution of technology and the vital role Blackbird AI plays in protecting information integrity.
Intelligent Machines (Audio)·IM 860: You Gotta Get Computer - Claude Surges to No. 1·Mar 04, 2026
“I mean, you use that all day, it's like you just build it in there. 99% of the code that I write is using the Codex app. Same, I mean, I live in there now. Yeah. Okay, well, that's, that's actually really interesting. I definitely want to talk about the app in particular, but I want to go back to the thing you just said, which is maybe if I, if I'm reading you right, you're, you're kind of like, we're pushing the frontier. We're seeing lots of people who are maybe broader than just like senior …”
Ridealong summary
The Codex app is transforming how developers write code, with 99% of users relying on it daily. While it’s designed for technical users, its powerful features are attracting a broader audience, blurring the lines between coding in different platforms. This tailored experience is reshaping the development lifecycle and has seen significant success among professional teams.
AI & I·OpenAI's Codex: This Model Is So Fast It Changes How You Code·Feb 18, 2026
“COVID forced us to adopt technology in a very rapid way that before, because of cultural change that we needed to overcome, because of regulatory issues and the fear of that, all those things disappeared. And you've turned this into something awesome, which is, I can now move into people's homes. You called it the first mile. I love that. What a transition. That must, it must have not just happened though, right? I mean, there was a big catalyst, but what took, what made that happen? Yeah. So, …”
Ridealong summary
Desperation is the mother of adoption, especially in healthcare. COVID-19 forced rapid technological changes, pushing telemedicine and AI into the spotlight as essential tools for patient care. With an overwhelming explosion of medical data, AI has become vital for healthcare professionals to provide effective treatment.
Embracing Digital Transformation·#319 From Telemedicine to AI: A New Era in Medicine·Jan 22, 2026
“token being A exactly in the same way that you would before. And so there's a couple of things that you can do to handle this. One is you can go in the direction of richer ways of measuring whether a model is correct or is doing the thing that you want than measuring log probability of a certain letter. The second thing you can do is repeatedly sample from a model and then see whether performance stays the same or degrades as you repeatedly sample from the model. So actually our work on …”
Ridealong summary
HealthBench measures model performance across 49,000 different axes, revealing how well AI expresses uncertainty and escalates issues to physicians. By repeatedly sampling models, researchers can now assess the worst performance outcomes, leading to more reliable healthcare AI. This innovative approach is reshaping our understanding of AI effectiveness in medical conversations.
"The Cognitive Revolution" | AI Builders, Researchers, and Live Player Analysis·Universal Medical Intelligence: OpenAI's Plan to Elevate Human Health, with Karan Singhal·Feb 25, 2026
“infrastructure and apps, right? So, like, what is a model company? Like, it's clearly infrastructure, right? Because it's like, you know, it's doing kind of core R&D, it's a horizontal platform, but it's also an app because it touches the users directly. And then, of course, you know, the growth of these is just so high. And so, I actually think you're just starting to see a new financing strategy emerge. And, you know, we've had to adapt as a result of that. And so, there's been a lot of …”
Ridealong summary
Emerging financing strategies in AI are blurring the lines between infrastructure and applications, allowing companies to rapidly gain market share. By securing funding for compute resources and integrating breakthroughs into user-facing applications, businesses are reshaping their growth models. This shift is reminiscent of past tech evolutions, but the speed of change is unprecedented.
Latent Space: The AI Engineer Podcast·Inside AI’s $10B+ Capital Flywheel — Martin Casado & Sarah Wang of a16z·Feb 19, 2026
“Yeah, we have one portfolio company called Exo, which is doing distributed inference instead of decentralized inference. So we've seen, obviously, today the most popular models are still these big lab models that run in a data center somewhere and you're paying them to access and they can sort of gate who can use it and how can you use it. But we're seeing more and more of these open source models get nearly as good as these big lab models, and you can run them very, very inexpensively on any …”
Ridealong summary
Exo's new technology allows AI models to run locally on devices, enhancing performance and reducing reliance on centralized data centers. This shift towards local computing aligns with the cryptocurrency vision of self-sovereignty, offering users greater control and efficiency amidst a GPU shortage. Dragonfly Capital's investment in this innovation reflects a broader trend in the evolving tech landscape.
The Milk Road Show·AI Agents Will Need Crypto to Work (Here’s Why) w/ Tom Schmidt·Mar 04, 2026
“Actually, let's come back after we go through this flow, how you would approach this if you were at say Meta or another, maybe a smaller company, how this workflow might work at a larger company that isn't just your own startup. Yeah, interesting. Let's come back to that. Cool, all right, so this is Claude coming back. I have a comprehensive understanding of the code base. I thoroughly analyzed StudyMate live code base and understand the current system, feature requests, and key areas that it's …”
Ridealong summary
Claude reveals a sophisticated method for understanding and planning app development by analyzing the code base and asking crucial questions. This approach contrasts with casual coding, emphasizing the importance of thorough preparation and learning opportunities. By leveraging structured inquiries, Claude sets the stage for effective project execution.
Lenny's Podcast: Product | Career | Growth·The non-technical PM’s guide to building with Cursor | Zevi Arnovitz (Meta)·Jan 18, 2026
“come back and look at the results. It's a rare thing that I do. I mean, once a month I finish a chapter or something, or if I write something important, where I want the maximum, let's say, quality check on that. But like you said, for most tasks, it's sufficient to use the light effort, yeah. Or the automatic one where it decides by itself, essentially. Right, right. And I mentioned MoldBot and the release of that tool. Have you spent much time digging into that? Well, yeah, MoldBot, I think …”
Ridealong summary
MoldBot, now known as OpenClaw, is sparking excitement in AI productivity tools, similar to the buzz around AlphaGo. This local agent allows users to run AI directly on their computers, helping with tasks like organizing calendars and emails. Despite some hesitance about trusting AI with personal data, its potential to demonstrate the capabilities of language models is undeniable.
The TWIML AI Podcast (formerly This Week in Machine Learning & Artificial Intelligence)·AI Trends 2026: OpenClaw Agents, Reasoning LLMs, and More with Sebastian Raschka - #762·Feb 26, 2026
“and evaluating outputs. You have felt this shift yourself using cloud code. This continues and accelerates knowledge work follows. Cowork chat GPT agent mode, and their successors will do for reports analysis and documents what cloud code did for code. The pattern is the same. Describe an outcome, let the agent execute, review and iterate. Gartner projects 38% of organizations will have AI agents as formal team members by 2028. Multi-agent coordination becomes the default architecture. Instead …”
Ridealong summary
AI agents will revolutionize knowledge work by 2028, with workers spending more time directing agents than performing tasks themselves.
AI agents will revolutionize knowledge work by 2028, with workers spending more time directing agents than performing tasks themselves.
Machine Learning Guide·MLA 028 AI Agents·Feb 22, 2026
“I can't show you that in the car. No that's I don't I'm not looking for a web page screen Yeah, I want you to tell me the answer yes. Yeah, that's one of them right right, but the other is you know when the cars Rolling or the list is limited to 20 the items versus when you stop you get you know 150 It's like well. I want to listen to rush and rush isn't alphabetically in the first 20 things on my phone, so if I want to listen to I'm gonna have to take my phone out of my pocket like this is …”
Ridealong summary
AI models struggle with short-term memory, leading to frustrating user experiences. A user shares their journey transitioning from GPT to Claude, highlighting the challenges faced when context is lost mid-conversation. This showcases the importance of maintaining context in AI interactions for effective troubleshooting.