Material Abundance: Radical AI’s Closed-Loop Lab Automates Scientific Discovery

Material Abundance: Radical AI’s Closed-Loop Lab Automates Scientific Discovery

Joseph Krause and Jorge Colindres, co-founders of Radical AI, unveil their "materials flywheel" – an integrated system combining frontier AI with autonomous labs to revolutionize materials discovery.


Watch Episode Here


Read Episode Description

Joseph Krause and Jorge Colindres, co-founders of Radical AI, unveil their "materials flywheel" – an integrated system combining frontier AI with autonomous labs to revolutionize materials discovery. They detail how this closed-loop system achieves unprecedented experimental throughput, addressing the costly and slow development cycle that plagues critical industries. Learn how their property-driven AI engine, multimodal data integration, and robotic labs are vertically integrated to create foundational materials for everything from semiconductors to hypersonic flight.

Sponsors:
Oracle Cloud Infrastructure: Oracle Cloud Infrastructure (OCI) is the next-generation cloud that delivers better performance, faster speeds, and significantly lower costs, including up to 50% less for compute, 70% for storage, and 80% for networking. Run any workload, from infrastructure to AI, in a high-availability environment and try OCI for free with zero commitment at https://oracle.com/cognitive

Shopify: Shopify powers millions of businesses worldwide, handling 10% of U.S. e-commerce. With hundreds of templates, AI tools for product descriptions, and seamless marketing campaign creation, it's like having a design studio and marketing team in one. Start your $1/month trial today at https://shopify.com/cognitive


PRODUCED BY:
https://aipodcast.ing

CHAPTERS:
(00:00) About the Episode
(03:45) Introduction and Company Overview
(05:01) Materials Science Problems
(10:35) Scale-up and Processing Challenges
(16:38) Customer-Driven Material Discovery (Part 1)
(18:26) Sponsor: Oracle Cloud Infrastructure
(19:35) Customer-Driven Material Discovery (Part 2)
(23:17) Company Mission and Vision (Part 1)
(31:54) Sponsor: Shopify
(33:50) Company Mission and Vision (Part 2)
(34:44) The AI-Lab Flywheel
(40:27) AI Models and Architecture
(49:49) Scientific Intuition and Experience
(58:11) Active Learning and Breakthroughs
(01:04:45) Data Challenges and Sources
(01:14:38) Search Space and Automation
(01:25:14) Inference Scaling and Properties
(01:31:01) Active Learning Implementation
(01:37:01) Move 37s in Materials
(01:44:31) IP Strategy and Business
(01:48:24) Air Force Partnership
(01:51:23) Culture and Closing
(01:53:22) Outro

SOCIAL LINKS:
Website: https://www.cognitiverevolutio...
Twitter (Podcast): https://x.com/cogrev_podcast
Twitter (Nathan): https://x.com/labenz
LinkedIn: https://linkedin.com/in/nathan...
Youtube: https://youtube.com/@Cognitive...
Apple: https://podcasts.apple.com/de/...
Spotify: https://open.spotify.com/show/...


Full Transcript

Transcript

Nathan Labenz (0:00) Hello, and welcome back to the Cognitive Revolution. Today, my guests are Joseph Krausz and Jorge Calindres, cofounders of Radical AI, a company that's building a materials flywheel, an integrated system combining frontier AI models with almost fully autonomous laboratories to dramatically accelerate the discovery and production of new materials. As someone who spent a year as an undergraduate research assistant weighing out small amounts of fine powders and running very tedious experiments, I have long dreamed of the day when science would start to be automated. That day, I'm pleased to say, is now arriving. Radical AI's lab can run a 100 experiments per day, roughly doubling both what I did in a full year as an undergrad and what Joseph did in a full year at the Army Research Lab. And they're using that unprecedented throughput to tackle some of humanity's most important materials challenges. The upside here is truly enormous. As Joseph and Jorge explained, advanced materials are the foundation of nearly every major modern industry, from semiconductors and aerospace to energy and defense. Yet, developing new materials remains painfully slow and expensive, typically requiring over a $100,000,000 and more than a decade to go from discovery to commercialization. This creates a so called valley of death between academic research, which focuses on fundamental understanding, and corporate r and d, which for the most part focuses on incremental improvements. In that gap lie potentially transformative materials that could enable futuristic technology, such as floating trains, interplanetary travel, and hypersonic flight, but which nobody has successfully brought to market. Joseph and Jorge walk us through their entire closed loop system, How they use property driven optimization to navigate the vast space of possible materials, how their AI engine incorporates multimodal data from papers to microscopic images, how their robotic lab conducts synthesis and characterization, why they believe that capturing experimental data at scale is the key unlock, and crucially, how they're vertically integrating all the way through to manufacturing. Along the way, we also discuss why material science has become more expensive over time, the l k 99 saga and what it reveals about the challenges of synthesis and reproducibility, how the Radical AI system learns from both successes and failures ultimately building toward a sort of scientific intuition, and their recent Air Force contract to develop high entropy alloys for hypersonic applications. What struck me most about this conversation is the scale of Radical AI's ambition. This is not a company trying to make paint a bit more durable. On the contrary, they are trying to create the materials that will define entirely new industries. 1 quick note before we get started, for those who want more on the machine learning of material science, my previous episodes with Tim Dignan and Jonathan Godwin of orbital materials could be great compliments to this conversation. Today, we focus primarily on the discovery engine that Radical AI is building at a system level. But those earlier episodes do go much deeper into the technical details of how AI models learn to simulate molecular dynamics and ultimately predict material properties. Of course, while the 2 companies could naively be viewed as competitors, I think this conversation makes it abundantly clear that there is functionally unlimited opportunity for better living through material science. And it's certainly my hope that both companies will go on to massive success. With that, I hope you enjoyed this conversation about accelerating materials discovery with a vertically integrated combination of AI and automation with Joseph Krausz and Jorge Calindres, cofounders of Radical AI. Joseph Krausz and Jorge Calindres, cofounders of Radical AI. Welcome to the Cognitive Revolution.

Joseph Krause (3:51) Thank you.

Jorge Colindres (3:51) For having us. Excited to be here.

Nathan Labenz (3:53) Yeah. Likewise. So you guys are doing some really interesting frontier work in the application of AI to material science, and, there's a lot to unpack. We were just joking beforehand that automating science is something that as an undergrad research assistant, I dreamed of for many hours as I was sitting there doing the tedious work of weighing out small amounts of fine powder, and that is I was told at the time that that was very difficult to automate and probably wouldn't happen in my lifetime. So the fact that it is starting to happen is, you know, obviously, part of a bigger vision gets me excited right off the bat. For starters, you know, our audience is very interested in AI by definition. Probably mostly doesn't know a ton about material science broadly. So maybe you could kind of just set us, like, a baseline. Why is material science hard? You know, what's going on? 1 of the observations, you know, from kind of doing my homework on the company is that the cost of material science breakthroughs seems to be going up. So what's the kind of lay of the land today that you guys are entering into with a a new paradigm?

Jorge Colindres (5:02) Yeah. Well, luckily for you, Nathan, your dreams are about to become, if not already a reality in today's world. And so it's gonna be an incredibly exciting, really, next 5 10 years. And we think and fundamentally believe as a company, science will never look the same as it's going to look over the next decade. And so we're very excited to be help leading that charge and building that. You know, if you think about why materials are so exciting, and I'm sure we'll get into this in the founding story of the company, but they really impact so many different industries in the world. Automotive and aerospace, manufacturing and defense, climate, energy, semiconductors, electronics, the most important industries in the world are all a direct result from materials r and d. But we always see really 3 problems. First, cost, as you called out, north of 100,000,000 to develop a new system from that computational framework through to really scaling that process. Second is time. It takes a long time to do a novel discovery. On average, we see typically 10 plus years to really push a novel discovery into commercialization. And sometimes it can be much longer spanning 2025. And then 3, fragmentation. Probably 1 of the biggest problems that contributes both to the cost and time that I just alluded to. In material science, we always describe it as a company as 2 different paradigms. On 1 end, you kind of have this very fundamental approach to science, what we call the academic pursuit of science. It drives your understanding of science, very fundamental in nature, and usually not focused on commercial application. There can be some tie in, of course, but that's not the objective of academic research. The objective is to drive your understanding of of what you're studying. And then on the other paradigm is really commercialized r and d or corporate r and d. And they're entirely focused on commercialization, which is great, entirely focused on optimization as well. And so they are driving to 1 to 5% performance gains so that they can change their margin and report that to Wall Street. They are not focused on discovering something like an auto tap superconductor, which I'm sure we'll get into. And this is the problem. In the middle is this wide open white space to do this novel fundamental discovery, but entirely focused on commercial applications. And then, of course, that's exactly what we radically are trying to do. So that's where material sits today and how we view the landscape and the opportunity that's sitting before us.

Nathan Labenz (7:35) Was it always that way? I mean, the obvious analogy that comes to mind for this sort of discovery is drug discovery. And, you know, broadly there, the story, I guess, seems to be kind of 2 parts that people typically tell. 1 is like, for 1 thing, it used to be easier because there was lower hanging fruit and maybe that low hanging fruit has sort of been picked. And then the other side is like, well, clinical trials are really expensive and so, you know, even once you have something that's really great, you still got a long way to go before you can actually get it to market and, you know, rightly or wrongly, obviously, we want to be somewhat safe with that stuff at least, but safe to say that the costs have grown dramatically. Is it kind of similar in material science? Would you tell a low hanging fruit depletion story? And is there something analogous to the clinical trial process that makes things harder than it would naively seem like it should be?

Jorge Colindres (8:26) Yeah. Absolutely. There are very strong similarities and then, of course, very core differences as well. Know, 1 important thing about materials is that we don't go into the human body. Well, most don't go into the human body. And so there is a big delineation that there are some that actually do, and there are classes of materials that, of course, have to be bio friendly. But there's a lot of similarities on search space, the size of the search space. Now things like proteins or or other things in bio are bigger, but materials are are quite literally impossible to move through as well. So they're both large. And then there is this challenge on going the full scale, what we call vertically integrating across. So actually going from the novel discovery and pushing that all the way through to producing that. We have different steps in that process from bio or drug discovery versus materials, but they both equally take around the same amount of time. Whereas bio might be going through clinical trials, materials are going through scale up and processing. And so how do you actually take this small reaction that you run-in the lab and figure out how to make tons of that material so it can be used in commercial application or industry. Again, use cases, different processes to do that approach, but you still get this high cost, you still get this long timeline, and you still get this fragmentation, candidly, where different people are doing different parts of that process. Just like in drug discovery, you might have a company focused on discovery of novel drugs And then a big pharma really thinking about, I'll pay for the clinical trial, I'll go through that process. And so you put all this process together, whether you're going through clinical trials or you're worrying about optimizing the microstructure for a novel alloy, both of these are time consuming. They are incredibly expensive to operate and go through, and they lead to this really value that doesn't allow a lot of discoveries to proliferate into real commercial application. And so there's a strong corollary between drug discovery and materials. And the differences we think are actually less explored inside materials today. It's the exact opportunity for a company like Radical.

Nathan Labenz (10:35) It's interesting because I would think that we'd be getting better at this sort of scale up and process over time. Right? Whereas you on the drug discovery side, the clinical trial, you can sort of tell yourself a story of like, well, you know, we're becoming ever more safetyist, ever more bureaucratic. You know? Everybody's got even more boxes boxes to check on all the steps of that process. And, again, maybe, you know, to some extent for good reason. But on the material side, it seems like we should be getting better. Like, is there a warp speed? You know, if there was a sufficiently great breakthrough or a sufficiently great market need, do we have the ability to be like, we're gonna do this in a year? Or is is there just some reason that that kind of focus and energy can't be brought to bear on the material scale up in the same way, you know, was for COVID vaccines?

Jorge Colindres (11:20) Whether there's an answer to that question or not, I can assure you that Jorge makes the entire team aware that it needs to be done sub a year, and our technology is going much faster than that. No. To your point, we this is where we think autonomous experimentation is incredibly important, and we'll get into this in deep detail. Jorge will walk through why experimental results are so important to the AI framework and why our flywheel is so differentiated. But it's not just about changing the way we do discovery from an AI lens. Meaning, how do we recommend more things that are actually more attuned to the application you're trying to solve for? It is those things, but we actually can just do more experimentation faster with self driving labs today. And as we're doing that high throughput experimentation, we are not just running something in high throughput repeatedly over and over. We are using an active learning approach so that every run that we do is incorporating the learnings from the last 1, just like a human scientist would do today. And so in the lab that we are building at Radical AI, this full self driving lab, we are targeting a 100 experiments per day right now. I probably did 50 experiments per year when I was at the army research lab as a material scientist. And in some extreme cases, directed programs from agencies and the government, you might push up 4 or 500 experiments a year, like the mock program, for example. But that's a week of research for us at Radical. And so it's this compounding effect of speeding up experimentation and then this huge unlock of the experimental data. And, Jorge, I can kick that to you and kind of how relevant that becomes on the other side of the engine.

Joseph Krause (13:05) Yeah. And in experimental data is really the big unlock, we think, to all of the big materials problems. But, really, it goes even just beyond experiments. It actually, to your question, Nathan, goes all the way down into scale up, into processing, and then ultimately into manufacturing. And so to answer your question around where's this big challenge, it is quite literally in scale up. You start with, at the experimental scale, a material that's the size of a quarter, and then you're trying to create, you know, 400 pound piece of material from that thing and at the same time retain all of the properties even though you're changing it in so many different ways. That's actually really, really hard to do and a different kind of challenge than what exists in drugs, for example. And so a big part of our belief is that we wanna be able to capture experimental data at the research level, right, at that quarter size level. We call them buttons. And then we wanna be able to thread that all the way through as we continue to scale up the material to make sure that we are retaining the properties that we want. And if we're not, we're understanding why did we lose those properties? What did we introduce environmentally, chemically, or so on and so forth to change those properties, and what can we do about it in order to not change them in the future? That's a big part of the unlock that we wanna go after.

Nathan Labenz (14:18) So do we then have sort of the equivalent of orphan drugs in the materials world? We have a bunch of orphan materials that sort of We definitely created at this small scale but just have not been ever scaled up for all these kind of just very practical difficulties?

Joseph Krause (14:34) I'll I'll let Joseph answer that as the scientist.

Jorge Colindres (14:38) Yes. I would say that's probably a bulk of the work that we do, and there's different lenses of that. Right? I mean, if you look back to some of our the biggest materials that have come out, they just might have taken a longer time like a transistor, right, and the use of silicon in semiconductors. Obviously, 1 of the most talked about, most impactful, most used material discoveries today. It also took a very long time for that to really push through and continues to exponentially increase today, right, with the node sizes that we're pushing into in semiconductors. And then on the complete flip side of that, you have something like an RTAP superconductor. It's not a new idea. People have been talking about it for decades. It is 1 of the more exciting material discoveries you could probably ever make. The applications of that are really incredible, but we struggle to understand the synthesis process of that. In the most recent thought, our tap superconductor l k 99, we fail to be able to make it experimentally. And and that was even just going from theory and 1 experiment to replication. Now take exactly what Jorge said and not only be able to reproduce it at, you know, a quarter size, but then go make, in in that case, millions of tons because the world is gonna want that at large scale. That is unbelievably complex. And so there are a lot of these things that from everything from computational discovery to just making it in the lab repeatedly go to die, and that's not even stepping through the long process of manufacturing, scale up, and then really putting it into industry. What product does that material end up in? How is it used? It's a very complex process. And to Jorge's point, when you can capture all that information and use it to design better materials, better compositions, better structures from those compositions, we think we can really 1,000, in some case, 300, 400 x to speed up discovery today.

Joseph Krause (16:38) This customer aspect of it is huge, you know, in our minds. And it really does get back to that white space that Joseph was talking about a couple minutes ago. And honestly, 1 of the prime examples that we always point to is Gorilla Glass. A lot of people don't know, actually originated as a product for Corning in the fifties. And so Corning developed something called Temcor. That was in the fifties, maybe the early sixties, and it sat on a shelf for half a century. And it wasn't until about 2007, 2006 when a guy named Steve Jobs walked over to the offices in in Upstate New York and said, hey. I'm developing a phone. It's gonna completely change the world. All of our phones right now have a plastic screen. And I'm Apple. I care about design. I don't want that. I want something that feels better, that looks better. It just has these properties that I'm looking for. And Corning said, you know what? I've got a product, and it's been sitting on the shelf for a while. And eventually, they took Chemcore, they turned it into Gorilla Glass, and that sits in every Apple product that they've created over the last, you know, 2 decades. And it also sits in the products of Samsung, LG, and, I think it's probably the biggest revenue driver for Corning at this point. And so the reason we love this story is because it really alludes to this sort of concept of, number 1, we don't really understand the materials we make a lot of the time. We don't really know what their good properties are. And number 2, we actually need a customer to tell us what they're looking for in order to map up those properties to customer expectations and where value can exist. So for us, a big part of it is, again, we wanna bring in those customer expectations. We wanna bring in those value drivers very early into the process so that as we're designing materials, as we're experimenting and doing r and d, and then ultimately scaling these up and manufacturing them, we're always keeping the customer in mind, and we're always satisfying their needs. That's a really important part for us.

Nathan Labenz (18:22) Hey. We'll continue our interview in a moment after a word from our sponsors. It is an interesting time for business. Tariff and trade policies are dynamic, supply chains squeezed, and cash flow tighter than ever. If your business can't adapt in real time, you are in a world of hurt. You need total visibility from global shipments to tariff impacts to real time cash flow, and that's NetSuite by Oracle, your AI powered business management suite trusted by over 42,000 businesses. NetSuite is the number 1 cloud ERP for many reasons. It brings accounting, financial management, inventory, and HR all together into 1 suite. That gives you 1 source of truth, giving you visibility and the control you need to make quick decisions. And with real time forecasting, you're peering into the future with actionable data. Plus, with AI embedded throughout, you can automate a lot of those everyday tasks, letting your teams stay strategic. NetSuite helps you know what's stuck, what it's costing you, and how to pivot fast. Because in the AI era, there is nothing more important than speed of execution. It's 1 system, giving you full control and the ability to tame the chaos. That is NetSuite by Oracle. If your revenues are at least in the 7 figures, download the free ebook, Navigating Global Trade, 3 Insights for Leaders at netsuite.com/cognitive. That's netsuite.com/cognitive.

Nathan Labenz (19:51) I'm very eager to get into the flywheel, but just a couple more kind of foundational understanding questions before we do. 1, on the l k 99 saga, it almost sounded like you were saying that that was maybe real, but we just don't know how to reproduce it. Like, I had sort of come to the conclusion that it was a mistake. Like, that it's a measurement error that they never actually had anything. What how would you describe the field's best understanding of where that kind of landed?

Jorge Colindres (20:21) No. I think you're correct. I think, you know, the most recent thing I read was accidental creation of an impurity. Copper sulfide, I think, was 1 of the things that played out from that. However, there there is this notion around we didn't even know how to reproduce it. Things like the temperature ranges, as well as even, like, what were we making it in the container? Because, of course, that's gonna infect the affect experimental results were unknown. We're undisclosed. And so it really gets into the importance of synthesis, how challenging synthesis is. And I think the best example is to think about a cook in a kitchen. Right? If I give 2 people the same ingredients to make a loaf of bread And 1 scores the bread and lets the starter rise and puts the right temperature in bake time, and they have a really good loaf of bread at the end of it. And if someone rushes the process in 30 minutes at 1000 degrees to speed it up, Well, they're not gonna make. Okay. Hope for bread. It's gonna be a rock, and it's gonna be pretty bad to eat. And so same exact ingredients, same exact process. I'm gonna heat it. I'm gonna score. Right? But even just changing the time variable there and the temperature, entirely different output from that. So whether l k 99 is real or not, I don't know. I don't think I have enough experimental results that we've run to determine that. What I do know is that the proposed method of synthesis and the lacking details from that method don't quantify as a new discovery because a new discovery needs to be reproduced. You don't have a new material until you can make it in a lab, and you don't have a new commercial material until you can take what you made in a lab and scale that up. And so I think we have not done that yet in RTAP superconductors. Whether that material exists or not, I think time will tell.

Nathan Labenz (22:09) Fascinating. And has the field just kinda moved on from that for the moment? Like, we're all just doing other stuff now? I mean, it seems like we would wanna run that to ground collectively. Right?

Jorge Colindres (22:19) The field might have moved on, but I can tell you that we think about that problem quite frequently. And when we think about the vision for Radical AI, look. We have a beachhead that we can get into. We have to. Otherwise, you'll boil the ocean with material science. It impacts every industry. As an early stage startup, you cannot do that. You must be laser focused and really drive towards execution than scale. But when we think about what our technology can really do, our materials flywheel is attempting to remove materials as the biggest bottleneck to our most important industries. We wanna build a world where you go from a human driven process to an AI and autonomy driven process. And this approach allows us to create what we call enabling materials. We actually create the industries that we sell into. We don't just optimize the current industries today. Just like Jobs created the touch glass for, you know, smartphone and other products today, that's the exact type of applications that we wanna enable in the long run as a north star for the company. And, clearly, our tap superconductors with their immense applications are somewhere situated on that roadmap.

Joseph Krause (23:32) Yeah. 1 of the things you should understand about the company and why we founded the company is it's not because we thought there was a big opportunity in materials. I mean, that's definitely true, and it's a big part of what we're going after. But we we seriously wanted to create a company that was going to move humanity steps forward. We wanna live in a world where we can ride on a floating train. We wanna live in a world where we can actually travel to a different planet. We wanna live in a world where you can do some of the things that we think should be basic, like being able to drive a Tesla from New York to Boston without having to recharge it. Like, those are the kinds of things that we feel like they should be available to us and within our generation. And so that's the whole point of Radical AI. We wanna make those things happen, but we can't because materials are constantly getting in the way. So that's the real driver for the company, and a superconductor is probably, from a materials perspective, the epitome of an enabler. It will create so many different things, and we like to think about the second and third order effects of what will come out of the materials we create in the sense that we can't actually predict what will come out of a superconductor. We won't know what sorts of things it will allow. We won't know what sort of products will come out of it. We won't actually know what the world will look like when we have room temperatures for those. But what we do know is that it will change the world, and it will make us able to do things and approach technology in an unprecedented fashion. That's the whole point of Radical AI.

Nathan Labenz (24:59) Could you maybe give a little bit more of a range of like and I'm conscious that you just said don't know exactly how the world will look given certain breakthroughs. Obviously, it is you know, nobody would have immediately classic example from, like, the smartphone glass. Right? It's like people didn't immediately know we were gonna have Uber and DoorDash as soon as we could touch the screen. So I, you know, certainly, get that it's gonna take time for these things to play out. I get the sense that, you know, obviously, okay. Room temperature superconductor is 1. How many of these sort of canonical things do we have where we're like, if we could just Yeah. Get this property that we don't have today, it would be like a massive step change unlock. I guess I have kind of a couple layers of this question. What is, like, what is the range of those things? You know, what specific examples, you know, are kind of forefront in mind? A layer down is, like, how how precisely do those get articulated today? Is it like, you know, this thing needs to be 4% lighter and then we cross some threshold, or is it like lighter is better in general? And then I I always think back back to the Elon mask begets mask. I'm sure he didn't coin that, but that you know, I I heard him kind of go on about that. So I'm also kind of interested in, like, general mental models. If people just wanna think, like, about the material world, about them in around them in a more sophisticated way, What other things are there that are along the lines of mass begets mass where you can just develop your intuition for, like, why materials matter and how they're getting better translates to material progress? So a lot there, but take your time.

Jorge Colindres (26:31) So there are a lot of ways to think about stepwise innovation. Right? There are a couple classes of materials that I would say everyone is attuned to, hard type superconductors, high end chip alloys, ultra high temperature ceramics, CMCs, semiconductor based materials. These are people have been thinking about these industries and the novel materials that will drive fundamental impact in all of them if you discover. And it really if you notice from that answer, comes down to the end application. And we think this is so important. Jorge alluded to this a little bit earlier. We don't make materials in a vacuum at Radical AI. Because if we do, then we actually don't know the problems that we're trying to solve. And if you're an academic research group or a national lab or a nonprofit, that could be a great approach to do that. We wanna make materials that leave a fundamental impact on the human race by directly solving a huge problem that exist in industry today or helps enable a new industry in the future. And so if you're gonna take that perspective, a lot of the things that we think we can do, like the examples I just gave, are grounded in, well, what do you want to come from that? I'll give you a perfect example of the versatility to your second question. How versatile are these materials? So something like a high entropy outlet, which is a core area of focus for Radical AI today. These materials have really incredible performance across mechanical properties, thermal properties, and then really their ability to blend them together. Meaning, you get a really high strength ratio while still being effective at high temperature. It's not 1 or the other, and that's important. And these materials, these high entropy alloys can be used in something like a hypersonic application where you are flying at speeds above Mach 5, very high temperature ranges, very high pressure, dealing with things like oxidation inside the atmosphere that are really challenging for a material to solve, we have struggled to solve that problem as a nation today, and then the exact same material can actually be used in a nuclear reactor, nuclear fusion reactor, that is actually getting bombarded with radiation from this power source, and our current materials today like tungsten can't withstand that radiation and erosion over long periods of time. And so these high entropy alloys are really 1 of the options. Amorphous metal being another option. That's a subcategory of a high entropy alloy that can try to solve these problems. So the versatility is very big because the properties are so desirable. And these are, again, what we call these enabling technologies. Right? These are unlocking not a 5% improvement game, right, or even a 50% improvement game. These are a 100 x better, 50 x better than what exists today. And this is what we get really excited about. And to go come to your third question and think about the way that these can, I guess you could say, extrapolate out and get very defined? Every industry in the world can define a materials problem. If you go into the athletics industry and clothing industry, well, they're looking for novel materials to wick away sweat and reduce odor, and there are silver nanoparticles inside Lululemon workout shirts today. So that's a real application and they're trying to optimize there. And if you go to the exact other side of the dimension and look at advanced hypersonic aircraft or space travel, They're looking for intense advanced materials that can block extreme heat, block radiation, withstand the immense pressure, all while doing that in an atmosphere that is quite corrosive to the material system. And the different problems, different defining mechanisms from them, but really important to be able to straddle all those constraints. And so in summary, I mean, the real core factor that you have to think about with optimization is what are the most important properties? And when we're defining materials today, rarely is it 1 property. There are probably a few examples you could think about if you wanted to that would be a single property. We always get asked from customers, yeah, I needed to have this mechanical performance in this environment for temperature and pressure and at this cost with this supply chain. For an example, inside the defense or aerospace industry. And so you're not just optimizing on thermal expansion. You're not just optimizing on its resistance to oxidation. You're actually optimizing on what does that cost of that solution, and then what is the supply chain to be able to employ that solution with where these raw elements are coming from. And so we are always seeing this optimization across a multitude of different properties. And because of that, there becomes so many different combinations, so many different industries, so many different applications in those industries that these materials, like a high entropy alloy, can be useful. And so in our opinion, there is an endless amount of problems to solve inside of material science. And you can go very specific like optimizing the silver nanoparticles in a workout shirt, and you can go very large, like building radiation blocking shields. So when we land on the surface of Mars, that doesn't become a problem for communications electronics. Everything in between there is very, very dense and why we think it's 1 of the biggest opportunities and markets in the world.

Nathan Labenz (32:04) Hey. We'll continue our interview in a moment after a word from our sponsors. Being an entrepreneur, I can say from personal experience, can be an intimidating and at times, lonely experience. There are so many jobs to be done and often nobody to turn to when things go wrong. That's just 1 of many reasons that founders absolutely must choose their technology platforms carefully. Pick the right 1 and the technology can play important roles for you. Pick the wrong 1 and you might find yourself fighting fires alone. In the ecommerce space, of course, there's never been a better platform than Shopify. Shopify is the commerce platform behind millions of businesses around the world and 10% of all ecommerce in The United States. From household names like Mattel and Gymshark to brands just getting started. With hundreds of ready to use templates, Shopify helps you build a beautiful online store to match your brand's style, just as if you had your own design studio. With helpful AI tools that write product descriptions, page headlines, and even enhance your product photography, it's like you have your own content team. And with the ability to easily create email and social media campaigns, you can reach your customers wherever they're scrolling or strolling, just as if you had a full marketing department behind you. Best yet, Shopify is your commerce expert with world class expertise in everything from managing inventory to international shipping to processing returns and beyond. If you're ready to sell, you're ready for Shopify. Turn your big business idea into cha ching with Shopify on your side. Sign up for your $1 per month trial and start selling today at shopify.com/cognitive. Visit shopify.com/cognitive. Once more, that's shopify.com/cognitive.

Nathan Labenz (34:06) So maybe that's a great jumping off point to talk about the flywheel. If a customer comes to you with 1 of these problems, we wanna, you know, make a Mach 5 aircraft. And to do that, it's gonna have to be, you know, no heavier or maybe even lighter and stronger and able to handle heat and maybe able to handle radiation. Who knows? Maybe this thing's going out into the outer reaches of space. I guess, how Take me through the the loop that you guys get into to work your way up to actually achieving that. Maybe starting also with just, like, how often is there that Steve Jobs opportunity where something maybe does exist out there? How would you even go and search? And then if it does exist, like, where does that, you know, allow you to shortcut into the flywheel versus if no such thing exists and you have to come up with your hypothesis of, you know, never before seen materials and work from that point?

Jorge Colindres (35:00) Yeah. So the flywheel consists of, at a high level, 2 different buckets. You have an AI engine on 1 end and then this fully robotic self driving lab on the other, and those work together. I'll let Jorge start on the AI front, what that looks like. It gets into a lot of how we do property driven optimization directly to your question, and we can talk about the experiments after that.

Joseph Krause (35:23) Yeah. So when when you think of the way scientists kind of do science today, there's really a lot of different things that come to in the picture. You're looking at images out of a microscope. You're reading a bunch of papers. You're using your intuition from previous experiments. You may be doing some computational modeling. That's really all of the tools and the tools that the scientists uses to eventually come to a confidence level where they wanna go out and run an experiment. And so we try to figure out how to create, an AI engine that incorporates all of that, all of the different modalities from a data perspective that come into play there, all of the different architectures that we need to use at a modeling level, to come into play there. And then this sort of scientific intuition aspect is another part of it that we try to capture. So we bring all of those things together into 1 piece of technology that we then kind of throw at these problems. And the way we do it is, as Joseph said, by doing what we call property driven optimization, which is we start with the properties that we're looking for. Those typically come from a customer. So it's the things that you that you mentioned. Right? I would like to use these sets of elements. I would like to have it hit this oxidation level. I would like to have it hit this ductility or this strength. And then we give all of that to the AI engine, and we prompt it, and we give it some input as well. And we let it go to work, and we let it design experiments that we think are going to be successful against those properties. That's when it then hands it off to the robotic lab, which I'll let Joseph talk about.

Jorge Colindres (36:50) Yeah. And this is where so Jorge's point is really important. I love the way he said kind of how do you use the tools at your disposal? What is an experimental scientist's job? Well, you come up with a bunch of analysis hypotheses. Excuse me. You think about the ways you're gonna test them, like Jorge just mentioned, all of that's happening on the AI side. And then you go and run experiments. You make things in the lab. You test the things that you made in the lab, and you look at all the results. And that is all happening inside our self driving lab today. So our lab will receive through our operating system, you know, a a design of experiment approach. Here's the composition. Here's how we're gonna make it. You know, here's what we're looking for, and it will start executing on said experiment. And so we will make a material. That material will come out. It'll go through material characterization. We do a bunch of different characterization techniques to understand what did we make and really what does it look like, which is important to microstructure and then therefore mechanical properties. And then we'll actually run property driven tests. How strong is the material? What is the stress strain curve of the material? How does it handle under creep or crack propagation? How hard is the material? And we will do those property tests inside the lab as well. When we collect all of this data from both the material characterization and the property characterization, we actually use machine learning to analyze all of this data in real time. So we are taking this information in, we train kind of these sub models, which Roy can talk about, to learn how to analyze just like a scientist would. We start to build this intuition, start to build this understanding of if I'm optimizing for strength, then I need to pay attention to grain growth and microstructure information. I'm going to actually look at that evolve in a scanning electron microscope or in a tensile test that I actually do on set. And so as we analyze all this information, we pull out the insight from that and then feed that as active learning approach back to the AI engine where you could think about the scientist takes back the results and says, oh, that was what the outcome was of experiment number 1. Let me now run experiment number 2, updating or tweaking that process based on the results that I just saw. And if you think about what I just described to you, that's what a human scientist does. Not some insane new changing the way we do the scientific method, but we are changing the way we kind of extract information from the scientific method. When I ran that process, I did all of that in a serial based approach. I looked at 1 XRD scan, 1 SEM scan, 1 tensile test, thought about those, and then went back and showed the next thing. And I might have remembered there was a paper I read. I might look into a patent database or I might not. When our AI scientist looks at all that information, it is analyzing and comparing those results to millions of XRD scans that it might be pulling from publications. It might have done past experimental results in our lab. It might have context on other experimental datasets that exist today, you know, and then using all of that to to make a new informed hypothesis. That operates on a dimension, candidly, the human brain just cannot do. And this is where the power of the flywheel becomes so impactful. It's not just an AI scientist or an AI engine as where I described. And it's not just a self driving lab as I just described. It's the combination of these techniques together that build this lucrative approach to not only developing and designing, but then making and testing novel materials. That is how we can actually take a step forward in discovery and in the future, you scaling these materials to larger systems.

Nathan Labenz (40:44) So, yeah, I guess I wanna dig into this on a bunch of different levels. Maybe for starters, like, when you're developing the AIs, you kinda mentioned, like, sub models. I'd be interested to understand the range of models that you have. Like, we're seeing increasingly stuff posted from, like, GPT 5 on Twitter that's like Yeah. You know, this thing helped me with front think through a frontier science problem. So I wonder to what degree those sorts of things are in this flow. And then, obviously, there's gonna be more specialized models. And there, I can imagine a lot of different, you know, approaches. Obviously, we've seen many things that are sort of graph neural network structures that have some sort of radius, you know, around particular atoms. What are they connected to and kind of model things in a more, like, literally physical way? I've also seen interesting structures where it's sort of a bunch of different things, and then you kind of mask out different tokens and ask the model to sort of fill in. Right? So you could have, you know, here's the chemical composition and here's certain properties and, like, here's how we synthesize it and mask those things out and kinda train the model to fill in those details. So, yeah, can you outline, like, the sort of family of different kinds of models that you're using and, like, what the inputs and outputs are to those?

Joseph Krause (42:00) Yeah. So we use a bunch of different models. We use GNs, like you mentioned, to do modeling. We have generative models. We use language models including the GPT family of models to do everything from understanding and extracting information out of scientific literature, even generating recommendations. We use computer vision in the lab. We use a bunch of different things in a bunch of different ways. And, obviously, the inputs and outputs of all those models are slightly different depending on the use case and the model itself. But I think the core thing for us is that there is so many different tasks and things that have to be done from a material science perspective that we actually have to use different architectures and different models in different ways. But I'm I'm happy to go into more detail on any of those.

Nathan Labenz (42:40) Yeah. Please do. I mean, I'm not even sure exactly what you would consider to be the most core 1 or a few tasks. I mean, obviously, you know, anything there's at least in my experience, there's always sort of 1 or a very small number of tasks where you're like, if we could get this working super well Yeah. We know we'd, like, be successful from there. Yeah. And then there's a lot of auxiliary things that are sort of you're more confident it's gonna work and it's more kind of efficiency and ironing out details. What are the really core ones that are sort of the biggest challenges? What are the inputs and outputs of those?

Joseph Krause (43:09) Yeah. So the 1 where we're most confident it will work is on the computational side. So that's using something like a GNN. In the material science world, we call them MLIPs. There's a lot of MLIPs. They're all fairly successful in different ways, they work in different manners. We know that that's a core part of the strategy. We use them. A lot of people use them, have been using them for a long time. But where we would wanna place our bets, where we really think there's gonna be a big differentiator is actually on the language side. We think that really what we're trying to do is understand the vast space of scientific knowledge that's out there and then reason about that, incorporate different modalities and different tools in order to really come about good conclusions so that we can generate hypothesis, design new experiments, send those down into the lab, learn from the success success and failures of those experience, and iterate and adapt over time in order to improve the recommendations that we sent down there. So we we do think that at the core language models is really what's gonna drive a lot of this scientific tuition, this reasoning about what makes a good experiment and what makes a not so great experiment over the long run. Where we think the other models will come into play is sort of bolster those models. You can imagine a GNN being tapped by an LLM, for example, in order to run a computational workload because it just needs to reaffirm some uncertainty or fill in a gap around some knowledge that it doesn't quite have even though it scanned thousands of papers. The same thing could be said for some analysis coming out of the lab. Right? We may wanna look at, a bunch of SEM images and do some computer vision in order to update our intuition, update our understanding of what kinds of experiments we're running in the lab and how are those performing at this sort of microstructure level. So that's sort of how we use all the different models. But, if we were to say there's 1 thing we have to nail, it's this sort of replication of scientific intuition and the designing of what would ultimately become an experiment.

Nathan Labenz (45:00) As an aside, what's the best anyone has ever been at the human intuition on this? Like, what is the sort of ceiling of human ability? People are very obviously enamored with John von Neumann. Is there a John von Neumann you know, I I also remember this story of I don't know if this is even true, but, like, supposedly the structure of benzene came to someone originally in a dream as the you know, they envisioned, like, a snake eating its tail or something. Yeah. And they were like, oh, it's a ring structure. Okay. We've, you know, we've moved a lot past that and we've got lots more concrete knowledge to, you know, to base our intuitions on. But, like, are there people in the world that are just like an obvious cut above everybody else in terms of intuiting, like, what material is gonna have what property or how to make it successfully when nobody else knows how to how to do it? What does that human landscape look like?

Joseph Krause (45:51) Yeah. It's a great question. The answer I think is is yes. And 1 of the best examples I could point to is actually our third cofounder, He's you know, he wouldn't say this, but I think we could go out and say this. He's probably 1 of the world's leading material scientists in the top 3, if not the very best. And when you speak to him, you just get this feeling. This guy just knows stuff that the rest of us don't know. He can just understand things and connect dots that are difficult for the average person, even a great scientist. And so, absolutely, there are people who have this intuition. But the funny thing about her is when you talk to him and you sort of really probe at where does this intuition come from, he breaks it down pretty simply for you. He says, look. I've been doing this for 40 years. And so I have run a bunch of experiments over my career, or I have had a bunch of students run those experiments and I have learned alongside them, or I have read papers from all of my colleagues and the experiments that they have run. And so this scientific intuition, yes, I'm sure there's a part of it that is innate, and there are some people who are just talented at this stuff and they can become better than the rest of us, But a lot of it is actually learned. It is learned from going about and doing these things and updating your understanding of what works, what doesn't work, and that really is that scientific intuition, that that we're trying to capture. Joseph, you probably have some some thoughts on this as well.

Jorge Colindres (47:15) Well said. There's so much info in the scientific process scientists will use that is from learned experience, whether that's at the beginning of your career or even taking something you've worked on for 20, 30, 40 years and applying that to a new industry. How do they like to test that material? What properties do they care about? Is there a different way that they need to consider manufacturing? You know, I had spent the better part of my undergraduate and early PhD days on 2 d materials for optoelectronic sensors. And then I ended up getting a fellowship at the Army Research Lab and went there and was working on neuromorphic computing. And to your boy, Von Neumann, I was working on getting rid of that architecture inside semiconductors. But the the testing requirements, the saturation current, the on off ratio, the conductivity, the resistance, the memory states that we were building, all those were learned things. And after a year of being an ARL, now it's better than probably 99% of other scientists in thinking around how to design new semiconductor materials that were directly at normal for computing, that were directly using ion gated channels. There was a group of people who were experts in that as well, all from intuition and experience. But exactly to Jorge's point, it comes from this learned process through experimental research, and that is very, very hard to capture. Actually, we don't capture of the year of work I did at ARL, 90% of the stuff I did didn't work, quote, unquote. Those negative results we always talk about, but there is no recollection of that. There is no capture of that. There's no, like, dataset that exists of here are all the e beam and and PVD and ALD experiments Joseph ran over the last 12 months. There is just 1 resulting material, which we pushed out and shared with people. And so if you can't capture that and the only place that exists is up here, and there's actually, like, no framework in place today to build this intuition except with exactly what Jorge just really walked through. And that's what gets us so excited. So to Jorge's point around Heard, you know, that's why, 1, he is such an important part of what we're doing and why when we met him, said we have to build this company together. But 2, has this really unique ability to how do you replicate that? How do you build scientific intelligence? Intelligence that is not just general from like, hey. I know the English language, and I know what science is, and I know the periodic table, and electron orbitals, and the way we actually make new elements or compositions. But how do I go deeper than that and build this intuition, the scientific intelligence around, yeah, I'm gonna use these 4 elements specifically because I am targeting an x property, and that property requires them. And I just know that. How do you teach an AI that? That's quite complex to do.

Joseph Krause (50:05) Yeah. And I think there's sort of 2 layers of this. There's, like, the easy aspect of it, which is, okay. Yeah. Science is really hard to do. It's super slow. It's very manual. We don't record it, and so we should automate that, and we should learn how to do that with AI. That's not a big leap for people to take. That that's an obvious thing, and we obviously do that. I think the bigger leap here and really the 1 that we're after is that is this this sort of understanding that there is an exo scale set of data that's out there, but is virtually impossible to capture. At least it has been in the past. And the only way to get at it is through a really, really inefficient process, which is being a 4 year academic. And that's why we are bottlenecked by all of these different things and that's why we don't have materials that solve some of biggest problems. Right? Because not only are you stuck in doing it over a 4 or 5 decade time span, But in that 4 to 5 decade time span, you can really only focus on 1 area, maybe 2 areas. And so you're just limited by the amount of things that you can do on top of it being an inefficient process. And yet, like Joseph talked about, we know the data is out there. And so a big part of our belief is if we can build a system that can actually tap into that data and then figure out how to use it in a good and efficient way, we can couple that with all of the advances that we've had in machine learning to really drive at this problem that takes 40 to 50 years and shrink that down so that it's sub a decade. And, of course, that will translate to solving materials problems to sub a decade as well. That's a big part of our belief.

Nathan Labenz (51:39) For the people that have this intuition, how sort of language grounded is it? How much can they articulate it versus how much does it live in a sort of fundamentally different space than language? Because I could imagine somebody might say, I I can't really tell you why I think this is gonna work. I just think that if we do it this way, it might work. Or you could have, you know, obviously a very mechanistic and clearly articulatable theory of why it's gonna work. In my brief chemistry experience, it was honestly a little bit more of a former from what I mean, and I had a limited time there and a limited view. But the 1 reaction that I worked on developing for a year, we never really had a clear mechanism, the sort of it was a catalyzed palladium catalyzed hydrocarbon oxygen oxidation at room temperature or slightly warmer, but, like, mild conditions, whatever. Anyway, that who cares? But the point was we never really knew what was going on at the reaction center. We had kind of vague sense of that. And so it was a very kind of, yeah, maybe we'll try this and we'll try that, you know, but without, like, a without any precise articulation. Yeah. How common is that, and what does that imply for whether, like, g p t 5 is gonna be able to sort of, you know, chain of thought its way into making these things happen versus whether there's just kind of different modalities of data that have to be, you know, collected and learned from?

Jorge Colindres (53:03) Yeah. You took the ant 1 of the answers there, which is modalities of the it's very important, and we'll talk about JC can go into deep detail around where that comes in AI engine and why it's very, very important. Language is an important piece. It's not just language as you just alluded to. You know, scientists learn a lot from the different data streams that they're pulling inside their experimental process. There's this 1 other thing that is a form of data, but it's not maybe what you call structured data from an experimental tool or from an output in the lab. And that is this, I guess, intuition, for lack of a better word, or experience you have in running experiments. I'll give you a perfect example. When I was working on these 2 d semiconductor materials, we would always look for the transparency of the film. And the technical reason why that we later came to figure out was they were monolayer 2 d TMD materials, not multilayer 2 d TMD materials, and that drove way better conductivity, etcetera. We didn't know that upfront. And so you could see this in an optical microscope. We weren't looking for this in an optical microscope. We weren't taking pictures of transparency and, you know, what what are the what are the changes in transparency until we kind of we're racking our brain around, like, why are we getting such better conductivity in these materials? They're the same material and we're making them the same way. And it actually came from an expert in the 2 d material field who said, oh, those are monolayer films. And we were like, what? Like, we know what monolayer films. I worked on monolayer films for 3 years, but never in that setting have we thought about looking for monolayer based films. And so this randomness, this accidental procedure that you hear all about famous science stories is really important to capture, this negative result based process. And that is almost like a separate form of data where we deeply believe it's not just making the things that work. It's not just using the intuition from multimodal datasets that actually allow you to predict new things. It is looking at, exploring, and capturing the 1 off random unexpected outcomes that might have a key insight in that a human scientist will look right past, but an AI engine would never look past. There's a very important piece of what is, I guess, you could call intuition as well. But on the multimodal data, that is also equally as important, and there's a huge kind of approach we take on that side.

Joseph Krause (55:37) Yeah. The multimodal aspect of it is is is just enormous because it's gonna get set this sort of, this idea that the data is out there. It's just in so many different ways and forms and shapes, and it's so unorganized. And if you can bring a little bit of order to it, and a little bit of of cleanliness to it, then you can start to really enrich your representations and really get to better predictions and and generations over time. But I do wanna go back to what Joseph talked about because this is, I think, something that's actually quite hard to do with machine learning. Right? When you have models that effectively learn out of a distribution of data, it becomes very, very challenging to have it do things that are out of distribution, which is effectively what Joseph is talking about. How do you go about this sort of, like, randomness when the entire purpose of the model is to learn from a distribution where those random things are generally just ignored by the model. Right? Or they're sort of said, like, now that's out of distribution, and I'm not gonna appear in that direction. And so we think about being very purposeful with that. We think about saying, look. In science, it's a little bit different. We have to do the sort of normal thing that we always do, and we have to learn from a distribution of data. And if we can improve that data over time and better and better enrich the modalities of data that come into that, we'll be good. But that's not entirely true. We have to also design systems that are effectively prized towards novelty and surprise because that is really where some of the biggest discoveries in science come from. They come from some weird corner of the chemical universe that everyone thought was kind of ugly and they didn't wanna look at and lo and behold, here's this amazing scientific discovery or pure accident sometimes. And so we have to design systems that not only learn really well from good distributions of data, but also figure out how to basically go hunch hunting. Right? How to develop hunches and then say, know what? I'm gonna go out and pursue that thing. I'm actually gonna go deeper here. And there's a bunch of different ways we we think about doing that. You could do some sort of randomization or diversification, which we do through, like, Bayesian approaches and Monte Carlo and stuff like that you can do. You can do noise sampling. Active learning is a huge part of this for us. But then there's another side of it, which is, again, the the sort of underlying design of the system itself, really has to be geared towards capturing scientific intuition at the end of the day because that is such an important part of it. And so it's what would drive a scientist to pursue something that is a little bit weird, but just, like Joseph said, at 10 or 20 x or even a 100 x the scale. That that's really, really important to us. And so we have a bunch of techniques that we use for that as well, reinforcement learning, and things like that. Human in the loop is a big part of it as well in the early days for us in sort of capturing this uncapturable data, like Joseph was talking about too. So there's a bunch of different techniques that we think are really, really important that go beyond just the normal, hey. Train a model on a good distribution of data.

Nathan Labenz (58:27) Yeah. That's really fascinating. I mean, that's also basically the core problem in all of AI, I would say, in some sense. Right? It's like for better or worse, you know, are we gonna take these things past the point where they model the past Mhmm. And into the point where you know, into the realm where they can sort of figure out the future. Yeah. I mean, I'd love to hear more about that in kind of any way you wanna go deeper on it. But 1, to try to map a little bit of what you have experimented with and learned onto the fully general problem, how do you see, like, base models, so to speak, versus, like, RLHF versus just pure success signal RL, kind of fitting into that? Because it it seems like people are that middle ground of RLHF is pretty good for the human assistant, and there's part of me that is concerned with, like, big picture, let's say, AI safety questions. It's like, man, as much as we can get the things to, like, imitate us, I'd I'd really like that because I I do worry about, you know, if we just go super hard on RL and give nothing but reward signal, what are these things gonna learn and what, you know, reward hacks might we find ourselves encountering? But the flip side, I would guess, is that if you just sort of imitate, you know, humans too much or just get the thumbs up from the human evaluator too much, you maybe sand down a lot of these rough edges. That certainly seems to be what we see from the chatbots. Right? People are like, the base model is more creative. I like that more in some ways, especially if I wanna get, you know, kinda outside the box ideas. The umbrella h f kinda gets boring, and then the RL again gets interesting, but sometimes in, like, kinda scary ways depending on exactly, you know, what sort of problem you've put in front of it. How would you compare and contrast your learning against that stylized story from the Yeah. The chatbot space?

Joseph Krause (1:00:13) You know, I think a lot of it is is definitely true. For us, a big part of it is is do you have good evals? At the end of the day, if you can develop good evals and that helps you go about this process because a lot of it is undiscovered and a lot of it is true research on the ML side and trying to understand how to bring these things together and move them in a direction that you want to over time with, like you said, not forgetting about safety and and some of the other things that that come into play as well, which obviously in material science and and in chemistry is an important thing. So eval is hugely important. And for us, there honestly is no better eval than the lab. I mean, that is as ground truth as it gets. Right? If you're looking for a good benchmark, a scientific lab is pretty good as a benchmark. It's a really good way to understand, well, did my model actually successfully do what it was supposed to do? And did it do so in a way that didn't come at the expense or the sacrifice of some these other key things that I really wanted to hold? So that's a really, really important part of it for us, and it brings us back to, again, some of these active learning loops that we were talking about earlier and how do we smartly incorporate that so that we can update these models, not just kind of, like, at the weight level because it it needs to go beyond that. It needs to be a much more dynamic process than, oh, we're just gonna retrain the model. There really needs to be a sort of, like, surrogate or agentic process that sits on top of it and says, yeah. I'm gonna use my model to kind of get things going, and hopefully, we're gonna improve that model over the long run. But in the meantime, I need to be able to be adaptive and responsive to the new learnings that are coming back out of, again, what is a best in class benchmark or a best in class eval aka the scientific lab. So that's a big part of it for us. And I think the other thing that I would bring up is that we do truly believe in trying to understand how these models work. We're making a ton of investments in mechanistic interpretability, for example, because we think that's gonna drive a lot of performance and a lot of gain. And the correct steering of these models towards the objectives that we have. And, again, if you think about what Radical AI is trying to do and the core purpose of the company, it's to solve really big materials problems for people. And so if our models are just not going in the direction we need them to, we need to be able to move them in that direction. If they're doing against the expectations or the wishes of an end application or an end customer, then we need to be able to curtail those things too. And so it goes a lot further for us than just, we'll give it to the model and let it do its thing. And, you know, it's okay if it's a black box, and we'll just learn how to improve that black box over time. We really need to be able to use it as a tool over the long run if we wanna solve some of these big hard problems.

Jorge Colindres (1:02:42) Especially going back to 1 of the things that you brought up earlier, like, this is the nexus or this is the 1 of the big problems in AI. You coupling on top of that that we don't have historical datasets to tap into. Like, even in the English language, you can, like, digitalize, you know, old books, right, and bring them in and learn about how Shakespeare wrote and and kind of build intuition or or knowledge off of that. We don't have Newton's notebook. And even if we did, it'd be scratches of all the things that they tried to do. There wouldn't be a written text that we can directly learn and and transcribe. And so coupling on top of that problem is this kind of idea that not only do we have to know where they need to go, but, like, how what can we bring in so that they know the different directions that they can go? Because we don't have a different text of the way we write books today and the way that Shade Sphere wrote books back in the day. If you don't have that intuition, you don't even have that understanding, it becomes really hard to think about how to index that. And I think you've seen this in software. Right? Like, where a lot of the best software is private. So a lot of the coding agents use what's open source to train on. And that's gotten really good because there is an open source community that pushes really good software. And so it can learn what really good software looks like and then utilize that data in an effective way. There is no open source science community. There's open source science from publication. I mean that there is no multimillion data point experimental set of, oh, these are all the scientists who have subscribed into, who have opted into publishing their lab notebooks. We don't even have that, let alone structured data like pulling a GitHub repo and being able to learn from, okay, here's how professional engineers write code, here's how it's organized. We can't do that. So if you layer that on top of exactly what Jorge just talked about, it is really important to know which way our models are going so that we can get in there and kind of, like, bring a scientist in to to trying to drive them to better discoveries. Because exactly like where I said, our objective is better discoveries that fundamentally impact humanity. It's not just about, you know, producing or recommending new things that are marginally an improvement. That's not the north star for the company.

Nathan Labenz (1:05:01) So I definitely wanna get into the mechanistic side in a second. Shout out to Eric, CEO of Goodfire, who originally introduced us. But maybe first on data. You've said a couple times, like, data is the limiting factor. And, you know, the big 3 the holy trinity, right, of of AI progress is data compute and algorithms. I was struck in looking at some of the specialized models by just how small they are, like, 30 some million parameters in some cases, like, you know, orders of magnitude smaller, 4 4 orders of magnitude, you know, smaller than some of the real big ones, even 5 maybe. That's lot of orders of magnitude. So that would, you know, reflect, guess, obviously, that data must be scarce. Otherwise, I would assume you'd have a bigger model. Where where does that future data come from? I mean, you said sort of like, you know, we don't have the history. I was kind of wondering I mean, I think of an analogy a little bit also to the computer use agents. I've remarked many times and still nobody has come forward to pay me to watch me use my computer. So I'm like, why is nobody just offering me a $100 a month or whatever to install some observer on my computer? Mhmm. And I kind of wonder, you know and there could be different x moving, but it's too messy or whatever. It's not it's maybe not worth the trouble. So the sources of data I had in mind for you guys were, 1, literally going and paying grad students to let you, like, you know, digitize their notebooks. Yeah. I don't know what the going rate for that would be. Grad students who, you know, would probably be interested in the money, I would think. I don't know if there's, you know, IP concerns there or whatever. So, you know, has that happened? Why not? Then there's just the straight, like, simulation data, which I understand is at the core of a lot of sort of at least filling in gaps for some of these, like, small kind of force prediction. What's the next time step in evolution of a collection of atoms kinda model? That would be 1 way you could, like, invest heavily in compute and get data, but I'm not sure, you know, what gaps that may leave or, like, why that where that falls short. And then, obviously, the lab is kind of the the final stop. And maybe I'm even missing categories. What do you think the portfolio is of contribution from those different types of data going forward? Are you ultimately kind of like, our progress is going to be rate limited by how many of these autonomous factories we can build and how many experiments we can run per day, or is there any other way to convert, like, financial resources into progress that doesn't go through that sort of factory build out?

Joseph Krause (1:07:26) Yeah. I mean, data is always gonna be the learning factor. Certainly, in our field, where we're so data limited, but I would I would even argue, generally speaking, across all areas of machine learning, data is gonna be your biggest driver of success. And if you can figure out a way to get over the hurdle, on the data side, then you're gonna be in a good in in in a good spot. Yeah. Improvements, you know, in models will help and certainly access to compute and and being able to do more with compute is a big part of it too. But, I think if you asked anyone to choose between each of those 3, they would almost universally say, yeah. Give me data. Give me data at the end of the day, and and the same would be set for us. Joseph can probably answer the mix of data and and where we think that will come into play for us.

Jorge Colindres (1:08:07) Yep. It's even harder than just a mix of data too because it's all unstructured. It's not even labeled, actually. Like, if we could get access to every lab notebook in North America today, that would be great. We'll we'll take it, by the way, anyone's offering that. We will we will take it because to Jorge's exact point, we'll always take data. And then we probably have to go hire an entire team to sit down with everyone who wrote those notes and understand what they wrote and what is it in reference to. I mean, even generalized simple things like running an X-ray diffraction tool. Like, the peaks mean things, and people might interpret them differently for their research than we interpret them. When we're doing amorphous based materials, we're using XRD as a con confirmation that there is no crystal structure. We're actually not even looking for relevant data. So if you gave that to someone who had never worked in a morphous materials, they'd probably be like, I don't there's no peaks here. I don't there's what's I don't get it. Right? Like, like, this is irrelevant to me. I can't train a model on learning XRD. Say, but you can train 1 on learning amorphous behavior because there isn't this, you know, of course, crystal structure. So maybe that's, like, a very technical example, but it's very important. It's like even if we had access to all of this, it's not labeled. It's not organized. It's not structured. It's interpreted in wildly different ways based on the discipline that you're working within, and most of it is, like, only a piece of the data. For example, I used to always write down the time stamp of halfway through my reaction because it didn't matter when I started, and I would always wind down temperature. I would, like, turn the temp right knob off at halfway, and then I couldn't control the cooling. So however long it took to cool to the end of experiment was irrelevant. So I gave you a time stamp, 1 minute, 1 minute and 6 seconds, 1 minute and 4 seconds, 58 seconds. What does that mean for us doing the same experiments in our lab? Those might be relevant metrics. They might not. If I sit there and tell you, oh, that's the halfway point and I cut temperature exactly at that point, well, now, yes, of course, the data becomes more relevant to what we're trying to do. But so this just decide this challenge in materials is not only is it multimodal, not only is it not captured anywhere, not only is it really important to, as Jorge said, evals, ground truth, benchmarks. We don't even have it structured and and and annotated to be able to use it even if all 3 of those things that is mentioned existed. And so this data problem, I think, will actually spur up an entire industry. We have a lot of ideas on the way to get around that, which we can dive into. And from a scientist side of understanding the data output I produce as a material scientist, that in of itself is is is quite a big challenge inside and out.

Nathan Labenz (1:10:55) Yeah. On the computer use side, I think maybe 1 of the reasons that it hasn't been done is sort of analogous to what you're describing, which is, like, at any given time, it's not immediately clear what I'm trying to even do as I'm using my computer. So there would have to be a sort of or probably would need to be. You could guess, obviously, but you would be greatly benefited if you could also capture the sort of, you know, internal monologue that is going on when the person's like

Jorge Colindres (1:11:22) Exactly.

Nathan Labenz (1:11:23) Going tab to tab or whatever. Like, what is it that you're even trying to do here? So I I take the point that that is generally and I can think back to my own, you know, not super well documented or super tidy or thorough notebook notes myself to

Jorge Colindres (1:11:36) And how do you judge quality there? Right? If my grandmother records her use of a computer and you record your because I hope they're gonna be different. I think they're gonna be different. How do you delineate that? What sets the bar of not great computer user and great computer user? Now bring that into a subjective field of science. Was I a good scientist or bad? Well, my boss and the people I used to work with think I was great. Other scientists in the field might think I was terrible. I would have not tried those experiments. I would have went in a way different direction. Would have never used molybdenum or used tungsten instead, whatever the property is. So who sets that bar? Who determines how it moves? How do you even begin to build it when there are so many different fields inside material science? So every single bar or benchmark is different. There's, like, these deep problems that come up and even just annotating the data that are really hard to answer. And I and, like, in the computer example, I don't know how I would answer that. I would generally start to say, well, someone who's very proficient at moving around and understands where all their applications are and, you know, can get their tasks done quickly. But what if someone goes to email and and opens a new email, pops it out into a new tab, goes to somewhere, writes in the notes at app, copies that, and then paste it in Gmail's email box. It's like, okay. Well, I wouldn't do that. Right? Like, that that's not the most efficient way to write an email. So how do you delineate that? There's probably someone who would argue it is. Oh, I make too many errors when I send emails. I have to write it in notes with Grammarly installed and check it first. Who's right? So this is so much ambiguity inside classifying or, I guess, it's just, like, labeling the data too. It just becomes really an exponential problem.

Joseph Krause (1:13:23) Yeah. I think I think what we're hitting on is that it's pretty abundantly clear that the highest value data comes out of that experimental setting out of lab.

Jorge Colindres (1:13:31) Yeah.

Joseph Krause (1:13:31) And yet, it's also where the biggest challenge lies. Right? Just gaining access to the data and doing it in a way that's orchestrated and clean is really, really challenging, not to mention expensive. And that's why we just we don't have the data. We have not seen people really build these datasets out because it is so hard to do. And actually, if you flip to the other side of the equation, the computational side, well, that's actually a lot more straightforward, and we've seen major major push to develop datasets there. But our big thought is, look, that only solves, like, 10% of the problem. You're not there. And it's because of this intuition again, where we know that the highest value is in the lab anyways. It's just really hard to get to. And so, yes, computational data is important, and it's helpful to us, and we're gonna do that too. But we're kind of missing the real big picture here, which is let's figure out the lab. Let's figure out how to get experimental data and incorporate it alongside the computational stuff that we're already getting pretty good at, and that is really gonna push things forward. That that's really what we're trying to do here.

Jorge Colindres (1:14:32) And if you do do that worth adding on, you actually control the problems I just brought up. It's no longer unstructured. It no longer lacks context. It's no longer unlabeled. All of those things become a reality. And so not only are we building the dataset, but we actually can use the dataset. And those 2 things are incredibly effective, in our opinion, at really developing new materials to point.

Nathan Labenz (1:14:54) Yeah. That's really interesting. You mentioned earlier the search space is vast. Could you maybe help me develop my intuition for the vastness of the search space? Like, how many, you know, different dimensions? Or are there sort of big branches in it that are kind of, know, just critical to understand forks in the road as you're as you're going down a thought process or ultimately, obviously, a material development process? And then, like, how much of that in terms of, you know, as you build out the capability of the lab automation, how much of that space can you effectively explore today? Are there parts of it that are like, we can't automate that yet? That's still kind of you know, because we can't automate plumbing in my 100 year old house. Right? So and, again, I remember doing chemistry stuff. Like, the 1 of the things I observed in my grad student mentor was like, this dude just knows how to do certain things, you know, in a very intuitive way that I don't know and that certainly at that time no robot had. So as you think about pushing the frontier of what you can automate experimentally, some of it is just horizontal build out at some point, right, to do more throughput. But presumably, there's also frontiers of, like, certain kinds of experiments or reactions or what have you that, you know, are just not yet automatable that become automatable. So, again, take as long as you want. How would you describe the vastness of this space and what parts of it can we explore? Can't we explore what, you know, what what frontiers do you think you'll be pushing the automation forward to to actually be able to to get into and and properly explore?

Jorge Colindres (1:16:27) Yes. It is a incredibly vast search space. It is impossible. It is improbable to think you're going to test every single potential combination. We always refer to when we're talking about the company to Eddington's number, I believe it's called, 10 to the 80 different observable atoms in the universe. And that is the number we use for the potential combinations. If you go into something like alloys specifically, which is of course our area, you know, there are 60 or so give or take elements on the periodic table that you're gonna use in novel alloy development. And so if you just, you know, and we when we talk about high energy values, we're doing 5 or 6 elements. This is called multi component alloy systems. If you just take 5 elements and and, you know, use that, it's like 5 and around 5,000,000 in change and potential combinations that you can develop, and that's if they're all equally balanced. So meaning 20 20 20 20 20 percentages inside that alloy. No alloy, to my knowledge, is is that equally balanced. And so within each individual element, of course, you have a 0 to 100%, and then even that many decimal places you wanna go into optimize. And so it is massive. The way we always describe it, it is a single grain of sand on an incredibly large beach, a beach that's the size of the earth. That's what you're really trying to identify and find and looking for a new alloy or or a new material. Now there are spaces that can be a little bit bigger than that. I think proteins are 1 to call out that everyone tracks. Everyone thinks about how big they are. I think it's something like 10 to the 130, I think, that the the search space is there. That's probably just a single grain across multiple earths or multiple universes even maybe. That's huge. We think that's a huge advantage for materials. It's actually more likely that we'll find, hopefully, something in our in our beach versus the bigger beach, but it is really big. And so when we think about, okay, how do we take that big number? And then we actually start testing things and actually get to, okay, how do you drive towards discovering now? Well, the first is indexing the search space, and all of that sits within the AI engine and how do we actually do property driven optimization to weed down and filter out with the best materials that we wanna try. And then when we think about the challenge on the experimental side, it's really the material tools themselves. It is not robotics or different automation systems like our rover track that we build that is our materials handling system. It's the tooling themselves that we need to conduct scientific experiments. Those companies have been around for a very long time, you know, some 25, some 50, some 75, 100 years. They have been human operated for decades, so they're built entirely around a human scientist with no automation in consideration. And the ability to then drive those tools with software are, in some cases, nonexistent. Or in the cases that is existent, there's some type of interface, very, very limited in the actuation you can drive. Or you might be able to, like, get outputs from the tool, but you can't control. Can't control the turbo pump. You can't control the heat source. You can't control the vacuum chamber and and, like, the pressure that's inside the the chamber. Therefore, the rendering an effective tool from an automation standpoint. So a lot of our time on the automation front is spent on how do we take existing tooling infrastructure and make them autonomous, make the tool itself be able to operate in an autonomous fashion. We do a lot of custom development ourselves. We have multiple tools in the lab that are entirely custom designed around the material tool to make it automated. We work with companies. We have multiple ongoing conversations with companies where we wanna build a custom tool together. So that's taking that outside automation and bringing it into the real engineering and design of the tool. And then there are some tools that we think will probably be pushed out from its lack of automation or from the ground up by companies because the way that they were built was just entirely human driven. And we have examples in our lab today where, you know, they're built in a very, very mechanical way, meaning humans flip switches and move knobs and turn dials. Those things we think will be replaced over the next 10 years in in these new tools that come out. And so from an experimental setting, a lot of our effort is spent in taking the existing tools today and making them automated. Remember, Radical AI is not a material tooling company. We don't wanna sell XRDs. We wanna sell materials. And so we work within companies and vendors and do a lot of that work ourselves. So that's how we do it on the experimental side. On the AI side, the search question is challenging because we do not do a combinatorial based approach. We really don't think combinatorial based science has driven enough value. We again, we move towards this property driven optimization that has this active learning component and is using a bunch of different tools and generating there. Jorge can kind of go into details if you want there, but we really kinda take a novel approach in our opinion to generating new systems to try.

Joseph Krause (1:21:51) I mean, the big the big aspect of it is is definitely the active learning side of it. But, you know, even in this sort of universe of, hey. We have, you know, dozens of elements that we might wanna use and and and how do we go about doing that? Some of that just goes away. Right? Because for example, if I really, really care about weight, I'm not gonna use a little bit. So you could just remove a little bit. So some of these things do just start to naturally shrink down. You still end up with a very, very, very large number, And this is why you need to have multiple machine learning models that can solve multiple problems kind of all along that trajectory so that by the time you get to the active learning loop, you're kind of sending really good stuff down into the lab. And yes, you're learning how to make them in the right way. But hopefully, the things that you send down to the lab are already good to begin with.

Nathan Labenz (1:22:39) Just 1 more brief follow-up on that, and then there's so many different interesting directions to go. Are there things right now where, like, you have to either patch with a human? It sounds like I know, like, Emerald Cloud Lab from what I understand has certain aspects of the process where it's like a person is told, like, go to this machine, pick up this file, move it to this other machine, put it in, and they're almost sort of Chinese rooming it to a degree with a person who doesn't necessarily even have to understand what's going on, but is just following, like, very point level instructions. Yep. Are there things like that that you are kind of bridging missing tools, and are there still sort of certain kinds of things that you just can't get to that, you know, that a a purely artisanal process, like, could do that are just, like, outside the bounds of what you can bring into the the closed loop as it exists today?

Jorge Colindres (1:23:30) The answer to that is yes today, meaning there are some things we don't have fully automated. So we do have scientists in the lab who who help orchestrate experiments. But the answer is also no in that while they end at that solution, like, that's our that's our starting point. We we are actually beginning our learning process from there. And of all the tools we have in the lab, everyone starts at the place I just described. Human scientists coming in, showing how you operate and running it, and moves to fully autonomous, no human in the way, inside running that tool specifically. And so for all of our tools, we will reach that place. And if a tool cannot or we think the value is not there to go through, you know, making it automated, we will go work with someone else to custom build a tool specifically to our needs. Now I don't mean custom build the science part of the tool. Actually, usually don't touch the science part of the tool. How the tool operates, we don't get into that. We are touching the loading chamber, the control of the vacuum, the control of when the laser turns on and when it turns off, The kind of shuffling of samples and prevention of contamination if there's multiple samples inside a chamber. All mechanically driven optimization focused, or I guess you could say autonomy focused functions of the tool. We are not redesigning how to do X-ray diffraction spectroscopy. That's not we don't do that as a company. We don't want to do that. Again, we're not a material tools company. We are material discovery and then, therefore, material manufacturing company. And that's what we want to move towards. The way that we can get there kind of can look different ways, and we are going through that process with all of our tools today. But, no, we don't have every single tool we've ever bought fully autonomous today. There are some tools that are going through that custom design process right now and and retrofitting process right now that aren't just fully automated yet. Cool.

Nathan Labenz (1:25:32) Alright. You mentioned the the idea of, like, hopefully, what you send down to the lab is good. Couple questions on that. 1, is there a sort of inference time scaling law that you can tap into where you're sort of like, if we run 10 times more experiments or, you know, in silico, 10 times more and I I guess, a lot of different kind of ways to think about into that. We have this desired property. We have this, you know, set sort of constraint around inputs. You know, generate me a bunch of protocols or generate me, like, the extra how am I what am I gonna dope it with or whatever. Right? Like, you could have, like, a lot of different guesses. Is there an a sort of known relationship between 10 x more of the comp use of the models, you know, the 10 10 x more compute translates to certain higher hit rate? And how does that relate to the sort of loss landscape, if you will, of materials? I'm not sure if my intuition should be if I come up with a material that has, like, pretty good but, you know, but not quite properties, does that mean I'm close where I can, like, you know, explore the very local space around that particular thing and expect that there's gonna be something better that I can kinda gradient descent my way into? Mhmm. Or I could also imagine that there's could be just, like, a ton of weird, like, discontinuities in the material space such that, you know, I might take 1 step in a particular direction and be, like, in a totally different property regime.

Joseph Krause (1:26:55) Yeah. I mean, there's there's always this exploit conversation that's coming into play, regardless of what stage you're at. You you could be at the purely atomistic level where you're again talking about trillions and trillions of different atomic systems that you might wanna look at, and you're gonna wanna explore exploit there. Right? You're gonna wanna explore the vastness of it. And then when you find something that feels pretty good, you're gonna wanna narrow down and go deep there. And as you kinda go down the ladder and eventually get yourself into the lab, it's going to be the same question over and over again. And a big part of it is at each layer, you need to do that slightly differently because the types of questions you're trying to answer in order to understand where you should be exploiting are actually just different. What you are trying to answer at the atomistic level is very, very different in terms of, well, what should I be synthesizing in the lab? Are they related? A 100%. They are absolutely related, but they are a different set of questions, and you need to have different techniques in order to answer those robustly and soundly. And so that's that's just something that we're constantly doing over and over again. In terms of inference time and and and and all of that kind of stuff, I mean, generally, the way I like to think about it is if you actually just forget Radical AI, you forget machine learning, you forget automation, you just say, hey. As a scientist, would it make sense for me to think about a bunch of things kind of in my head, come up with a bunch of hypothesis, narrow those down in my own head, maybe on some scratch paper, maybe do a little bit of modeling, and then walk into the lab? Or should I just say, you know what? I'm going straight into the lab. Which 1 is likely to yield the best result? That's probably the former, and I think that same thing holds and applies. Again, you do have to always keep in mind that science doesn't always work in a logical rational way. There is just a natural degree of of randomness and counterintuition that has to come into play, and you cannot forget that. We have to design every single system with this idea in mind. We have to abide by the general rules, but also understand that we need to be able to break them too. And so everything is designed around that core thought process. But yeah. I mean, to to answer your question, if we can offload some of these things in simulation, that naturally makes a lot of sense, and we will do that. We just have to do it in different ways at different stages along the process.

Nathan Labenz (1:29:06) Can you talk a little bit more about the active learning? I mean, we've got these sort of, like, few shot prompting, no type things. Again, there's, you know, test time, training. There's a bunch of different paradigms where people are trying to push this, but it seems like, again, this is 1 of the core questions is, like, can we really get them over the hump to where they can take on this new data that is genuinely out of distribution

Joseph Krause (1:29:29) Mhmm.

Nathan Labenz (1:29:29) And, like, do an effective update on that. What more can you tell us about, like, how you're getting that to work?

Joseph Krause (1:29:38) Yeah. We do active learning, again, kind of in every every state. So at the atomistic level, we do active learning, and that's different active learning than what we do down in the lab. But we do do it all the time to improve our recommendations, to improve the likelihood that as we kind of move into the next phase, it's better and better. So at the atomistic level, we might be doing some active learning around uncertainty. And so we might be trying to generate these recommendations around forces. If And we're not feeling pretty good, we might run some DFT and then kind of update our understanding of how we're making our predictions at the atomistic level. In the lab, that's a totally different thing, but we're still doing active learning. We might run an experiment and realize, you know what? Even though we scanned a bunch of literature, even though we've done all of these calculations, we didn't actually realize that the pressure and temperature is just a little bit different today. And so we actually need to update it for that next experimental run. And so we do do active learning all of the time, and that's mostly because science is not necessarily predictable, to a perfect t. It actually does change in a bunch of different ways, and it kinda comes back to this core thing around human aspects, is number 1, we just don't have the data. It's out there, but we don't really have it. And the more of it we can get and the more of it we can get in a good fashion, the more likely we are to move successfully through that process. But then number 2, never forget that there always has to be some degree of randomness that we have to lean into order to open ourselves up to the greater discoveries. And so we do do active learning kind of all along the way. We think it's just an important part of the process that, you know, naturally happens in science. And if we're trying to replicate and improve and and upscale the way that we do science, we have to include that into it as well.

Nathan Labenz (1:31:16) Yeah. That's the core stuff these days. So we've we've made it pretty far in terms of learning the the past distribution. So this how to take a surprise on board and really do the right thing in response to that. That's, yeah, it's central in in all areas. In terms of, like, interpretability, you know, a bit earlier, Joseph, you kind of told the story of the, you know, the grizzled vet who was like, oh, that's a, you know, that's a monolayer film you've got there. That's why that's happening. To sort of tell a little hackneyed story there, it seemed like you had, you know, gradient descended your way into a heuristic, I guess. Like, if I understand correctly, you had sort of realized without a theory that, you know, when they look more transparent, that seems good. Right. And it seems like a lot of models today, certainly thinking at the language model level at least, kind of operate in this sort of heuristic zone where they've, like, learned the patterns, but they haven't really grokked the underlying laws. And then, you know, you've got this veteran that comes in and says, oh, well, I've grokked that situation, and I can give you not just an you know, I can actually tell you why your intuition has has any value with a, you know, something that's rooted in the actual physics of the question. I basically think that, like, a lot of that is going on in models as far as I can tell. I know I always think back to the original gracking paper from OpenAI where, you know, you get heuristics really fast. You can memorize the the training set really fast. It takes, like, orders of magnitude longer in that original experiment to actually do the grokking and really implement an algorithm that does, you know, the modular division. And then it took interpretability to even know what's going on. You know, is it still just guessing, or does it have some actual, like, right way that we can trust to solve this problem? Can we be confident that it will actually work for, like, all unseen cases? So where are you in the end? You know, I guess, where where do you think the models are, and to what degree have you been able to, like, do the interpretability at this point to have a sense for where the models are on these, like, grocking curves? Are we how how much of the time are you getting stuff out that's still, like, heuristic and not really reflecting any real understanding versus, you know, are there things that you can point to and say, like, demonstrably, like, we see real understanding here that, you know, is principled in any particular model in any particular area?

Joseph Krause (1:33:44) So I wouldn't say that we are in a place where we can demonstrably point to a very clear and definitive sense of understanding. What I would say is that depending on the models that we're working with, you see varying levels of this. And it it's almost like in the inverse where you see sort of signs of memorization. You see signs of obviously hallucination and that kind of stuff as well. And so I think the fact that you could kind of see the inverse means that you know there should be a way to get to to the opposite end of it as well. And this is why we are so bullish on interpretability really playing a big role here is because we think that there is a a realistic world where where these models can actually develop a rich representation of a materials problem and that we can then therefore sort of, like, learn how to exploit that model and steer it in the right direction. But I wouldn't say that we have any proof or anything like that just yet.

Jorge Colindres (1:34:40) And the 1 thing I would add on from a scientist perspective is and and this is 1 of the reasons we got excited about working with with Eric and the Goodfire team is is this potential ability to try to unearth things like this is they've she I don't know if we know. Actually, I I do know that we don't know what the

Nathan Labenz (1:35:00) I know that we don't know.

Jorge Colindres (1:35:01) Yes. The of what we might want to look at to help impact our discovery, like the monolayer film on an optical microscope and seeing the transparency of that. If you had asked me before, I I would never have said to even look for that. And so, like, out of the machine learning world, like, at a base level, like, at a scientific level, We I don't even think we fully know, like, which areas we can look into, what heuristics we can try to capture that might be impactful, and in some cases, might not. Like, there there's probably a world where the transparency of something does not matter to the end resulting property that you're going after. So there's the 4. It's not effective, but maybe it's transferable. Right? And maybe there is something that you can actually learn from that process that comes into a different heuristic that you use. And I think at a scientific level, as a human scientist, this is why good fire and interpretability in this concept to me is so exciting. This ability for AI to maybe teach you what those heuristics are for lack of for a very simplistic phrase and actually use those to drive different results that you would never be able to insinuate. Now there are numbers, there there there are weights and biases in a way. That's really cool to me because you you remove the human bias, the human perspective from that, and you and it's just math and and on the underlying models. That's a really cool concept. I'm not sure we've had too much of that ever in science per se. Maybe much more in theory where you can kind of have this optionality and do different type of simulations, but certainly limited and experimental. And I think that's a really, really cool perspective to bring into it from purely the scientific perspective, not deep in the machine learning weeds and what does it mean and how does it impact models from a scientist perspective of what is the output of those models, What does interpretability give me? I just think that's such a cool concept that is very infant in its exploration, and we believe, as Jorge said, will be a really important field for science at large over the next 2 to 2 to 5 years.

Nathan Labenz (1:37:16) Have there been any, like, move 30 sevens in the material science domain and or, you know, the alternative being, like, there's a lot of just, like, doing kind of grinding work a lot faster and, you know, at a at a scale that just couldn't happen otherwise. I guess, have there been would you be confident if 1 like, would you immediately know it if there was 1? And if there haven't been any yet, you know, is that a sort of 2 to 5 year timeline in your mind as you just scale out data? Eventually, you think you'll get there? Like, what what's the the outlook from material science move 37?

Jorge Colindres (1:37:52) Great question. I think we're moving towards that. I'm not sure if there was a earth shattering 1 yet. I mean, like, like, an R type superconductor. Like, just use an example we talked about earlier where everyone in the world is gonna realize and then feel the direct result of that discovery that comes out of a lab. I I can't say that at 30 move 37 there yet. There probably are examples of materials that are being optimized, that are being discovered in the computational realm, in the AI realm. I think it's very lacking still on the on the experimental side, hence why we think our flywheel is imperative in doing materials discovery. But I think 1 cool thing, we always tell the AlphaGo story. We think it's a really good story about indexing information. How do you bring that to science? You know, it it had learned Go games. Right? I forget the like, millions of Go games or whatever the exact number was and made this 1 in 10,000 move. And then the subsequent versions of that AI didn't need to learn Go games. Right? It actually made similar moves without having done millions of of of learnings in Go. I'm excited for us to get to that place as well. And I think we are building the technology to do the first version of AlphaGo today. That infamous move of 37, everyone watching, AI made a mistake. It was a 1 in 10,000 move. We didn't see it coming. I think we'll have something like that. I think Radical AI will have something like that in the next 12 to 24 months if if we follow and hit our objectives in our road map. I think once those become par for the course, become more normal, then moving into the we didn't look at experimental data or we used experimental data from alloys and start making discovery in room temperature superconductors, that's that's what we think is gonna happen. We're gonna amass the largest experimental dataset ever built in human history. That's what we think. And so if we do that, we are very, very excited about the learnings that come from that And then the other areas that we can make move 30 sevens in, I don't need millions of trained examples exactly in that specific material space. To us, that is the holy grail of material science. That is removing materials as the biggest bottleneck to our most important industries. That is enabling a world that is not limited by material, but only limited by imagination and the laws of physics. That's what we wanna build, and we are so excited to watch that proliferate and and candidly, very excited to be leading the charge on the intersection of AI in discovery and fully robotic self driving labs to supercharge and inform that discovery. We think that flywheel together will really craft a world that, as we always say to the team, you can't imagine today the amount of Ubers and Airbnbs and Facebooks and even things like Nvidias that come from this world that we can enable. We think it'll be 1 of the most important companies to ever exist because of that, and that's the world that we're excited to live in.

Nathan Labenz (1:40:57) In many podcasts, I would say that's a great place to end it. But in this 1, I have a couple more questions because that is an inspiring vision, and we can link to the careers page in the in the show notes if you'd like. How about the future of the integration of modalities? This is kind of a hobby horse question of mine. But, you know, I do, to be very real about it, lower impact work. But, you I know, started this company way, Mark. We help small businesses create video content for local advertising. We had, like, basically, a pre AI version of this, and then we've tried to infuse AI into everything that we do. And 1 of the biggest things that I've observed is that the integration of language and visual understanding has been a real step change when you no longer had to go through a language bottleneck to say, okay, like, I have this image. I wanna change it to this other image or I just want an image, you know, but but if I have to say all that in a 100 word prompt, you know, if I train a language model, which I've done to, you know, prompt effectively, that's just such a lossy thing. Right? And what has made some of the more recent models just go viral and just amaze people so much more is that they clearly have this joint understanding where now I can turn my image into, you know, anime whatever, but it still retains the meanness of it while layering on this other kind of coat of paint, whatever. Right? But it that that depth of understanding where it it knows what you're asking for in language, but it also really clearly gets the image itself, you know, at a pixel level, that has seemingly created a a major step change. What I understand today about the system that you guys have built is, like, there's a lot of models that are doing a lot of kind of sub parts. Do you think that this kind of goes the Tesla route eventually or the Gemini 2.5 flash image out or whatever route where it's like there's 1 joint embedding space and kind of language and a sort of intuitive physics to potentially, like, even graphed physics of materials all exist together, or is that, like, too fanciful to expect at this point?

Joseph Krause (1:43:10) I don't think it's fanciful at all. It's it's probably going too far to say that we'll only ever need that 1 model to do everything. But, I mean, internally, we are developing something like what you described. And so it will be natively multimodal nature. We'll have encoders for all of the different modalities that that we wanna work with, and we will run it through this model. And we'll be able to have things output from it that are quite useful to us. So, no, I don't I don't think it's fanciful at all to say that we'll have something like that. In fact, I think, you know, in the next 6 months, we'll be using that internally here at Radical AI. And certainly within the next 6 to 12 months, we will likely publish and release something along those lines.

Nathan Labenz (1:43:53) The interpretability of that is gonna start to get super interesting.

Joseph Krause (1:43:56) That is 1 of Eric's prized things. We have been talking about it for a while if we could work on something something like that.

Nathan Labenz (1:44:02) Yeah. Cool. And that sounds like it's not super far off. How do you think about the IP associated with this? I mean, you're gonna accumulate the world's biggest experimental dataset, as you said, and then I don't know if you meant, like, open source necessarily the model that you're creating there when you talk about publish it, but and I also just don't know a lot about, like, the IP that exists in the materials science space. So how do you guys think about what data and what value, you know, you wanna kind of keep internal to the company versus what you can share? Again, I don't even know if you can you is it possible to, like, patent, you know, an alloy? Is it especially something amorphous that would you be patenting the process that creates that? I mean, you can't really patent something that's amorphous, right, and has no, like, repeating structure. Would think that would be hard, maybe the process to create it. So, yeah, I guess, what is the sort of outlook for what is gonna be core IP not shared and and what can be shared? How do we get the how do we diffuse the benefits of this while still giving you guys the ROI needed to, like, scale out the the autonomous systems to actually scale up the the flywheel, the dataset, and eventually get those move 30 sevens?

Jorge Colindres (1:45:15) Yeah. I mean, it's an awesome question. And when we started the company, we spent a lot of time thinking, okay. What what do we wanna build? And not just what do we wanna build, what is needed if we are trying to solve the material science problem? Not what's the easy low hanging fruit, not what's a business model that we know VCs will get behind. What is needed to build a company that does the vision you've heard us articulate today. And what we netted out was, well, if you just sell software, you're never gonna make raw materials. And so that can't be it, so we're not gonna do that. And there's been a bunch of companies who have tried and failed doing that. If you try to license every material you make, that can be a business model. But the problem is then you need to be creating new discoveries multiple a year for decades to keep up with and continue to drive revenue because, of course, reverse engineering and expiration of patents becomes a serious problem when you're just licensing. And so then we looked at what the largest material companies do in the world. What does 3 m, Dow, Applied Materials, BASF, Materion, etcetera? What do they do? They sell materials at scale. That's what their business models are. Except they're doing it in areas that, frankly, to be completely blunt, don't impact society. Look. I understand the need to have a new Sharon Williams pink color every year. My wife is a massive fan. I I know. I'm a customer. But that's not why we're going radically. Honestly, they can continue to handle that. We are trying to make the species interplanetary multi planetary. We're trying to drive lossless nuclear energy. We're trying to rethink transportation with floating bullet trains. Like, we're trying to create a world that we want to live in. And so if all of that is true, then not only do need to sell materials at scale, but you need to push the materials that you're selling at scale into into enabling the industries and creating the industries of the future. And so when we do that, our focus is on the IP of making it. And in material science, yes, you get patents on composition, but it's really the IP is really around the trade secret on how do you make it at scale and for the lowest cost at scale. It's all we call it processing in material sense. All in that process. It's where all of the real meat of it is. You know, if I tell you a high entropy that we have today, like the exact composition, you're not really gonna go be able to make it at scale per se. Right? And so this is really important piece of this. And so if that is the business model, then that changes the way we think about the AI models. And I'll let Jorge, you know, articulate that specifically.

Joseph Krause (1:48:00) No. I think Joseph hit it on the head. I mean, the the machine learning models will change. They will. They're gonna get better and better over time. We're gonna create some of them. Other people will create some of them as well. We may release some of them. Some of those we probably won't release, to be entirely honest with you. But the 1 thing that Radical AI will never not do is go after the hardest problems that we think are gonna be needle movers for the world. At the end of the day, that's the whole point of the company, and so that will stay core to us through and through.

Nathan Labenz (1:48:27) Love it. You wanna talk a little bit about the I know we're just about at time here, and I really appreciate all the time you've shared with us. Just yesterday, you put on a a news item about working with the US Air Force. You wanna tell a little bit about that?

Jorge Colindres (1:48:40) Yeah. Super. Very, very excited about this. It is in high entropy alloys for hypersonic applications. We are conducting a lot of really hot not just high throughput experimentation, but pushing into how to test and optimize these materials for hypersonic applications. Materials, again, are 1 of the biggest bottlenecks in the most important industries. And in the hypersonic industry, China, Russia, both field hypersonic systems today, they do that because they never stop investing in material science. Actually, it's worth mentioning China, when they make a new discovery in the material, will set up an entire manufacturing hub around said material just to figure out how to scale it. It is an unbelievable emphasis on putting materials at the bedrock of all their most important innovations, And we have struggled with that in The US and other places, and we are trying to drive that nexus back through. And so we spent a lot of time in DC, a lot of time with the office of science technology policy, the National Science Foundation, Department of Defense, Department of Energy, even going to the hill and telling people in government, look. Public private partnership and a focus on materials is imperative. Whether you care about hypersonics, you care about nuclear fusion, or you care about the datasets that are gonna go inside the ML models to generate a bunch of other new materials. That needs to be driven from a government perspective, and you need to a lot rely on private enterprise to do exactly that, which is where we fit in. The direct to phase 2 is, in our opinion, kind of our belief in trying to push this forward. And separately to that whole plan I just mentioned, really an ability to execute on the first material system that we've chosen. You know, we had a thesis that high entropy alloys can be very impactful for hypersonic technology. They're very hard to discover. It's a very large search space. There aren't a lot of experimental results, and our work with the air force is directly around trying to solve that problem and really generate these new materials for them to then go look at exploring putting into hypersonic systems. So that is what we've been working on, and you can find a lot more information in our press release and and more information on that. But we're really excited about that first project, which is, again, specific to the air force and hypersonics, but not just stopping there. We have a lot of other things in the works with the US government that we are trying to push forward so that we can identify, look, materials are not a subsector of industry or innovation. Materials are at the cornerstone of innovation, whether it's the stone age to the bronze age to the silicon age. Every modern technological wave has come from novel material advancement, and the future will be exactly the same. And that's kind of the perspective that we're trying to push inside government and inside policy and inside why materials are really important as a nation and really as a society, as the human race.

Nathan Labenz (1:51:31) That's great. That could be the the right place to end on. Anything we didn't touch on that you would wanna make sure people know about before we go?

Jorge Colindres (1:51:39) 1 thing we call out so we are really intense about our culture at Radical AI because we're trying to solve some of the hardest problems in the world. And so you have to have people that have really a ridiculous curiosity that this question is ability to go to first principle and ask why, but are also unbelievably relentless. We fail every single day, and we will continue to fail purposefully every single day for the next decade and and beyond that. And that's actually where the unlocks come from. It's kind of building on those failures, of course, to reach success as as cliche it is. So if there are people out there listening who really haven't, you know, an appetite to challenge everything down to to to the laws of physics and wanna do that in a way that is trying to tackle our most important problems, recognizing that there's gonna be a lot of failures along the way. But in the chance that you succeed, you leave a fundamental stamp on the human race of impact and innovation, then please reach out. We are actively recruiting right now, and we are looking for not only the best people in what they do, but people that are willing to dedicate their life's work to ambition. If you're looking for a job, this is not the place for you. You should not apply to Radical AI. You will not fit in here. If you are looking to work on a mission, then please reach out. We'd love to talk to you. And Jorge, anything you wanna add on top

Joseph Krause (1:53:09) of No. I think that was perfectly well said. Nothing to add.

Nathan Labenz (1:53:13) Amazing. Well, this has been an excellent conversation. I've learned a lot from it. I think people are gonna find it fascinating. So, obviously, wish you guys nothing but success in terms of breakthrough after breakthrough, and look forward to changing material reality around us as we unlock 1 thing after another. Joseph Krausz and Jorge Calendras, cofounders of Radical AI, thank you for being part of the Cognitive Revolution.

Joseph Krause (1:53:35) Thank you. Thanks for your

Nathan Labenz (1:53:37) If you're finding value in the show, we'd appreciate it if you'd take a moment to share it with friends, post online, write a review on Apple Podcasts or Spotify, or just leave us a comment on YouTube. Of course, we always welcome your feedback, guest and topic suggestions, and sponsorship inquiries either via our website, cognitiverevolution.ai, or by DMing me on your favorite social network. The Cognitive Revolution is part of the Turpentine Network, a network of podcasts where experts talk technology, business, economics, geopolitics, culture, and more, which is now a part of a 16. We're produced by AI podcasting. If you're looking for podcast production help for everything from the moment you stop recording to the moment your audience starts listening, check them out and see my endorsement at aipodcast.ing. And finally, I encourage you to take a moment to check out our new and improved show notes, which were created automatically by Notion's AI meeting notes. AI meeting notes captures every detail and breaks down complex concepts so no idea gets lost. And because AI meeting notes lives right in Notion, everything you capture, whether that's meetings, podcasts, interviews, or conversations, lives exactly where you plan, build, and get things done. No switching, no slowdown. Check out Notion's AI meeting notes if you want perfect notes that write themselves. And head to the link in our show notes to try Notion's AI meeting notes free for 30 days.

Great! You’ve successfully signed up.

Welcome back! You've successfully signed in.

You've successfully subscribed to The Cognitive Revolution.

Success! Check your email for magic link to sign-in.

Success! Your billing info has been updated.

Your billing was not updated.