All right. Hey, everybody. How are you doing? My name is Greg. I lead student outreach here at Anthropic, and I am so excited to be sharing the stage with some of the brightest young minds in AI. Just a little context for this panel. So, at Anthropic, we've given out API credits to thousands of students to help them build things at school, and so what you're about to see is a very small glimpse at what students have been creating with those API credits.
It's a very wide variety of things, as you're about to notice. Some of these projects are very humorous and funny. Some of these projects are very serious and important. Some of these students are working on one project, and some of these students have been building an app every single week throughout all of 2025.
So, I think if I was gonna sum up what I've learned from running this program, it's that the future is really, really bright in the hands of these students. So, without further ado, I'm gonna invite up our first speaker, Isabel from Stanford. All right. Thank you for having me.
It's a privilege to be here. My name is Isabel. I'm a senior at Stanford, where I study aeronautics and astronautics, and I'm doing my honors in international security. And today, I'm here to talk to you about my honors work, which is on finding nuclear weapons in outer space, and how I use Claude to help me do it.
So, for those of you that may not know, Article 4 of the Outer Space Treaty bans the placement of nuclear weapons in outer space. Now, other arms control agreements that you may have heard of, like START and New START, include provisions for verification and monitoring. So, nations are shown to be compliant with their treaty obligations using inspection systems.
We have on-site inspections where inspectors will go and look at each other's delivery vehicles and inspect for the presence of nuclear warheads. We don't have anything like that for outer space. mostly because we signed the outer space treaty in 1967, and there were no technologies to do that kind of inspection, right?
How would you go about approaching a satellite in orbit that might be carrying a nuclear weapon and inspecting it for the presence of such a device? Daunting for the 1960s, daunting today. And this became a problem recently, in 2024, in April of last year. The Biden administration announced that the United States assesses that Russia is developing a space vehicle that carries a nuclear weapon.
Now, this was pretty destabilizing for the international community. We've had a lot of dispute in the UN Security Council recently about how to handle this potential violation of the Outer Space Treaty. Given that we don't have a verification mechanism for compliance with the Outer Space Treaty, I started to wonder if it would be possible to implement such a system.
Particularly given that the US Space Force tracks 44,800 space objects today, how would you begin to know which one of those is the suspected nuclear weapon? So this brings me to my research question. Is it feasible to perform an in-space inspection mission where you inspect a target satellite for the presence of a nuclear warhead on board?
Daunting question. It has a lot of interesting technical and political facets to it, but for one particular aspect of it, I was able to use Claude to my advantage. So I looked at specifically the feasibility of detecting the nuclear weapon with an x-ray system. So you fly an x-ray source and detector on two different inspector satellites in space, have them rendezvous with the suspected nuclear warhead target, and scan it for the presence of a nuclear weapon and board.
I wanted to know if this would ever be possible. No one's ever tried using x-rays in space. There are interesting questions around whether the space background environment is - there's too much noise in space to detect the source signal. So I built a computational simulation to see if this would ever be possible.
And to do it, I used Claude. I used this very complicated CERN software package called giant4. I am not a particle physicist. I did not know how to approach this software package and write the C++ code. But I was able to make a desktop application to do my simulation using Claude.
And it was incredibly exciting. It worked. So what you're seeing in this picture is like a very, very quick snapshot of an x-ray image taken in space. And you see a little hole in the middle that shows you that there's very, very dense fissile material on board the target of the scan.
So indeed, in this simulation, there was a simulated nuclear warhead on board the satellite target. The outcomes for this are pretty significant and interesting. There are a lot of people in the national security intelligence community in this country that are interested in developing this kind of capability to inspect adversary spacecraft on orbit.
To understand their capabilities, particularly whether they might carry a weapon of mass destruction. So having done this research, I actually am going to be able to brief it in Washington, D.C. to some policy makers at the Pentagon and state. I'm really thrilled about that opportunity. And certainly, the desktop application with this level of fidelity would not have been possible without modern AI tools to make this kind of research accessible to an undergrad in less than a year.
My takeaways for you, kind of as a student, doing research in the era of AI is just that primarily, there is no learning curve that is too steep any longer, right? Even the toughest problems, space technology is notoriously hard, nuclear weapons, existential threats. We can address these critical crises with the tools that we have today with emerging technology.
And so I want to challenge all of the minds here and other students to think about what are the world's toughest problems? What are the problems that you thought were unaddressable, that feel like existential crises to you for the next generation? Those are the ones that we should be using our brand new, shiny, exciting AI assistants to work on, because that's how we're going to help make the world safer and more secure, or at least outer space, more secure.
So thank you. I'm going to pass it off to the next presenter now, but if you have any questions, I'd love to talk after the presentation. Okay, so it's kind of tough to follow up finding nuclear objects in space. So I'm going to tell you about how I did not know the difference between the terminal and the code editor, and why Claude is the reason why I was able to learn how to code.
I'm a student at UC Berkeley, my name is Mason Ardidi, and I'll go ahead and get started. So I want to talk about what we think of as the traditional way to approach learning how to code. I'm going to call this the bottom-up way, where we start by taking basic classes, learn our basic skills, and then build apps with those skills.
Slowly but surely, we level up our skill set and build apps that are more complicated. I learned a little bit differently. I'm going to call this the top-down approach, where I had an idea as I get inspired randomly, and I had no idea how to solve it. It was software I've never coded before.
So I tried to have AI make it for me. Hey, make this app for me. And then when it inevitably fails, I learn how to do it myself, slowly but surely learning through different layers of abstraction until I actually understand what's going on. Now, where did this leave me seven months ago?
It left me not knowing what the difference between the terminal and the code editor was. I put npx create next app latest in my page file. I had no idea what I was doing. But slowly but surely, I asked, why is this happening? What am I doing wrong? And I was able to learn more complicated skills.
Let me show you a demo of something I'm capable of doing now. Okay, welcome to CalGBT, which is a better way to schedule your Cal courses using AI. It's going to question like, show me math classes with a high average grade, since I want to be lazy in my math and get an easy curve.
Here, it's going to show us five different classes that have an average of A or more, And in fact, it even showed us classes with a grade point average of 4.0. Can't really get much better than that. Now, let's say it's getting late in the enrollment cycle. And I want to see classes that still have open seats.
Show me history classes that still have open seats. And this is drawing directly from Berkeley time. So it's live data. And here it is, it's showing you history classes, five seats, 20 seats, or three seats. We can even ask questions that are more deep, like, what is the meaning of life?
And do with that answer as you will. But this is CalGBT. My name is Mason. And enjoy your day. I'll show you another one, which I developed at the Pear and Anthropic hackathon as well. Okay, welcome to Get Ready, which is a new way to visualize and understand new code bases.
Let's take a look at Anthropic's SDK for TypeScript, for example. You'll see soon. And we'll be able to interact with the chart and see how all of these files interact with each other. So here we have a mapping of some of the most important files. We chose not to display all of them, just the most important ones that the users will interact with the most.
And we have these lines to show how they're interconnected. And we do this through the function calls that are actually, like, in each file. So, like, if this demo TypeScript file is referencing the batch results, that's where the line comes in. And then over here, we have just a quick description on what the file actually does.
And we have our comments on the code base. Okay, and on top of these two, I built many projects over the course of my learning how to code. Now, what is the point of me showing you all of this? I'm not here to brag. I'm here to say that Claude is the reason why I was able to learn how to code.
Without Claude, without these AI tools, including cursor, windsurf, whatever you guys want to use, none of this would have been possible. And the key takeaway for me is that you can build anything you want nowadays. You just have to ask the right questions, learn through the different layers of abstraction.
I think this is representative of a new style of building and a new class of builders. Where my flow, personally, is I find a problem that I'm inspired by and want to fix. I realize the solution is something that I have no idea how to do. And then I have a high level chat with Claude, execute steps in the actual editor, and then record and post a demo when it's not perfect, hopefully bringing users and revenue later on.
But this iteration cycle, instead of taking years for an undergraduate degree or doing other things, can be one day to one week maximum if you really want to. So I'll keep it short and sweet and leave you guys with a couple of things to think about, which are on my mind right now.
Which is, how can we build to get users and revenue? Not for technical perfection and impressiveness. How can I build things as fast and as simply as possible? As demonstrated by this prompt, give it to me in the simplest and most concise way possible. What ideas actually inspire you?
And how can we build it today? And lastly, not on the slide, but what does it mean to really know how to code? Does it mean understanding every single line and every single function? Or does it mean being able to build something that actually improves people's lives? I'm going to continue to post more information.
If you want to connect with me, you can scan this QR code. But my name is Mason. Thank you guys. All right. What is up everyone? How are we all doing? We good? Yeah. My name is Rohil. I'm a freshman, or just finished freshman year at UC Berkeley in the MET program, studying EKS and business.
So CS and business. And I'm here to talk to you guys today about SideQuest, which is a project that a couple of friends and I made at the Pear Exanthropic Hackathon recently. So let me tell you guys about a big problem today, is AI embodiment. So we see like in Hacker News and the newest news all around that we're trying to create bots that interact with our world.
And most recently, we have seen these robot dogs that are able to deliver you a water cup or something like that. But these systems do not compete with humans ourselves. Humans are like built to interact with our world. And that brings me to here, which is that today we have humans hiring AI agents to do their work for them.
I'm sure all of you guys have probably employed some sort of AI agent to do your work for you. But today with SideQuest, we are flipping the script, we are flipping the script. And we have AI agents hiring humans to do their work for them. So AI agents obviously are amazing at interacting with the digital world.
And humans are amazing at interacting with the physical world. So why can't these AI agents just hire the humans? So that brings me to the architecture of SideQuest, which is basically, like, let me give you a hypothetical example. Let's say an AI agent is trying to host a hackathon.
So now they have all the logistics covered, but they need to put some advertising material up. They need some flyers up so that people can find out where this hackathon is, where to go. But they don't have any physical means to do that. So what they do is that they ping the nearest human to that area and tell them, "Oh, pick up this flyer, put it in this location, and livestream that video to me.
And as soon as I can see that you did it, then I'll give you money." So that's exactly what's happening in SideQuest, and I'll show you a short demo. Hello, world. My AI friends and I are hosting a hackathon. Let's check if the flyers are up. So we see a flyer here.
Flyer detected. But we do not see a flyer here. No flyer detected. Bruh, I need a human to put on some flyers in room two. Let's do this. It looks like there's a quest. So I have to collect three posters from table eight. Let's do it. So over here, there's a live video stream that Claude is actively looking at and verifying whether you're doing a new task.
I found table eight. Let's see the posters. Boom. Scanned. It says I have to set them up in Strong Yes. We're here at Strong Yes. Now, let's set up the poster. And perfect. I think that should be good. Let's scan it. Booyah. I made a hundred bucks. Let's go.
And boom, we're done. We're ready for the hackathon. Yep. And that's SideQuest. So let me talk a little bit about what I learned with building with Claude. Is that first, Claude is really smart, as with like any of these AI systems these days. And they can reason through many messy edge cases.
So we as humans, we don't need to prompt every little nitty-gritty thing. We can start thinking about bigger picture parts of building products. Secondly, we should design like with a back and forth workflow with these AI systems. Like originally, we are thinking like upfront, oh, how should I build this whole big thing?
But that's a really big task. You can break it down. Ask Claude, oh, like what are the different things that I need to do to to work on something? And let's build this step by step. So with this iterative process, you can build like very robust systems. So bottom line is that you should trust AI and trust Claude that they aren't things that you have to micromanage.
They can think on their own as well. And now some takeaways for builders to be like like this cool guy, not this grumpy guy, is that you should think of AI as a system rather than just a feature builder. That this is someone that you can like talk to, reason with.
And secondly, as like thinking bigger picture about us as humans, is that we should be system designers first, or architects of like the things that we're building. Because in the future, we aren't going to be the ones writing the small code. We'll be the ones dictating what code to write.
So that brings me to the end. Thank you guys so much. Have a great day. Bye bye. All right. Good afternoon, everyone. I'm Daniel. I study computer science at USC, and I've also built projects across Amazon, IBM, and various startups. Yeah, very honored to be here as a student speaker today.
For more context, I help USC lead some of the entrepreneurship programs. And over the past year, Claude has been integral to many of our projects, powering innovative solutions across various domains. When Anthropic announced the hackathon at USC, a lot of the students, including my teammates Vishnu, Shabayan, and myself, were naturally very eager to join in and explore new directions with Claude.
Today, I'm honored to share our journey and insights with you. So let's first start by looking at the problem. Current LLMs are great at giving answers, but when decisions really matter, one general response just isn't enough most of the time. Whether it's business, healthcare, or policy, high-stakes decisions require diverse input and deep analysis.
Today, getting those perspectives means prompting an LLM multiple times, which could be slow, inconsistent, and very manual. Knowing that Claude excels at complex reasoning as one of its most impressive capabilities, that's the gap that we aim to solve for our hackathon. Introducing Claude Cortex, a system designed to emulate a panel of experts, each analyzing the problem from a different angle.
It dynamically creates specialized agents tailored to your problem context and enables parallel processing for diverse insights. The output here is a more synthesized and well-rounded recommendation, enhancing output quality for decision making. It's basically like having your own strategy team for each prompt. So yeah, let me show you how it works with a really simple example to test out the agents.
So let's say I want to learn how to use Lengraph specifically by researching its documentation. I also want to share that finding with my teammates. I would type that in as a single prompt and let the master agent interpret that request and spin it different agents, which in this case will need a browser agent to search and extract relevant information from Lengraph's documentation, a research agent to summarize the key concepts in plain language, as well as a notes agent to generate clear explanations, which it then shares with my teammates automatically.
Each agent will work independently, but they can communicate with one another, creating a multi-agent system that gives more comprehensive insights. Now for sectors where data security and compliance are paramount, Claude Cortex offers a secured mode by integrating with AWS Bedrock. It ensures that all operations meet privacy standards, making it ideal for sensitive environments.
The rest of our architecture is also very straightforward. The front end was built with Next and Tailwind. The back end leverages Fast API and Lengraph for orchestrating multi-agent workflows. And Claude, of course, powers our agents reasoning with the addition of browser use, which allows agents to fetch real-time web data and enhance their analytical capabilities.
Claude Cortex represents a shift in the way we use language models, moving away from simply generating responses to structuring parallel reasoning pathways and delivering more comprehensive insights. It's versatile, making it valuable across various sectors, from corporate strategy to public health safety. Now, the key takeaways from building Claude Cortex are very intuitive, but the main two points here that I want to emphasize are that when agent outputs were more focused and well-structured, like JSON format, Claude's synthesis became more nuanced and high quality.
It struggled, however, when upstream agents were more vague and just dumped text blobs into the stream. And then dynamic task creation allows for flexibility. What that means is we first started off by creating five predefined agents for every scenario. However, we later realized that having a master agent to decide what tasks and agents to create allowed for more accurate and relevant information.
What we're building with Claude Cortex sits at a broader trend. Claude is powering a large number of student-led products at USC. We've seen tools for lawyers to process case files faster, apps that help people retain and connect knowledge more effectively, and software that can automate documentation and progress updates.
Claude's ability to read deeply, summarize clearly, and follow structure is what makes all of this possible. Looking ahead, as a student building with Claude, the most powerful applications I've seen aren't just asking Claude for answers. They're using it as infrastructure, something that you can wire into your workflows and something that you can orchestrate like a system.
And that's the shift that we see as well. We imagine agents that can collaborate with one another, tools that can reflect, and context that can compound. In summary, Claude Cortex isn't another AI tool. It's a leap towards a more intelligent, secure, and multi-dimensional decision-making process. As we continue to refine and expand its capabilities, we invite you to explore its potential and join us in shaping the future of AI-driven solutions.
Here's the team behind Claude Cortex. We're all student builders and we're all, yeah, builders and student leaders at USC, and we would love to discuss more, so please feel free to reach out to us whenever. I'm Daniel Gao, and it's been a pleasure sharing our work with you. Thank you for your time and attention today.
Thank you. Thank you.