I know it's tough being after the break and after what was a nice conversation, fireside chat. And I was talking with Harrison backstage on how much of these beliefs that he has for the future we have been implementing and the plan is to show with some of you today.
So Cisco is a company that many of you may know that there is a lot of networking stuff and collaboration and security. We are over 40 year company and there is customer experience inside Cisco. Customer experience for us go on a wide range of services that we provide. I'll share with you a little bit on the first topic, then deep dive on what the use cases and how we're using technology together with LangChain on the partnership to accomplish the results that we're getting, and then share with you some key learners and takeaways along the way.
Some of the good scarves, some of the learners don't do this, do that kind of thing, okay? So fourth thing so we can get all the way so you can all understand what I'm talking about. Customer experience at Cisco. Cisco is a $56 billion plus company with half of the recurring revenue of the company.
So we have more than $26 billion that runs on this. So talk with LangChain like, hey, you're not joking here. So there is a lot of implications for that and everything on customer experience is meant to maximize the value that the customer gets on investment they do on Cisco products or services or technologies.
So we have the typical land, adopt, expand and renew framework that everybody has on the industry to go there. That is the process that we go through. But when you talk about any company, there is always process, people and technology. In those kind of conferences, a lot of people think about the technology.
There is a lot of hype but they always come together with the teams that implement that and the process that goes with it. So I meant to start with this. So we have land, adopt, expand and renews and actually there are people and organizations that goes for that. So we have a customer success organization that actually is in charge or when someone acquires something from Cisco, buy a product.
There is an adoption team that makes sure that you adopt whatever you buy for, right? And there is an organization that is renews to make sure that when the end of the term comes, you are happy enough to actually renew your subscription. And usually more often than not, during that period, we expand our relationship with the customer.
There is obviously technical support, which a lot of you match to this, and other organizations like implementation, service delivery. I'm talking about 20,000 people organization north of this. So there's a lot of things that could be improved. And from that lens, what you're doing is adopting AI to help us, not only on optimizing the process, but optimizing for people and maximize their returns to the business.
And that's a very important thing because if you look on the left hand of this slide, I'm talking about how we make customer experience unimaginable. And by that, I'm going to hyper-personalization, proactive and predictiveness. But we are not talking about agentic AI overall. It doesn't come out of the blue and like, hey, agentic AI is a new thing.
We start with machine learning predictive AI models over a decade ago. So treat very well your data science team because they're going to be critical now. When it comes from the LLMs, Gen AI, which are very good for everything that has to do with interactions on language, there is an L on the middle of this for a reason, LLM is really bad for predictions.
So when you bring this together, we can have multiple agents that go for workflows. So with that said, our vision and the future of Cisco customer experience, and you have been leading on this for over two years now, and a lot of companies are coming to help understand how we're doing this.
And that's part of the reason why in this conference on the very strong partnership we have with and Harrison and his team, we are going for using agentic AI to elevate CX to become an agentic CX by providing personalized, predictive and proactive experience to all of our customers, together, not separate.
Hyper-personalization, how we can predict fails before they happen, how it can be proactive of field notice, or even best practice that goes there. We are leveraging multi-agents. By multi-agents, you're going to see on the next slide what I mean by that. There's human and machine type of agents, Gen AI, and traditional AM.
We are providing services that goes for users that call them, like you can call on a video call, on a chat interface, or a phone call, or to have a tool. And the agentic CX provides a lot of embedded value, like advanced technical support, predictive intelligence operations, and all of this is meant to help customers to have recommendations that are proactive.
So, we want to avoid them to face an issue, if we can, upfront. We want to give them predictive insights, and make this hyper-personalized foresight on a sense that every customer should have an agent on themselves, which has context for that. So, context is a very important thing for us, and we are going beyond what we call about MCP context.
It's a big thing that is out there. With that said, let me go, now that you're all PhD, what we do at Cisco, let me talk a little bit about use cases, and why we are drive from a use cases approach, as opposed to a tools approach. So, just sharing an experience for here, when you start about a year and a half ago, obviously, GPT was the big name in town, and everybody was trying to try, GPT 3.0 was really bad back then, but it still was novel, right, completely novel.
But, before we have everybody trying to do a chatbot for whatever reason there was, so the first thing is, before we even go to the use cases, we define the criteria that would make the use case belong in the first place. Because, we have an advisory part of our services, we enter a customer that the customer had 412 use cases for AI, and when you talk with them, it ends up being five that actually collaborate to the business.
So, boy, this is less than 10%. So, we define that criteria to say, any use case that we do on customer experience, remember, it's 20,000 people organization, there's a lot of creative people and a lot of ideas. Like, it must fit one of those three buckets. We must have use cases that help customers get immediate value and maximize what they invest in us, and that's where renews and adoption as an agent goes in.
Same applies on how you make the operations of these people more secure and reliable. That's where everything with support would go in, and then the whole correlation and agentic workflows provide visibility and insights across the whole life cycle. So, there is a method to the madness here, if you think, and if you leave it alone, people are going to do their own thing, and it's going to come to you and say, "Hey, how cool it is." So, how this manifests to the business.
So, define the use case criteria first, then put the use case on top of this, and that's how we've been structured to work with LinkChain, because now we stitch the pieces. Because the agents, developers, and multi-organizations make sense to the customer in the end. So, with that said, we obviously have a high-level stack.
I can go into details. We have the team on the booth that goes deeper, including on what we demoed yesterday. So, we start with the need, for our use case, to have flexible deployment models. What I mean by that, we have customers, like federal customers, health care, and some others, that require on-premises deployment.
By on-premises, I'm not talking a VPC on AWS. I'm talking a physical data center with the devices. We have clouds, and we have hybrids on it. So, we need to choose some criteria, like security, compliance. There are customers in Europe that have heavy regulations. There are some others that are more B2C.
The bulk of our business is B2B, business-to-business, not only business-to-consumer. So, when you do for that, we start to power the best-in-class AI technology. So, we chose Mistral-Large, and we worked very closely with Mistral, even to the development of their models, to run on-premises. And we have both Sonnet, the latest from Sonnet, 3.7, and shartgpt from 4.1 all the way to 0.3 for some of the use cases.
That is empowered all together for Langchain. In the demo, you're going to see that you have the same agentic framework running on-premises on a data center, and 100% on the cloud, without any change. So, what he's saying on the belief, we've been doing this in production for a long time at scale.
So, what Harrison said is not high level, it actually works. It's an interesting thing, and we used the agentics on multiple agents on the top, and we did end up doing custom AI models. By custom AI models, I'm talking about both creating machine learning models that are for predictions, that we train for the signals, and fine-tuning LLMs, especially the on-prems, to accomplish high accuracy in some of our use cases.
With that said, here I want to do one slide before I deep dive on some of the tech aspects. So, remember that I mentioned to you the process, land, adopt, expand, and renew, and the people. So, now I'm going to land the technology on top of that. So, if you look at the use cases, we have shipping deployed in production at scale for over six months.
Renews agent that applies to the renews team with predictive insights. This is LLM combination with machine learning. We have support with the virtual tech engineers that augment our support people to actually go, what's the next best thing? Automate resolution of the low priority case without human touch at scale.
I'm talking about 1.6 to 1.8 million cases a year. 60% of that is fully automated. And beyond that, integrating this directly on the product and sentiment analysis across the whole life cycle. Which is an important thing. But I want to highlight with you, we have stuff that is deployed.
And Harrison is absolutely right. Experimentation and production are two different beasts. But we also talk about limited availability. So, we interact with the end user. We have the subject matter experts and the cohort to work with us. So, I understand what questions they're going to ask. Because the renews people is going to ask renews related questions.
The adoption people, adoption related questions. It sounds obvious, but a lot of people don't think this way. They interact with the customer after they develop the cool stuff. We shouldn't do that. You should go to them before. And say, "Hi, what do you need?" And then you build the AI to help them.
So, but at the same time, even though we have limited availability. And extending with the length chain for the supervisor approach that we're going to show. We also have a lot that is on experimentation. Because we learn to build pipelines for the new use cases that are coming. So, those three things run in parallel.
Ideally, on three different teams. And I'll touch upon this soon. So, let me take one of those as an example on the renews system. The business value that we start this is why do I need to build the renews agent? Well, over half of my business is recurring revenue.
And I have a lot of time that these people is wasting just trying to chase dashboards and tools. So, anything that I can give back and remove operational burden for them means and translates directly correlation with the financial results. Less time spending on doing useless stuff means more renews that they're chasing, means more business that go on.
Not going untouched means higher results. There is a direct correlation to the business. Not hard to justify the investment. You get it. Right? But at the same time, we wanted to correlate this with real-time sentiment to the customers and provide summarization that are hyper-personalized to the person. So, you are renews on financial services industry.
So, I only explain what's the new trends for financial services, which are completely different than healthcare or the government, people that deal with. So, with that said, we had over multiple 50 different data sets that were going around. a zoo of tools, as you can imagine, for every renew event.
And the results for us was actually a reduction on 20% of the time, used in less than three weeks, and limited availability. So, the business impact was immediately, and we have high accuracy of risk recommendation. I was the one that the team loves and hates me at that point, that I said, we're going to go 95% accuracy and higher.
And we accomplished that. We explained this yesterday, and people on the booth can tell you how we got there. So, let me go a little bit on the weeds here. How did you do that? So, let me go a little, explain how the agentic flow works for us. So, here's a question, typical question for the renews person.
What's the upcoming renew status for customer X, Y, Z, and the actions required to minimize its potential risk? It's a very fair question, but if you think about it, I need to know what's the customer, then I need to know what products the customer bought, I need to know what dates these products were bought, so I can get the circle of renews.
I need to understand what the current status is, and I need to map all the signals for the risk. And if the customer has multiple products, he may be happy with the product A and not happy with the product B. And the renews may be compromised or not. So, there is a lot of signals and intelligence that goes from that.
So, having a single agent for this was not ideal, because you don't get to the accuracy level that we are targeted to. So, we needed to go before supervisor was a thing on the Harrison this morning. We came with a supervisor approach, which is basically receives the NLP and decomposes the NLP.
And because the question is about a renews question, it hits the renews agent immediately. Then the renews agent gets the context, and then he calls himself the adoption agent and the delivery agent to understand what's the status of the customer now. So, I can answer half of the question on the pipeline.
And you can see this on the demo, on the Langsmith and all the traces going, and how we decompose the questions there. So, we can learn and leverage some of this. But that's active in parallel behind the scenes. Not 100% autonomous yet, because we still have human in the loop.
But sentiment analysis is something that I can trigger any time. Right? If there is a question, I can proactively go and trigger the sentiment analysis and all the signals and come back. Hey, this customer hates us on that product, he loves us, or something like that. Right? And at the same time, I want to make sure what is the install base that this customer has.
Is a competitive guide into this, because we open a breach, we are too expensive, or we miss a functionality, whatever that is. So, I'm talking about Cisco, but I appreciate the fact that a lot of you have a similar scenario on your products, especially if you have recurrent revenue and a lot of things that goes on.
So, then when you get out of the first part of the question, and you get the real focus on customer experience. Add customer experience on any company is all about workflows. A support ticket follows a workflow. A renewal process follows a workflow. Adoption follows a workflow. So, if you think about this, LLMs are not very good on workflows.
They are good on language. They are not metadata. They are trained on information that goes there. When you do a genetic with a tool like LangGraph platform, it helps a lot with that. So, we went on this, and we go to workflows, and if the second part of the question is about risk, I hit a predicted machine learning model.
Why? Because it's very deterministic. DLLM is very probabilistic. So, we combine it both to get to the accuracy level, and we leverage the length change to make context carrying back and forward, and LLMs to trace it back. Going to the point that we receive the answer, we do the final reasoning formatting, and answer it back to the internal user.
Now, you can go deeper on this, and I don't have time for that, but I want to share you something. This, our CX Agentic AI in action, I have seven agents on that example. Over time, we may decompose them on some other agents. That's okay. The point of Agentic AI, a lot of people think about agents.
For us, it's less about the agent itself. It's more about the flexibility for the workflow. Think about this example. This is a question, how can I maximize the value for what I've invested in Cisco in the last two years? That's a very fair question, but it's meant for an external customer, not an internal user.
Same agents, different workflow. How can I dynamically change this for the understanding? That's how the power of the supervisor and dynamic agents goes from there. So, you can see the agentic power coming in reality, and this is running in production. So, if you have time, I recommend you to go in the booth.
You have Vince and Amand that presented yesterday, myself and others. That's going to show how we're running this on environment in production. How many agents we have. The interaction between the supervisors and the agents. We use multiple models, as I said, deployment on-prem and on the cloud. And a predictive machine learning pipeline using predicted models on ML integrated with LLMs to accomplish the results that we want.
With that said, I want to wrap this conversation, share with you some key learnings and takeaways that we went through that process for us. So, first thing that I would recommend you. As I mentioned before, please define the use cases and the metrics first. Don't jump on the bandwagon because there is a new tool on the weekend because next weekend is going to have a new one.
And your team is going to get excited and AI is moving at unprecedented speed, which is great. It's amazing to be on. It's happening in our lifetime. But at the same time, if you define the use case, it's much better for you to measure it. At the same time, rags, prompts, field shots, supervised fine-tuning chains come after you have the use case.
There is a reason for them to access. I'm stating the obvious, but you wouldn't believe how many times this is not used. Last, on the right side, experimentation writing prototype is key. Sometimes, if you do have a team that's only focused on production, they have different metrics than the experimentation team.
The experimentation team has a latitude and degree of freedom to try and fail and fail fast. So use that and have a dedicated team for evaluation too. I talk to my team and they know that you don't make the dog the custodian of the sausage. It doesn't work like that.
So you want the evaluation team to have the golden data sets and be able to say, "Hey, this stuff is not hitting the performance you need, or the cost you need, or the metrics you need." Because if it's the same team, people blend among themselves. So create this isolation, which helps you to achieve what you want.
Last but not least, achieving high accuracy with stacks to SQL SQL and enterprise use case is really, really hard. The three-letter acronym is called SQL, and another three-letter acronym is called LLM. They don't go on a date. They don't get along. So, boy, believe me, it is hard. So we actually leverage this snowflake context, semantic context on Cortex, just for the reference of metadata.
But then normalize the data first. And if you believe on something that I'm saying, avoid using LLM for doing joints on SQL. It royally sucks. You're going to get there. Avoid hype and inter-Asian context and collaboration is critical. It goes beyond MCP. MCP is great, but MCP needs to be evolved.
It's part of the industry that's starting. It's a Swiss cheese for now. We are working on others. And both Cisco and LLM chain has been championing an initiative that we put out there that's called Agency, which is a full architecture that we open source. The code is there. You can use it if you want.
That goes beyond the only sharing of LLM protocols, MCP being one of those. Could be HOA in all of those. But it's how you leverage a semantic layer and syntactical layer across agents with a directory. When you go to the internet, the first thing that happens is you go to a DNS server, right?
There is no notion of DNS servers with agents yet, if you think about it. So it brings all these notions of how you authenticate. You have an agent directory. We authenticate and make sure that you have this. There are companies that are working there. There are startups that are bringing this.
LLM chain and Cisco with others on that slide. We are proposing a full architecture and open source so the industry will go to a Gentic AI fest. With that said, I would like to thank you all for all the time that you have here. It's amazing too. Amazing conference.
We have our teams on the booth. We are going on the weeds and have the demo running in production. If you go through the traces, brace yourself, and let's go together. Thank you very much. Thank you very much. Thank you very much. Thank you. Thank you. Thank you.