There is no denying that AI is real. Yes, it’s over-hyped. Yes, entire portfolios are being ‘AI-washed’ in the same way everything suddenly became a ‘cloud’ product over a decade ago. But it is real according to our 2024 State of AI Application Strategy research.
While most organisations (69%) are conducting research on technology and use cases, 43% say that they have implemented AI at scale. That’s either generative or predictive.
Somewhat disconcerting is the finding that 47% of those already implementing AI of some kind have no defined strategy for AI. If we’ve learned anything from the rush to public cloud, it should be that jumping in without a strategy is going to cause problems down the road.
To help you define that strategy — especially when trying to understand the operational and security implications — we’ve put together a list of three things you should consider.
- AI Applications are Modern Applications
It doesn’t need to be said, but let’s say it anyway. AI applications are modern applications. While the core of an AI application is the model, there are many other components—inferencing server, data sources, decoders, encoders, etc.— that make up an AI application.
These components are typically deployed as modern applications; that is, they leverage Kubernetes and its constructs for scalability, scheduling, and even security. Because different components have different resource needs — some workloads will benefit from GPU acceleration and others just need plain old CPUs — deployment as a modern application makes the most sense and allows for greater flexibility in ensuring each of the workloads in an AI application is deployed and scaled optimally based on its specific computing needs.
What this means is that AI applications face many of the same challenges as any other modern application. The lessons you’ve learned from scaling and securing existing modern applications will help you do the same for AI applications.
Strategic Takeaway
Leverage existing knowledge and practices for application delivery and security but expand to include approaches that recognise that different components of AI applications may have varying resource needs, such as GPU acceleration for compute-intensive tasks or CPU resources for less compute-intensive workloads. Modern application deployments allow for flexibility in allocating resources based on the specific requirements of each component, optimising for performance and cost-efficiency.
- AI Applications are Different from Modern Applications
Yes, they are modern applications but there are differences between the two that impact architecture, operations, and security.
First, AI applications exchange unstructured data. Those prompts have no format, no length or data type requirements, and the eager adoption of multi-modal LLMs only adds to the chaos that is a ‘request’. In the sense that most AI applications wrap a prompt and response in a JSON payload, I suppose you could say it’s structured, but it’s not because the actual payload is, well, undefined.
Second, AI applications communicate almost exclusively with a model via an API. That means bot detection solutions that use ‘human’ or ‘machine’ as a base criterion for access are not going to be as helpful. Security services helping to weed out ‘bad bots’ from ‘good bots’ are going to be an important part of any AI strategy. The reliance on APIs is also why, in our 2024 State of Application Strategy research, we found that the top security service planned for protecting AI models is API security.
Lastly, interaction patterns for AI applications are often dynamic, variable, and unpredictable. Generally, today’s security services watch for anomalies in mouse click and typing rates per page, because the services can infer ‘bot’ behaviour based on deviations from established human average standards. That doesn’t work when someone is using a conversational interface, and may type, retype, and submit questions on a highly irregular basis. Given that many security solutions today rely on behavioural analysis — including API security — that means some adjustments will be necessary.
Strategic Takeaway
You will need additional security capabilities to properly govern AI applications. Rethink traditional security approaches that may not adequately capture the nuances of conversational interactions. Explore innovative approaches such as real-time monitoring of interaction patterns and adaptive access control mechanisms based on contextual cues. Recognize the critical role of APIs in facilitating communication with AI models. Invest in robust API security solutions to protect against unauthorised access, data breaches, and malicious attacks.
3. Different AI Applications use Different Models
Like the eventual reality that is multi-cloud, it’s highly unlikely organisations will standardise on a single AI model. That’s because different models can be a better fit for certain use cases.
That’s why we are unsurprised to learn that the average enterprise is already using almost three (2.9) distinct models, inclusive of open-source and proprietary models. When we look at the use of models based on use cases, we start to see a pattern. For example, in use cases which rely heavily on sensitive corporate data or ideas — security ops and content creation — we see significant trends toward open-source models. On the other hand, looking at a use case for automation, we see Microsoft gaining use, largely due to its ability to integrate with the tools and processes already in use at many organisations.
This is important to understand because the practices, tools, and technologies needed to deliver and secure a SaaS-managed AI model are different than that of a cloud-managed AI model, which in turn again are different than that of a self-managed AI model. While there are certainly similarities—especially for security—there are significant differences that will need to be addressed for each deployment pattern used.
Strategic Takeaway
Analyse the use cases within your organisation and identify patterns in the adoption of different AI models. Consider factors such as data sensitivity, integration capabilities, and alignment with existing tools and processes. Tailor your approach to deployment and security based on the specific characteristics of each deployment pattern.
There are a lot of considerations for building, operating, and securing AI applications, not the least of which is all the new requirements for model security and scalability. But many of the lessons learned from deploying modern applications across core, cloud, and edge for the past decade will serve organisations well. The core challenges remain the same, and applying the same level of rigour to scaling and securing AI applications will go a long way toward a successful implementation.
But forgoing attention to the differences and leaping in without at least a semi-formal strategy for addressing delivery and security challenges is bound to lead to disappointment down the road.