Stop doing small AI experiments at the edges; build the governance and platform needed to use AI safely in your core business. If it hasn't happened yet, one day in the near future your CFOs will start asking the important question: "Are we seeing the value from all of our AI spend?" And pressure to show that value will be even stronger when the cost per tokens - which is in large part subsidized today (think of Uber in the early days when rides were $3-4) - inevitably will reflect its real price over time. Many organizations today are deploying peripheral agents in regards to the their core business, to experiment in ways that AI can accelerate their business. But the truth is that those organization will never see the outsized outcomes that AI can bring to the table until they start using AI in their core business. A peripheral use case can never justify the enormous spend that is currently being poured in AI. And here is the chicken-egg problem: organizations know that the core business needs to be transformed with AI in order to get those outsized benefits and justify the AI spend, but their are not ready yet to deploy agents at the core because they either don't trust the non-deterministic nature of agents, they don't have the right governance and infrastructure in place to be comfortable doing that, and also while they have developed early experiments across one or more teams that specialized in AI, they haven't developed an internal platform that allows to scale those experiments from the teams into the enterprise. And sometimes, it's all of the above! So how do we deploy AI in the core business? - We need to establish concrete AI, data and agentic governance. - We need to establish a platform that allows us to manage identify for agents, their permissions, what tools and APIs they can consume, being able to observe and monitor outcomes, and so on. - We need to start transforming the core business little by little with AI and agents that are not going to put the business and the customers at risk. And here at Kong we partner with our enterprise customers to bring that: an AI platform that allows them to start showing outsized value in their core business with AI, to not only justify the AI spend, but also to create an internal platform for AI that can scale our agents and their interactions with APIs, data and LLMs in a governed way. You can learn more about our platform at https://lnkd.in/en-BndJX
Deploy AI in Core Business with Governance and Platform
More Relevant Posts
-
AI does not hit a capability ceiling first. It hits a cost ceiling. For the last year, the narrative around agentic AI has been nonstop execution. Persistent agents. Continuous workflows. Systems running 24/7 with minimal human involvement. Now the constraints are showing up. Usage caps, session limits, and guardrails are not signs the vision failed. They are signs the economics are finally showing up at scale. Every loop, tool call, and token has a price. Context windows expand, KV caches persist, and inference cycles stack on top of each other. In a demo, that feels manageable. In production, it becomes a cost and latency problem very quickly, especially once retries and validation are factored in. This is where architecture matters. Many current approaches assume persistent context, long lived sessions, and centralized model calls. That works until scale forces a correction. Teams start throttling usage and limiting autonomy, not because they do not want continuous agents, but because continuous inference is expensive. There is another path. Design for selective execution. Minimize unnecessary context. Trigger models intentionally. Push more processing closer to the data. This is where on prem and hybrid approaches start to make sense, not as replacements for cloud AI, but as ways to control token spend, reduce latency, and make system behavior more predictable. The narrative has been that AI removes constraints. What we are seeing is that it introduces new ones, just at a different layer. JetHead Take: The future is not about running less AI. It is about being more efficient with what you have. The teams that win will treat inference, context, and data movement as economic decisions, and will use architectures like on premise and hybrid models to control cost, latency, and scale more intelligently. #ArtificialIntelligence #AIAgents #EnterpriseAI #MLOps
To view or add a comment, sign in
-
-
Placing some agents on prem to reduce tokens used by cloud solutions could definitely reduce cost. Love the hybrid approach
AI does not hit a capability ceiling first. It hits a cost ceiling. For the last year, the narrative around agentic AI has been nonstop execution. Persistent agents. Continuous workflows. Systems running 24/7 with minimal human involvement. Now the constraints are showing up. Usage caps, session limits, and guardrails are not signs the vision failed. They are signs the economics are finally showing up at scale. Every loop, tool call, and token has a price. Context windows expand, KV caches persist, and inference cycles stack on top of each other. In a demo, that feels manageable. In production, it becomes a cost and latency problem very quickly, especially once retries and validation are factored in. This is where architecture matters. Many current approaches assume persistent context, long lived sessions, and centralized model calls. That works until scale forces a correction. Teams start throttling usage and limiting autonomy, not because they do not want continuous agents, but because continuous inference is expensive. There is another path. Design for selective execution. Minimize unnecessary context. Trigger models intentionally. Push more processing closer to the data. This is where on prem and hybrid approaches start to make sense, not as replacements for cloud AI, but as ways to control token spend, reduce latency, and make system behavior more predictable. The narrative has been that AI removes constraints. What we are seeing is that it introduces new ones, just at a different layer. JetHead Take: The future is not about running less AI. It is about being more efficient with what you have. The teams that win will treat inference, context, and data movement as economic decisions, and will use architectures like on premise and hybrid models to control cost, latency, and scale more intelligently. #ArtificialIntelligence #AIAgents #EnterpriseAI #MLOps
To view or add a comment, sign in
-
-
This is an important shift that more teams are starting to feel in production. AI doesn’t remove constraints. It moves them. For a while, the focus was on what models could do. Now the conversation is becoming about what they should do, when, and at what cost. The teams getting real value aren’t just building smarter systems, they’re building more intentional ones. Selective execution. Scoped context. Human validation where it matters. Not less AI, just more disciplined use of it.
AI does not hit a capability ceiling first. It hits a cost ceiling. For the last year, the narrative around agentic AI has been nonstop execution. Persistent agents. Continuous workflows. Systems running 24/7 with minimal human involvement. Now the constraints are showing up. Usage caps, session limits, and guardrails are not signs the vision failed. They are signs the economics are finally showing up at scale. Every loop, tool call, and token has a price. Context windows expand, KV caches persist, and inference cycles stack on top of each other. In a demo, that feels manageable. In production, it becomes a cost and latency problem very quickly, especially once retries and validation are factored in. This is where architecture matters. Many current approaches assume persistent context, long lived sessions, and centralized model calls. That works until scale forces a correction. Teams start throttling usage and limiting autonomy, not because they do not want continuous agents, but because continuous inference is expensive. There is another path. Design for selective execution. Minimize unnecessary context. Trigger models intentionally. Push more processing closer to the data. This is where on prem and hybrid approaches start to make sense, not as replacements for cloud AI, but as ways to control token spend, reduce latency, and make system behavior more predictable. The narrative has been that AI removes constraints. What we are seeing is that it introduces new ones, just at a different layer. JetHead Take: The future is not about running less AI. It is about being more efficient with what you have. The teams that win will treat inference, context, and data movement as economic decisions, and will use architectures like on premise and hybrid models to control cost, latency, and scale more intelligently. #ArtificialIntelligence #AIAgents #EnterpriseAI #MLOps
To view or add a comment, sign in
-
-
This is where the conversation is finally getting real. AI doesn’t remove constraints. It shifts them into cost, latency, and control.
AI does not hit a capability ceiling first. It hits a cost ceiling. For the last year, the narrative around agentic AI has been nonstop execution. Persistent agents. Continuous workflows. Systems running 24/7 with minimal human involvement. Now the constraints are showing up. Usage caps, session limits, and guardrails are not signs the vision failed. They are signs the economics are finally showing up at scale. Every loop, tool call, and token has a price. Context windows expand, KV caches persist, and inference cycles stack on top of each other. In a demo, that feels manageable. In production, it becomes a cost and latency problem very quickly, especially once retries and validation are factored in. This is where architecture matters. Many current approaches assume persistent context, long lived sessions, and centralized model calls. That works until scale forces a correction. Teams start throttling usage and limiting autonomy, not because they do not want continuous agents, but because continuous inference is expensive. There is another path. Design for selective execution. Minimize unnecessary context. Trigger models intentionally. Push more processing closer to the data. This is where on prem and hybrid approaches start to make sense, not as replacements for cloud AI, but as ways to control token spend, reduce latency, and make system behavior more predictable. The narrative has been that AI removes constraints. What we are seeing is that it introduces new ones, just at a different layer. JetHead Take: The future is not about running less AI. It is about being more efficient with what you have. The teams that win will treat inference, context, and data movement as economic decisions, and will use architectures like on premise and hybrid models to control cost, latency, and scale more intelligently. #ArtificialIntelligence #AIAgents #EnterpriseAI #MLOps
To view or add a comment, sign in
-
-
Enterprise AI does not only break on model quality. It breaks on organizational reality. Your stack may retrieve the right file and still fail because it CANNOT: 🙄 find the right owner, 🙄 route to the real approver, 🙄 follow how work actually moves, 🙄 explain why an escalation happened, 🙄 or keep up with org drift after the reorg. In this new piece, we lay out 6 repeatable failure modes we see across enterprise AI rollouts: ⚠️ retrieval is not routing ⚠️ authority hallucination ⚠️ workflow on paper is not reality ⚠️ governance by policy alone ⚠️ context-less escalation ⚠️ static behavioral snapshots Our argument is that these are not edge cases. They are structural. And they do not get solved by better prompts or better RAG alone, because they sit in a behavioral layer that content-level systems cannot see. That is the gap we think Large Behavior Models can help fill. If you work on AI governance, agent orchestration, enterprise search, or platform strategy, this is the layer worth paying attention to. Read: https://lnkd.in/ejSbkQKF
To view or add a comment, sign in
-
Many enterprise AI failures are being explained at the wrong layer. People often assume the issue is model quality, retrieval quality, or policy design. Sometimes it is. But often the deeper problem is that the system does not understand enough about how the organization actually functions. Based on years of work around organizational behavior, knowledge flow, and enterprise execution, I tried to define six recurring failure patterns that I think deserve more attention: - Retrieval is not routing - Authority hallucination - Workflow on paper is not reality - Governance by policy alone - Context-less escalation - Static behavioral snapshots What made me write it more explicitly is that versions of these same problems have also been pointed out publicly, from different angles, by people like Aaron Levie, Arvind Jain, Ethan Mollick, Andrej Karpathy, Hamel Husain, Josh Bersin, and Ali Ghodsi. So this essay is not a product pitch as much as an attempt to define a layer of enterprise reality that I think the industry still under-models. If you work on enterprise search, agent orchestration, AI governance, or enterprise AI strategy, this may be worth reading. https://lnkd.in/eVnwyrSr
Enterprise AI does not only break on model quality. It breaks on organizational reality. Your stack may retrieve the right file and still fail because it CANNOT: 🙄 find the right owner, 🙄 route to the real approver, 🙄 follow how work actually moves, 🙄 explain why an escalation happened, 🙄 or keep up with org drift after the reorg. In this new piece, we lay out 6 repeatable failure modes we see across enterprise AI rollouts: ⚠️ retrieval is not routing ⚠️ authority hallucination ⚠️ workflow on paper is not reality ⚠️ governance by policy alone ⚠️ context-less escalation ⚠️ static behavioral snapshots Our argument is that these are not edge cases. They are structural. And they do not get solved by better prompts or better RAG alone, because they sit in a behavioral layer that content-level systems cannot see. That is the gap we think Large Behavior Models can help fill. If you work on AI governance, agent orchestration, enterprise search, or platform strategy, this is the layer worth paying attention to. Read: https://lnkd.in/ejSbkQKF
To view or add a comment, sign in
-
DeepSeek V4 just shifted the cost structure of enterprise AI. On Friday, the Chinese AI firm released its new flagship model with dramatically expanded context windows and improved efficiency. The model processes longer prompts at lower cost than competing closed systems. It's open source, available for commercial use. 𝗧𝗵𝗿𝗲𝗲 𝗰𝗼𝗺𝗺𝗲𝗿𝗰𝗶𝗮𝗹 𝗶𝗺𝗽𝗹𝗶𝗰𝗮𝘁𝗶𝗼𝗻𝘀: • 𝗜𝗻𝗳𝗿𝗮𝘀𝘁𝗿𝘂𝗰𝘁𝘂𝗿𝗲 𝗮𝗿𝗯𝗶𝘁𝗿𝗮𝗴𝗲 𝗻𝗮𝗿𝗿𝗼𝘄𝘀. When open models match frontier capabilities at a fraction of the cost, the defensibility of closed API pricing erodes fast. • 𝗖𝗼𝗻𝘁𝗲𝘅𝘁 𝗯𝗲𝗰𝗼𝗺𝗲𝘀 𝗰𝗼𝗺𝗺𝗼𝗱𝗶𝘁𝘆. Long-context processing was a moat six months ago. Now it's table stakes, and the competitive edge moves to application layer intelligence. • 𝗗𝗲𝗽𝗹𝗼𝘆𝗺𝗲𝗻𝘁 𝗼𝗽𝘁𝗶𝗼𝗻𝗮𝗹𝗶𝘁𝘆 𝗺𝗮𝘁𝘁𝗲𝗿𝘀 𝗺𝗼𝗿𝗲. Teams that locked into single-vendor APIs are discovering they've traded flexibility for convenience at exactly the wrong moment. The gap between open and closed models isn't closing, it's inverting on specific workloads. That changes procurement conversations and build-versus-buy calculations across the stack. Most enterprise AI strategies assume frontier models stay expensive and proprietary. V4 tests that assumption.
To view or add a comment, sign in
-
Most AI orgs aren’t wrong. They’re just built on a model that made sense for a long time. You had separate teams for data, AI, cloud, engineering. You stitched them together with process. Things moved in sequence. There were handoffs, reviews, approvals. It wasn’t fast, but it was controlled, and that mattered. It’s also not reasonable to expect companies to flip their operating model overnight for something many still see as unproven at enterprise scale. And honestly, this shift is uncomfortable. On the surface, it feels like you’re adding overhead. More teams. More duplication. It goes against the instinct most firms have built over the last 20+ years, which is to centralize, standardize, and control. So the natural reaction is to keep everything grouped and coordinated. But this is where it starts to break because AI doesn’t stay in one lane. It cuts across all of it. And when you try to run it through that same structure, the work just gets stuck between teams. Everyone’s involved. No one really owns it. You end up dotted-line to death. I think a lot of talk round AI failing in production is about the op-model/process side (vs just the tech side). It was set up to fail to begin with. You still need teams that build the roads. That has to be consistent and reliable. But you also need teams that do the driving. And that work happens in small, high-agency teams that own the outcome end to end, with a mix of skill sets. Not coordinating across five groups. Not waiting on approvals. Just taking the road and moving. One owner. A couple of people who can actually build. A clear outcome. That’s it.
To view or add a comment, sign in
-
-
🔄𝐓𝐫𝐚𝐧𝐬𝐟𝐨𝐫𝐦 𝐘𝐨𝐮𝐫 𝐃𝐚𝐭𝐚 𝐰𝐢𝐭𝐡 𝐃𝟖𝐓𝐀𝐎𝐏𝐒. At D8TAOPS, they focus on how businesses actually manage, govern, and use their data across real environments. With decades of enterprise experience behind the platform, D8TAOPS brings a different approach to data ingestion, curation, security, and governance through the D8:APPS framework. Based in Beaverton, OR, D8TAOPS is designed to simplify data operations while creating a governed foundation for AI, machine learning, and generative AI initiatives. Rather than forcing replatforming, D8TAOPS works across your existing systems to prepare, connect, and control data where it already lives. Whether in healthcare, finance, manufacturing, or other regulated industries, the goal is the same: help teams move faster, reduce manual effort, and build data systems that are actually ready for production AI. ➡️https://d8taops.com/
To view or add a comment, sign in
-
-
I spent time with Forrester’s Sudha Maheshwari pressure-testing what enterprise AI governance needs to look like in 2026. One stat stopped the conversation cold: Shadow AI is growing 2–3x faster than sanctioned AI. This isn’t a discipline problem. It’s a design problem. Governance programs built in 2023 assumed AI would enter through a central procurement door. That door is gone. Every SaaS update, browser tab, and vendor tool is now an AI entry point, with no visibility, no oversight, and no audit trail. If your model is still “review and approve,” it’s already behind. You can’t govern what you can’t see. The 2026 shift is clear: discover first → govern continuously → enable safely That’s the control plane enterprises are now building. If you're rethinking your approach, this conversation goes deep into what that actually looks like. Full conversation [on demand here]: https://lnkd.in/gXCZtHg6
To view or add a comment, sign in
-
Explore related topics
- How to Scale AI in Enterprises
- How to Make Artificial Intelligence a Business Imperative
- How to Scale AI Responsibly in Business
- How To Scale AI In Regulated Industries
- Strategies For Integrating AI Across Teams
- How to Streamline Enterprise AI Integration
- Developing Scalable AI Use Cases
- Best Practices For Scaling AI In Large Companies
- How to Build AI Adoption Awareness in Large Firms
- Scaling AI Solutions Without Sacrificing Quality
Explore content categories
- Career
- Productivity
- Finance
- Soft Skills & Emotional Intelligence
- Project Management
- Education
- Technology
- Leadership
- Ecommerce
- User Experience
- Recruitment & HR
- Customer Experience
- Real Estate
- Marketing
- Sales
- Retail & Merchandising
- Science
- Supply Chain Management
- Future Of Work
- Consulting
- Writing
- Economics
- Artificial Intelligence
- Employee Experience
- Workplace Trends
- Fundraising
- Networking
- Corporate Social Responsibility
- Negotiation
- Communication
- Engineering
- Hospitality & Tourism
- Business Strategy
- Change Management
- Organizational Culture
- Design
- Innovation
- Event Planning
- Training & Development