|
|
A Maturity Model for
|
| Tier | Capability Level | What It Does | Scope |
|---|---|---|---|
| Prompt Based AI | AI Task Advisor | Task-level guidance using SoTA models (GPT-4o, Claude Sonnet). Business users get actionable output via natural language — no engineering required. | Task |
| AI Fine-Tuned | AI Assistant | Pre-trained open-source models (LLaMA, Mistral) adapted to specific business use cases. Builds customised AI Assistants for targeted operational needs. | Task → Process |
| RAG for Custom AI | AI Collaborator | Retrieval Augmented Generation integrates real-time proprietary data. Reduces hallucination, improves accuracy, enables process-level AI collaboration. | Process |
| AI Agents | Autonomous System | Execute multi-step tasks autonomously (including transactions). Multi-agent coordination. Adaptive to new information. Bridges digital and physical worlds. | Process → System |
| LLM as OS / AIOS | AI Operating System | LLMs embedded at device/OS level. Orchestrates agents and tasks at kernel level. Context switching, scalability, and interaction-oriented autonomous systems. | System |
| Predictive AI | Foundation | Statistical Learning, Data Science, Deep Learning — powers purpose-built predictive applications. Foundational, runs in parallel with Generative AI. | Purpose-Built Apps |
Prompt engineering is about crafting strategic inputs — business questions, context, constraints — to guide state-of-the-art AI models (GPT-4o, Claude Sonnet, Gemini) towards outputs that are valuable for faster decision-making.
Enterprises can start here by deploying "AI Advisors" that allow business users to complete tasks through natural language interaction. No engineering required. Further techniques like Chain of Thought (CoT) reasoning help AI models break down complex problems into manageable steps — producing more accurate and reliable outputs.
Best for: Task-level guidance, research acceleration, document drafting, summarisation, Q&A over business content.
→ Generic to Specialised · Task scopeFirms adapt pre-trained open-source models (LLaMA by Meta, Mistral, and others) to meet specific use case or business operation needs. Fine-tuning is an initial step for organisations to build customised solutions that address unique business challenges.
The result is AI Assistants that enhance performance in targeted areas critical to business operations — going beyond what generic prompt-based AI can deliver. Domain-specific language, tone, and knowledge is baked in.
Best for: Specialised customer support, technical documentation, industry-specific advisory, compliance assistance.
Research Paper: Fine-Tuning (arXiv) → Task to Process scopeBase LLM models, while very powerful, are "Generic" in their intelligence — posing key challenges when used directly for business needs. Hallucination and providing outdated information are the two most cited problems in enterprise deployments.
Retrieval Augmented Generation (RAG) enhances AI models by integrating real-time data from trusted, proprietary sources to support business user needs. Use of business-owned knowledge bases ensures up-to-date and accurate information — dramatically reducing the risk of hallucination while improving decision-making quality.
Critically, source systems update dynamically — avoiding costly LLM retraining. RAG operates at a Business Process level (e.g. AI-powered Sales workflows based on CRM data, compliance checks, engineering document Q&A).
Best for: Enterprise knowledge management, compliance workflows, engineering document intelligence, CRM-powered sales assistance.
Research Paper: RAG (arXiv) → Process scope · AI CollaboratorAI Agents go beyond response generation — they execute Actions. When compared to RAG or fine-tuned models, AI Agents represent a significant capability leap: they can execute multi-step tasks (including transactions) without constant human intervention, use complex inputs by breaking them into manageable components, and adapt based on new information and feedback.
AI Agents can work with other Agents (Multi-Agent mode) or with Humans, while integrating with business processes and workflows. In the digital world, Software Agents operate across systems. When coupled with Robotics, they extend into the physical world — enabling capabilities like automated manufacturing, logistics, and vehicle operations.
Best for: Complex business process automation, autonomous purchasing, multi-system workflow orchestration, agentic customer service, physical world automation.
→ Process to System scope · AutonomousLLM Agent Operating System (AIOS) is a new paradigm that integrates Large Language Models directly into the operating system architecture. Conceptualised by Andrej Karpathy, this approach takes the agentic model to a new level by embedding AI capabilities at the core of system operations — not as an application layer, but at the kernel level.
Key components include an LLM OS Agent (orchestrator) that delegates tasks and coordinates information across agents. From a capability perspective, AIOS offers Context Switching between Agents and Tasks, Scalability beyond the Application layer to Kernel level, and the adaptability of LLMs to support multiple or various tasks on a single platform or device.
The innovations at this frontier are happening rapidly — this is the tier to watch closely in 2025 and beyond, particularly in automotive (vehicle OS), industrial, and mobile contexts.
Research Paper: AIOS (arXiv) → System scope · Embedded at device levelThe AI Maturity Model will continue to evolve as Generative AI developments accelerate. But the framework above provides a durable, practical structure for enterprises to understand where they are, clarify the next step, and build AI capability systematically — rather than reactively.
The key insight is that AI maturity is not a binary choice between Predictive and Generative, or between simple prompts and full autonomous agents. It is a journey — and most organisations will operate across multiple tiers simultaneously, applying the right capability to the right business problem.
Let us know what you think in the comments!