I. Introduction: The Dawn of Agentic AI
Artificial intelligence (AI) is undergoing a period of rapid and profound transformation. We have witnessed AI systems evolve from primarily recognizing patterns in data, characteristic of traditional machine learning (ML), to generating novel content, as seen with the rise of generative AI. Now, the field is venturing into a new frontier: systems that can not only process information or create content but can also act and achieve goals autonomously within complex environments. This emerging paradigm is increasingly known as "agentic AI".1 These systems represent a significant departure, focusing on perception, reasoning, planning, and task execution with minimal human intervention.3
At a fundamental level, agentic AI refers to systems capable of making independent decisions and taking purposeful actions without constant, explicit human direction. They demonstrate a notable degree of independence and adaptability, particularly in dynamic and unpredictable settings.3 Unlike basic ML algorithms, which are primarily designed to identify patterns or predict outcomes based on historical data, agentic AI can define, set, and actively pursue specific objectives.3 These are AI systems engineered to accomplish designated goals with limited supervision, often by emulating human-like decision-making processes to solve problems in real-time.4
The development of agentic AI marks a significant leap forward. It builds upon the capabilities of generative AI, especially large language models (LLMs), extending their utility by applying generated outputs towards achieving concrete goals within interactive environments.4 This signifies a pivotal shift from earlier AI systems, which were often reactive or deliberative based on predefined rules, towards architectures that are learning-driven and context-aware.6 This advancement is characterized by enhanced reasoning faculties, sophisticated interaction capabilities, and the capacity to autonomously tackle complex, multi-step problems.1 The journey towards agentic AI can be seen as an evolutionary progression, where foundational AI concepts and traditional machine learning paved the way for generative AI, which in turn provided crucial building blocks—like the reasoning and language understanding capabilities of LLMs—for task-specific AI agents. Agentic AI systems, often involving multiple such agents and orchestrated autonomy, represent the current cutting edge of this trajectory. Understanding this lineage is key to appreciating the sophistication and potential of modern agentic systems.
The core distinction setting agentic AI apart is its proactive pursuit of goals coupled with an autonomous decision-making capacity within dynamic and multifaceted environments. This moves AI beyond the reactive or pattern-matching nature of its predecessors.3 The ability to independently set and pursue goals necessitates a high degree of autonomy, adaptability, and complex reasoning. This shift has profound implications for how humans will interact with AI, moving from a model of direct, step-by-step instruction to one of delegating complex objectives to autonomous systems.
This exploration will delve into the intricacies of agentic AI, examining its defining characteristics and the sophisticated learning mechanisms that empower these intelligent agents. It will then take a focused look at their transformative applications within the domain of robotics, a field uniquely positioned to benefit from autonomous action. Finally, the discussion will address the inherent challenges and the exciting future prospects that lie ahead for agentic AI.
II. Defining Agentic AI: More Than Just Machine Learning
To fully appreciate the impact of agentic AI, a clear understanding of its definition and core characteristics is essential, particularly in how it relates to, yet surpasses, previous AI paradigms.
Detailed Definition of Agentic AI and AI Agents
Agentic AI is broadly defined as an artificial intelligence system capable of accomplishing specific, often complex, goals with limited human supervision. These systems are composed of, or act as, AI agents—machine learning models designed to mimic human decision-making processes to solve problems, often in real-time.4 A defining feature of agentic AI is its inherent "agency," or its capacity to act independently and purposefully within its environment.4 Such systems employ sophisticated reasoning and iterative planning to autonomously navigate and resolve multi-step problems.1
AI Agents, in this context, are often software programs or modular systems that interact with their environment by perceiving it, collecting data, and then using that data to perform self-determined tasks aimed at achieving predetermined goals.7 While humans typically set the overarching goals, the AI agent independently chooses the most suitable sequence of actions to achieve them.7 These agents are frequently driven by powerful foundation models like LLMs and Large Image Models (LIMs), particularly for narrow, task-specific automation.6
It is important to note that the term "agentic AI" itself is part of an evolving lexicon and can represent a spectrum of capabilities. While "agentic" describes the quality of autonomous, goal-driven behavior, "Agentic AI systems" often refer to more complex, advanced implementations, frequently involving multiple specialized agents collaborating under an orchestrated autonomy.6 Thus, agentic AI can be considered a subfield of artificial intelligence, with the AI agent being the fundamental unit of study and operation within this subfield.1 This distinction is crucial: an individual AI might exhibit agentic capabilities, but an Agentic AI system often implies a more sophisticated architecture designed for complex, coordinated action.
Key Characteristics
Agentic AI is distinguished by a confluence of several key characteristics that enable its autonomous and adaptive behavior:
- Autonomy and Independent Decision-Making: This is the cornerstone of agentic AI. These systems operate independently, analyzing available data to determine the best course of action without requiring constant human input or intervention.11 They exhibit a capacity for independent decision-making and action even in dynamic and unpredictable environments.3 This autonomy is critical for scalability and for applications where persistent human oversight is impractical or impossible, such as in some robotic deployments or automated customer support scenarios.6
- Goal-Oriented Behavior & Proactiveness: Agentic systems are designed with specific objectives in mind and continuously monitor their environment to take actions that advance these goals.12 They can proactively define their own sub-goals as steps toward achieving broader objectives, dynamically breaking down complex problems into manageable parts and adjusting strategies based on real-time feedback.11 This includes anticipating potential problems and taking corrective actions preemptively, shifting AI from a purely reactive tool to a proactive, strategic entity.12
- Perception: Agents collect data from their environment through various means, including sensors (in robotics), APIs, databases, or direct user interactions. Technologies like Natural Language Processing (NLP) and computer vision are vital for interpreting this data.4 The quality, diversity, and timeliness of perceived data significantly influence the agent's effectiveness.11
- Reasoning & Planning: Once data is collected, the agent processes it to extract meaningful insights, interpret queries, detect patterns, and understand the broader context. This reasoning phase is crucial for determining appropriate actions.4 Knowledge representation techniques, such as knowledge graphs and symbolic logic, play a key role in organizing information for effective reasoning.5 Based on this reasoning, the agent develops strategies and plans—often using methods like decision trees, reinforcement learning, Monte Carlo Tree Search (MCTS), or evolutionary algorithms—to achieve its goals, evaluating various options and selecting the optimal path.4
- Execution/Action: Following the planning phase, the agent executes the chosen actions. This can involve interacting with external systems (e.g., calling APIs, controlling robotic actuators) or providing responses to users.4 This continuous cycle of perception, reasoning, planning, and action, followed by learning from the outcomes, is fundamental to how agentic systems function and improve over time.4
- Adaptability and Continuous Learning: A hallmark of agentic AI is its ability to learn from interactions and experiences, progressively refining its decision-making processes and overall performance.4 Using techniques such as reinforcement learning, model fine-tuning, or incorporating human feedback, these agents can adapt to new challenges, identify inefficiencies, and self-correct their behavior.11 This capacity for self-improvement is crucial for effective operation in dynamic and unpredictable real-world environments, distinguishing agentic AI from static, pre-programmed systems.12
- Context Awareness: Agentic AI systems demonstrate a keen understanding of context. They can adjust their responses and actions based on ongoing user interactions, data from external sources, or real-time environmental inputs.11 For instance, an AI-driven medical assistant might interpret a patient's symptoms not in isolation, but within the broader context of their medical history, lifestyle, and even genetic factors, leading to more accurate assessments.11 This contextual understanding makes agentic AI significantly more effective and nuanced in its real-world applications.
- Collaboration (Human-AI and AI-AI): Agentic AI is not limited to solitary operation. It can collaborate effectively with humans by interpreting intent, receiving feedback, and delegating tasks where appropriate.11 Furthermore, a significant advancement in agentic AI is the development of Multi-Agent Systems (MAS), where multiple specialized AI agents coordinate their activities, communicate with each other, and dynamically allocate sub-tasks to achieve a common, broader objective.6 This collaborative capability, especially in MAS, allows for the tackling of problems that would be beyond the scope of any single agent and is a defining feature of advanced Agentic AI systems.
The true power of agentic AI arises not from a single technological breakthrough, but from the synergistic integration of these diverse and advanced AI capabilities—perception, reasoning, learning, interaction, and collaboration, often supercharged by the capabilities of LLMs. Achieving true agency is a complex systems engineering challenge that requires this multifaceted approach; no single component alone is sufficient. Progress in any of these underlying AI sub-disciplines will directly contribute to the advancement of agentic AI.
How Agentic AI Differs from Traditional AI/ML Models and Generative AI
The unique characteristics of agentic AI set it apart from earlier forms of artificial intelligence:
- Traditional AI/ML: These systems typically operate within predefined constraints and follow predetermined paths. While capable of complex pattern recognition and prediction, they often require significant human intervention for adapting to new situations or making decisions outside their initial training scope.4 Basic ML algorithms, for example, are primarily designed to identify patterns and predict outcomes based on existing data.3
- Generative AI (GenAI): GenAI models, such as LLMs, focus primarily on creating novel content (text, images, code) based on learned patterns from their training data.4 Agentic AI builds upon and extends this capability. It applies the outputs of generative models, like the reasoning or planning abilities of an LLM, towards achieving specific goals and actively interacting with its environment.1 In many agentic systems, GenAI serves as a crucial component or a precursor technology rather than being the end goal itself.1 The transition from Generative AI to Agentic AI signifies a critical shift from content creation to action and impact in the real or digital world. While GenAI produces digital artifacts, agentic AI can control robots, alter physical objects, or manage systems without necessarily producing a tangible user-facing artifact.1 This capacity for direct real-world interaction means agentic AI has the potential for far more significant consequences, both beneficial and detrimental, than generative AI alone.
The fundamental distinction lies in agentic AI's emphasis on autonomy, proactive goal-driven behavior, dynamic adaptation to changing circumstances, and, increasingly, multi-agent collaboration. Traditional AI is generally more constrained, and Generative AI is primarily concerned with content generation. Agentic AI, in essence, can do things in the world, not just predict, classify, or generate.
III. The Learning Engine: How Agentic Agents Evolve
The capacity to learn is not merely an add-on feature for agentic AI; it is a fundamental component that underpins their autonomy, adaptability, and ability to handle complex tasks in ever-changing environments.4 Unlike static systems that rigidly follow pre-programmed rules, agentic AI systems perceive their surroundings, plan their actions, adapt to new information, and crucially, learn from their experiences and feedback.5 This continuous learning process allows them to update their internal models and behaviors, leading to enhanced decision-making capabilities over time.7
Reinforcement Learning (RL) as a Core Mechanism
Reinforcement Learning (RL) stands out as a pivotal learning paradigm for agentic AI, providing a robust framework for agents to learn optimal behaviors through interaction and feedback, rather than explicit instruction.
- Fundamental Concepts of RL: The typical RL scenario involves an agent interacting with an environment over a sequence of discrete time steps.17 At each step, the agent observes the current state of the environment, which is a representation of its situation.17 Based on this state, the agent selects an action from a set of available choices.17 The environment, in turn, transitions to a new state and provides a reward (or penalty) signal to the agent.17 The overarching goal of the agent is to learn a policy—a strategy or mapping from states to actions—that maximizes the cumulative reward it receives over the long term.17 This trial-and-error interaction, driven by the pursuit of rewards, allows the agent to autonomously discover effective strategies without being explicitly programmed for every eventuality, a critical feature for operating in unknown or dynamic conditions.
- How RL Enables Learning in Agentic Systems: RL empowers agentic AI by allowing agents to evaluate the outcomes of their actions and progressively adjust their strategies (i.e., update their policies) to favor actions that yield higher long-term rewards.4 This iterative process of action, feedback, and adaptation is central to the agent's ability to learn and refine its behavior based on real-world interactions.5 A key distinction from supervised learning is that RL does not rely on labeled input-output pairs, nor does it involve explicit correction of sub-optimal actions. Instead, learning is guided by the scalar reward signal, which implicitly indicates the desirability of the agent's behavior.18
- Specific RL Techniques Fueling Agentic AI: Several RL algorithms and frameworks are instrumental in developing agentic capabilities:
- Q-Learning: A model-free, value-based RL method where the agent learns a "Q-value" for each state-action pair. The Q-value represents the expected future cumulative reward of taking a specific action in a given state and then following the optimal policy thereafter. The agent iteratively updates these Q-values based on the rewards it receives, eventually converging towards an optimal action-selection strategy.17
- Policy Gradient Methods: These methods directly learn the policy function, which maps states to actions (or probabilities of actions). Algorithms like REINFORCE use Monte Carlo methods to estimate the gradient of the expected reward with respect to the policy's parameters, and then update the policy in the direction that increases this expected reward. More advanced techniques like Proximal Policy Optimization (PPO) aim to improve the stability and sample efficiency of policy gradient updates.17
- Deep Reinforcement Learning (DRL): DRL represents a significant advancement by integrating deep neural networks (DNNs) into the RL framework. DNNs serve as powerful function approximators, enabling RL agents to handle complex, high-dimensional state and action spaces, such as those encountered in robotics (from raw sensor data) or complex games.14
- Deep Q-Networks (DQN): DQN combines Q-Learning with a deep neural network to approximate the Q-value function, allowing agents to learn effectively in environments with vast state spaces, like learning to play Atari games from pixel inputs.17
- Actor-Critic Methods: These hybrid methods combine the strengths of both value-based and policy-based approaches. They typically involve two neural networks: an "actor" that learns and implements the policy (deciding which actions to take), and a "critic" that learns a value function to evaluate the states or state-action pairs, providing feedback to the actor. This separation can lead to more stable and efficient learning.17
The progression from foundational RL algorithms to sophisticated DRL techniques, including advanced policy optimization and actor-critic architectures, has been crucial for enabling agents to tackle the complexities inherent in real-world tasks and environments.
Role of Supervised Learning
While RL is central to decision-making and adaptation, supervised learning (SL) plays a critical role in equipping agentic AI with foundational capabilities, particularly in perception and knowledge representation:
- Perception: SL is extensively used to train models on labeled datasets for tasks essential to environmental understanding. In robotics, Convolutional Neural Networks (CNNs) are trained on vast image datasets to perform object recognition, scene understanding, and motion tracking, enabling robots to "see" and interpret their surroundings.13 Similarly, NLP models, often trained with SL, allow agents to understand human language, whether it's user queries, instructions, or text encountered in the environment.13 For example, AI chatbots are often trained on large datasets of customer service conversations and their corresponding resolutions to learn how to respond appropriately.16
- Knowledge Representation & Predictive Models: SL is used to build predictive models from labeled data, forming a part of the agent's knowledge base.16 LLMs, which are foundational to many modern agentic systems, are typically pre-trained using supervised (and self-supervised) learning methods on massive text and code corpora. This pre-training endows them with a broad understanding of language, facts about the world, and reasoning patterns, which serve as a powerful knowledge base.1
Role of Unsupervised and Self-Supervised Learning
Unsupervised learning (UL) and self-supervised learning (SSL) contribute to an agent's ability to learn from the vast amounts of unlabeled data prevalent in the real world:
- Pattern Discovery: UL techniques allow agents to discover hidden structures, patterns, or relationships within unlabeled data. This can be used for tasks like clustering similar environmental states, identifying anomalies in sensor readings, or segmenting customer behaviors without prior labels.16
- Self-Supervised Learning (SSL): SSL is a powerful paradigm where models generate their own supervisory signals (implicit labels) from the input data itself. This reduces the dependency on manually labeled datasets, which are expensive and time-consuming to create. SSL has been particularly transformative in pre-training large foundation models, including LLMs and vision models, by enabling them to learn rich representations from enormous quantities of unstructured text and image data.4
The Synergy of LLMs with Learning Mechanisms
Large Language Models are increasingly becoming active, adaptable components within the learning loop of agentic systems, rather than just static knowledge repositories:
LLMs serve as powerful engines for reasoning, planning, and natural language interaction within agentic architectures.1 Agentic AI systems often integrate elements of RL where LLM-based agents interact with their environment, potentially using external tools, and receive feedback that informs their future actions. This feedback can be used for "instant learning" or for more structured fine-tuning of the LLM itself.1
Recent research highlights this synergy. For instance, the FLAG-TRADER architecture uses an LLM as the policy network in an RL setup for financial trading. The LLM processes market information (linguistic processing) and outputs trading actions. Policy gradient methods then update the LLM's parameters based on the rewards generated by these trading actions.22 Similarly, the ML-Agent framework explores LLM agents learning through interactive experimentation on ML tasks using online RL.23 The ARTIST framework enables LLMs to autonomously decide when and how to invoke tools within complex reasoning chains, learning these strategies through outcome-based RL.24 This direct integration of LLMs into the RL loop allows agents to combine the broad knowledge and reasoning capabilities of LLMs with the adaptive learning-from-experience paradigm of RL, potentially leading to more robust, generalizable, and sophisticated agentic behaviors than either approach could achieve in isolation.
The learning processes in agentic AI are often a hybrid, synergistically combining different machine learning paradigms. RL is typically employed for decision-making, strategy optimization, and learning control policies through interaction. Supervised learning is crucial for perception (e.g., training computer vision models for a robot's eyes or NLP models for understanding commands) and for pre-training foundational knowledge in models like LLMs. Unsupervised and self-supervised learning enable agents to extract patterns from unlabeled data and are fundamental to the initial training of large-scale foundation models. This combination is necessary because different learning approaches are suited to different facets of an agent's complex functionality.
Furthermore, the concept of "learning" in agentic AI is multi-layered. It extends beyond simple parameter updates in a neural network to encompass higher-level cognitive functions such as dynamic goal decomposition (breaking a large goal into smaller, manageable sub-goals), strategic adaptation (changing plans when circumstances change), and potentially even learning how to learn (meta-learning). This makes agentic AI qualitatively different from systems that only learn a fixed mapping or pattern, endowing them with a more flexible and human-like learning capability.
IV. Agentic AI in Action: Revolutionizing Robotics
The principles of agentic AI—autonomy, goal-orientation, continuous learning, and adaptation—are particularly transformative in the field of robotics. They are enabling a shift from robots that execute pre-programmed sequences of actions to intelligent systems capable of perceiving their environment, making independent decisions, learning from experience, and collaborating with humans and other robots in complex, dynamic settings.1 Reinforcement learning, especially Deep Reinforcement Learning (DRL), provides a powerful framework for designing sophisticated and often hard-to-engineer behaviors in robotic systems, allowing them to acquire skills through interaction and feedback.19
Autonomous Navigation and Perception
For a robot to act agentically, it must first accurately perceive and understand its surroundings. This involves sophisticated multi-modal perception capabilities:
- Computer Vision (CV): CNNs and other deep learning models are extensively used for object recognition, scene understanding, and motion tracking, enabling robots to interpret visual information from cameras.13 Advanced CV systems can detect subtle visual cues, such as minute defects in manufacturing or indicators of crop health in agriculture.13
- Natural Language Processing (NLP): NLP capabilities allow robots to understand human instructions given in natural language or to process textual information present in their operational environment.13
- Sensor Fusion: Real-world environments are complex and often noisy. Sensor fusion techniques integrate data from a variety of sensors (e.g., visual, auditory, tactile, LiDAR, radar) to provide a more robust and holistic understanding of the environment. This is particularly crucial for reliable operation in challenging conditions.20
- Spatial AI: An advancement beyond basic mapping, Spatial AI integrates Simultaneous Localization and Mapping (SLAM) techniques with cognitive reasoning. This allows robots not only to build maps of their environment and track their position within them but also to develop a context-aware understanding of the objects and relationships within that space, including their functions and potential interactions.27
A prime application area for agentic navigation and perception is Multi-Agent Autonomous Driving. Autonomous vehicles inherently require agentic capabilities to operate independently and, increasingly, to collaborate with other vehicles and infrastructure.8 A significant challenge in this domain is the efficient exchange of semantically rich multimodal information (e.g., sensor data, intent signals) between vehicles, especially over bandwidth-limited wireless channels. Directly transmitting raw sensor data is often inefficient and can overwhelm communication capacities.14
One proposed solution is the Retrieval-Augmented Multimodal Semantic Communication (RAMSemCom) framework. RAMSemCom aims to facilitate the effective exchange of essential semantic multimodal knowledge—spanning visual, textual, and auditory information—among autonomous vehicles and other agentic AI entities like collaborative robots or edge devices. Inspired by concepts like Retrieval-Augmented Generation (RAG) and Retrieval-Augmented Perception (RAP), RAMSemCom employs local caching and selective transmission of only the most critical portions of information. This approach can utilize DRL-based retrieval strategies to optimize what information is shared, thereby enhancing bandwidth efficiency while preserving crucial semantic details necessary for effective coordination.8 Such systems involve complex communication patterns, including intradomain inter-agent communication (e.g., car-to-car sharing of traffic states or intended paths) and intra-agent communication (e.g., between perception, planning, and execution modules within a single vehicle).14
Complex Manipulation Tasks
Agentic AI is also enabling robots to perform complex manipulation tasks that extend over long horizons and require sophisticated reasoning, precise execution, and robust error recovery. Traditional approaches often struggle with error accumulation over extended sequences and lack effective mechanisms for verifying progress or recovering from failures.8
The 'Agentic Robot' framework offers a brain-inspired solution for vision-language-action models in embodied agents designed for such tasks.8 This framework introduces Standardized Action Procedures (SAP), a novel coordination protocol that governs the interactions between specialized components. Drawing inspiration from Standardized Operating Procedures (SOPs) in human organizations, SAP establishes structured workflows for the distinct phases of planning, execution, and verification.25 The architecture comprises three key specialized components:
1. A Planner: This is a large reasoning model (often LLM-based) responsible for decomposing high-level natural language instructions into a sequence of semantically coherent subgoals.
2. An Executor: A vision-language-action (VLA) model that generates continuous control commands for the robot based on real-time visual inputs and the current subgoal.
3. A Verifier: A temporal verifier that enables dynamic self-verification of task progress and success without requiring external supervision.
By enforcing consistent procedures through SAP, the Agentic Robot framework aims to dramatically reduce error accumulation and provide robust mechanisms for failure recovery. This approach has demonstrated state-of-the-art performance on benchmarks like LIBERO, particularly for long-horizon manipulation tasks.25
Human-Robot Collaboration (HRC)
Agentic AI is fostering more intuitive, adaptive, and safer human-robot collaboration.
- Shared Autonomy: This paradigm involves splitting agency between the human and the robot, where human input, often through natural language, can dynamically refine the robot's control space and behavior.8 The LILAC (Language-Informed Latent Actions with Corrections) framework is an example. It allows users to provide online natural language corrections (e.g., "move a bit to the right," "no, towards the red block") while the robot is executing a task. Language input is fed to a learned model that produces a meaningful, low-dimensional control space that the human can then use to guide the robot more precisely. A key advantage is that LILAC can learn to perform complex manipulation tasks from a small number of demonstrations (10-20) compared to the thousands often required by fully autonomous imitation or reinforcement learning approaches.30
- Safety in Physical HRC: Ensuring safety is paramount when humans and robots share a physical workspace. Artificial electronic skins covering robot bodies can make physical HRC safer. These skins, equipped with pressure sensors, can detect contact. By implementing adaptive sensitivity thresholds—dynamically adjusted based on factors like the robot's link velocity or effective mass—robots can maximize their operational speed and productivity while still guaranteeing safety according to standards like ISO/TS 15066 (which prescribes permissible forces and pressures during human contact).32 For instance, in a pick-and-place scenario, dynamic thresholds allow the robot to react appropriately to transient collisions, choosing to stop or avoid, and lead to higher productivity compared to conservative, static threshold settings. This is because dynamic thresholds reduce unnecessary stops for minor, harmless contacts.32
Multi-Robot Coordination
Agentic AI is fundamental for enabling systems of multiple robots to collaborate effectively on complex tasks that a single robot could not accomplish alone.
- Challenges: Key challenges include efficient and dynamic task allocation among heterogeneous agents, autonomous navigation in shared spaces, and achieving collective coordination in dynamic and potentially unpredictable environments.
- Agentic UAV Swarms: An exemplary application is the coordination of Unmanned Aerial Vehicle (UAV) swarms. In such systems:
- LLMs are used to interpret high-level human instructions (e.g., "patrol this area," "track that target") and decompose them into specific, executable commands or subtasks for individual UAVs or groups of UAVs.8
- A Coordination Field Mechanism can be employed to guide UAV motion and task selection. This approach models target points as attractive sources and obstacles as repulsive sources in a potential field. Robots then navigate by following the gradient of this field, enabling decentralized and adaptive allocation of emergent tasks, as well as real-time obstacle avoidance.21 This field-driven strategy offers a decentralized structure, promoting real-time adaptability and responsiveness to dynamically changing tasks and conditions, making it well-suited for complex urban scenarios.21 Furthermore, LLM-based Multi-Agent Systems (MAS) provide a general framework for groups of intelligent robotic agents to coordinate, share knowledge, distribute tasks, and align their efforts towards common objectives, effectively scaling their collective capabilities.15
The diverse applications in robotics underscore a clear trend: the increasing reliance on multi-modal understanding (integrating vision, language, and other sensory data) and multi-agent collaboration to address the complexities of real-world environments.14 This is driven by the fact that real-world robotic tasks are rarely unimodal or solvable by isolated agents. Effective operation necessitates sensing and acting upon diverse information types and often requires coordination with other entities, whether human or artificial. Consequently, future breakthroughs in agentic robotics will likely be heavily influenced by advancements in multimodal AI, robust sensor fusion techniques, and sophisticated multi-agent coordination and communication protocols.
Another significant development is the evolving role of LLMs in robotics. They are transitioning from being primarily NLP tools to serving as central components for high-level reasoning, strategic planning, and task decomposition.21 LLMs are, in effect, becoming the "brains" that bridge abstract human goals with concrete robotic actions, potentially lowering the barrier to programming complex robot behaviors and making robot instruction more intuitive.
Finally, to manage the inherent complexity and ensure the reliability of these sophisticated agentic robotic systems, there is a growing emphasis on developing structured frameworks and protocols. Examples like SAP in the Agentic Robot framework, the RAMSemCom communication framework, and Coordination Field mechanisms for UAVs illustrate this trend.14 As robotic tasks become more intricate, involve more agents, or span longer time horizons, unstructured approaches tend to suffer from error accumulation and unreliability.25 Therefore, the maturation of agentic robotics will likely involve the development and adoption of more standardized architectures, interaction protocols, and verification methods, akin to the design patterns and communication standards that underpin robust software engineering. This is crucial for building scalable, interpretable, and trustworthy robotic systems.
V. Navigating the Frontier: Challenges, Ethics, and the Road Ahead
While agentic AI holds immense promise, its journey towards widespread and robust deployment is accompanied by significant challenges, profound ethical considerations, and a rapidly evolving research landscape.
Current Challenges and Limitations
The practical implementation of agentic AI systems faces several hurdles:
- Data Dependency and Bias: The performance and reliability of agentic AI are critically dependent on the quality, diversity, and timeliness of the data they are trained on and interact with. Deficiencies in data can lead to biased decisions, inaccuracies, and outright errors.36 If training data reflects historical biases, agentic systems can perpetuate or even amplify these biases, leading to suboptimal or unfair outcomes, particularly in sensitive applications like customer service, financial loan assessment, or law enforcement.36 Robust data governance, including meticulous data collection, cleaning, and bias detection, is therefore a non-negotiable prerequisite for developing trustworthy agentic AI.
- Explainability and "Black-Box" Models (XAI): Many advanced AI models, especially deep learning systems and LLMs that often power agentic AI, operate as "black boxes." Their internal decision-making processes can be opaque, making it difficult to understand precisely how they arrive at specific conclusions or actions.36 This lack of transparency poses significant challenges for debugging, ensuring reliability, gaining user trust, and complying with regulatory requirements, especially in high-stakes fields like finance and healthcare.36 There is a pressing need for the development and adoption of Explainable AI (XAI) techniques (e.g., SHAP, LIME) that can provide human-understandable justifications for agent decisions.36 However, a persistent challenge is the accuracy-interpretability trade-off: often, the most accurate models are the most complex and hardest to explain, while simpler, more interpretable models might sacrifice some performance.41
- Scalability, Integration, and Computational Resources: Integrating agentic AI into existing enterprise ecosystems and scaling these systems effectively presents considerable operational and technical challenges.36 This involves not only the initial deployment but also ongoing efforts in retraining models, configuring integrations with legacy systems, and maintaining performance across dynamic and evolving workflows.36 Such endeavors require substantial investment in infrastructure and specialized technical expertise (e.g., AI engineers).7 Furthermore, the computational demands of agentic AI can be immense. As agents move towards more complex reasoning and learning, the need for compute resources can grow exponentially; some estimates suggest next-generation AI may require orders of magnitude more compute power than current models.2 Agentic AI systems are inherently computationally and energy-intensive, particularly during training (e.g., for RL) and large-scale deployment.45 This resource intensiveness could become a significant barrier to widespread adoption, potentially favoring larger organizations or necessitating breakthroughs in hardware efficiency and algorithmic optimization.
- Security Vulnerabilities and Loss of Control: The enhanced autonomy and connectivity of agentic AI systems make them attractive targets for malicious actors. They can be vulnerable to various attacks, including data poisoning, adversarial manipulation of inputs, or direct hijacking, leading to altered decision-making, disrupted processes, or compromised sensitive data.36 There is also the risk of "over-autonomy," where agents act beyond their intended boundaries due to unforeseen interactions or flawed goal specifications, or "behavioral drift," where agents deviate from their original programming over time as they adapt to new data or environmental changes.49 In critical scenarios, a loss of control over highly autonomous agents could lead to cascading failures with severe consequences.47 Therefore, robust security practices, comprehensive threat modeling, secure development lifecycles, and effective human-in-the-loop control and override mechanisms are vital.
- Complexity of Deployment and Development: Designing, developing, and deploying robust agentic AI systems is an inherently complex undertaking. It requires specialized expertise in various AI subfields, systems engineering, and domain-specific knowledge.7 Defining the goals, constraints, and operational parameters for an autonomous agent in a way that ensures safe and effective behavior across a wide range of potential situations is a non-trivial challenge, especially given that their autonomous nature can lead to diverse and sometimes unpredictable solution paths.1
These challenges are often deeply interconnected and can amplify one another. For instance, a lack of explainability in a "black-box" model makes it more difficult to detect inherent biases or to understand how a security vulnerability was exploited. Data dependencies can introduce biases, which, when acted upon by a highly autonomous agent, can lead to far-reaching and unfair consequences. The complexity of multi-agent systems can further exacerbate issues of accountability and make it harder to trace the root cause of errors or malicious actions. Addressing these challenges effectively requires a holistic, systems-thinking approach, where improvements in one area (like model accuracy) are considered in conjunction with others (like explainability, security, and ethical alignment).
Ethical Considerations
The rise of agentic AI brings a host of ethical considerations to the forefront:
- Accountability and Oversight: As agentic AI systems take on more decision-making responsibility, determining accountability when errors or harm occur becomes increasingly complex.36 The autonomous and often unpredictable nature of these systems can blur traditional lines of responsibility. Clear governance structures, well-defined chains of accountability involving developers, deployers, and users, and robust oversight mechanisms are essential to address this.37
- Privacy and Data Protection: Agentic AI systems frequently require access to vast amounts of data, often including sensitive personal or proprietary information, to function effectively. This raises significant concerns about data privacy, potential misuse of data, and the risk of data breaches.37 Compliance with data protection regulations (e.g., GDPR, CCPA) is paramount, and robust technical measures like encryption, anonymization, and secure data handling protocols are necessary.37
- Fairness and Bias Mitigation: Ensuring that agentic AI systems operate fairly and do not perpetuate or amplify existing societal biases is a critical ethical imperative. Proactive measures are needed, including regular bias audits of training data and model behavior, fostering diversity in development teams to identify potential blind spots, rigorous testing frameworks to ensure equitable outcomes across different demographic groups, and continuous monitoring and refinement of systems based on real-world performance.37
- Human-Centered Design and Job Displacement: Agentic AI should be designed to augment human capabilities and enhance human well-being, rather than simply replacing human workers.43 Maintaining meaningful human control over critical decisions and providing mechanisms for human intervention are key aspects of a human-centered approach.43 Nevertheless, the widespread adoption of advanced automation driven by agentic AI has the potential to displace jobs in certain sectors, particularly those involving routine or repetitive tasks.39 This necessitates proactive planning for workforce transitions, including investment in reskilling and upskilling programs, and the development of new career pathways that leverage unique human skills in collaboration with AI.39 Humans will likely transition to more strategic, supervisory, creative, and complex problem-solving roles.44
A fundamental tension exists between the drive to increase agent autonomy and capability, and the need to maintain human control, trust, and ethical alignment. The core value of agentic AI lies in its ability to operate with minimal oversight.3 However, this very autonomy fuels concerns about loss of control, accountability, and ethical decision-making.36 The emphasis on human-in-the-loop systems, explainability, and robust governance are all responses to this inherent tension.40 The future development of agentic AI will likely involve navigating this spectrum carefully, designing systems that are highly capable yet reliably controllable, and autonomous yet ethically aligned. This may lead to the deployment of varying levels of autonomy depending on the risk profile and criticality of specific applications.
Future Trends and Ongoing Research
The field of agentic AI is dynamic, with several key trends and research directions shaping its future:
- Evolution from Generative AI towards AGI?: Agentic AI is widely regarded as the next significant evolutionary step beyond Generative AI. It is characterized by much stronger reasoning and interaction capabilities, enabling more autonomous behavior to tackle complex, multi-step problems.1 Some researchers and observers posit that agentic AI, with its ability to learn, adapt, and act purposefully in diverse environments, represents a tangible step towards Artificial General Intelligence (AGI)—AI systems that can understand, learn, and apply knowledge across a wide range of tasks at a human-like level of competence.1
- Advancements in Multi-Agent Systems (MAS) and Collaboration: The future of agentic AI is increasingly seen as rooted in the orchestration of multiple, specialized AI agents that interact within cohesive frameworks. This approach aims to leverage the collective intelligence and diverse capabilities of various agents to solve complex, multi-faceted problems that would be intractable for a single agent.6 Active research includes developing taxonomies for MAS failures (such as MAST) to build more robust and reliable systems 58, and creating infrastructure for collaborative agentic AI, like the "Web of Agents" concept, which aims to foster interoperability between different agent ecosystems.56
- Next-Generation Agentic Reinforcement Learning: A significant area of research focuses on more deeply integrating LLMs and other advanced models with reinforcement learning. Frameworks like ARTIST (Agentic Reasoning and Tool Integration in Self-improving Transformers) are being developed to tightly couple agentic reasoning, RL, and dynamic tool integration for LLMs. ARTIST enables models to autonomously decide when, how, and which external tools (e.g., code interpreters, web search APIs) to invoke within multi-turn reasoning chains, learning these complex strategies through outcome-based RL without requiring step-by-step supervision.22 Other approaches, like the ML-Agent framework, explore LLM-based agents that learn through direct interactive experimentation on machine learning tasks using online RL.23 These advancements aim to make agents more adaptive, robust, and capable of sophisticated problem-solving.
- Emerging Capabilities in 2025-2026: In the near term, agentic AI is expected to gain its strongest foothold in structured task automation across various industries, including HR (e.g., resume parsing, candidate matching), customer service (e.g., automated issue resolution), and operational workflows in retail and other sectors.2 A significant increase in actual agentic AI deployments is anticipated, particularly in fields like cybersecurity, where agents will be used for autonomous network monitoring, threat detection, proactive security measures, and incident response.62 This will also likely spur the "AI-against-AI" paradigm, where agentic systems are used by both attackers and defenders.62
- Theoretical Frameworks for Agentic Systems: The development of agentic AI is supported by and contributes to various theoretical frameworks. Foundational models for sequential decision-making under uncertainty, such as Markov Decision Processes (MDPs) 18 and their extensions to Partially Observable MDPs (POMDPs) that handle situations where the agent has incomplete information about the true state of the environment 65, remain highly relevant. Architectures inspired by human practical reasoning, like the Belief-Desire-Intention (BDI) model, provide conceptual structures for designing agents with explicit mental states and goal-directed behavior.67 Concurrently, a rich ecosystem of practical AI agent development frameworks is emerging, including LangGraph, CrewAI, AutoGen, Swarm, FIPA standards, Agno, Atomic Agents, OpenAI Agents SDK, LlamaIndex, and Semantic Kernel, each offering different tools and abstractions for building and orchestrating agentic systems.69
The potential evolution towards AGI, if agentic AI is indeed a precursor, will likely be driven not merely by scaling existing models to be larger, but by fundamental breakthroughs in these core agentic capabilities. These include achieving robust and scalable multi-agent collaboration, enabling continuous and adaptive learning in open-ended and dynamic environments, and developing principled and flexible mechanisms for tool use and embodiment. These are qualitatively different challenges from those faced in traditional supervised or generative learning; they revolve around how intelligence is embodied, how it interacts with a complex world, how it learns from those interactions in context, and how it coordinates with other intelligent entities. Solving these "agentic" problems is key to unlocking more general and capable AI.
VI. Conclusion: The Agentic Revolution is Here
Agentic AI, characterized by its autonomy, goal-orientation, and sophisticated learning mechanisms, is undeniably poised to redefine numerous industries, with robotics being a prime example of its transformative potential. We are witnessing a significant shift from AI systems that primarily automate specific tasks or generate content to systems capable of complex problem-solving, dynamic adaptation, and orchestrated workflow execution with minimal human oversight. In robotics, this translates to machines that can navigate unpredictable environments, perform intricate manipulations, collaborate seamlessly with humans, and coordinate actions within multi-robot teams, pushing the boundaries of what automated systems can achieve.
However, this exciting frontier is not without its challenges. The development and deployment of agentic AI bring to the fore critical considerations regarding data bias, model explainability, system accountability, security vulnerabilities, and the societal impact of widespread automation, including potential job displacement. These are not trivial concerns and demand a proactive and thoughtful approach.
The path forward necessitates a strong emphasis on responsible development and the cultivation of robust human-AI collaboration. Human oversight, transparent governance frameworks, and a steadfast commitment to human-centered design principles are paramount to ensuring that agentic AI systems augment human capabilities and align with broadly accepted societal values and ethical standards.43 The future is unlikely to be one where humans are entirely replaced by autonomous agents; rather, it will likely involve synergistic human-AI teams. In such teams, AI agents will handle the complexities of data processing, task execution, and dynamic adaptation, while humans provide strategic direction, ethical guidance, oversight, and manage exceptions or situations requiring nuanced judgment.44
Agentic AI is no longer a purely futuristic concept; it is an actively evolving field with tangible applications and a trajectory of rapid advancement.2 Ongoing research into more sophisticated multi-agent systems, next-generation reinforcement learning techniques, and the deeper, more versatile integration of large language models into agentic architectures promises to unlock even more powerful capabilities. The potential for agentic AI to drive new levels of innovation, efficiency, and problem-solving capacity is immense. Realizing this potential in a beneficial and sustainable manner hinges on our collective ability to navigate its complexities and guide its development with wisdom, foresight, and a strong ethical compass.
Works cited
- arxiv.org, accessed June 1, 2025, https://arxiv.org/html/2504.18875v1
- Agentic AI: The New Frontier of Intelligence That Acts | WisdomTree, accessed June 1, 2025, https://www.wisdomtree.com/investments/blog/2025/04/21/agentic-ai-the-new-frontier-of-intelligence-that-acts
- What Does “Agentic” Mean? A Definitive Guide to the AI Buzzword ..., accessed June 1, 2025, https://www.moveworks.com/us/en/resources/blog/what-does-agentic-mean#:~:text=Agentic%20AI%20can%20make%20independent,and%20adaptability%20in%20dynamic%20environments.&text=Unlike%20basic%20ML%20algorithms%20that,set%20and%20pursue%20specific%20goals.
- What Is Agentic AI? | IBM, accessed June 1, 2025, https://www.ibm.com/think/topics/agentic-ai
- The Rise of Agentic AI: A Deep Dive - re:cinq, accessed June 1, 2025, https://blog.re-cinq.com/posts/agentic-ai/
- AI Agents vs. Agentic AI: A Conceptual Taxonomy, Applications and Challenges - arXiv, accessed June 1, 2025, https://arxiv.org/html/2505.10468v1
- What are AI Agents?- Agents in Artificial Intelligence Explained - AWS, accessed June 1, 2025, https://aws.amazon.com/what-is/ai-agents/
- arxiv.org, accessed June 1, 2025, https://arxiv.org/abs/2505.10468
- (PDF) AI Agents vs. Agentic AI: A Conceptual Taxonomy, Applications and Challenges, accessed June 1, 2025, https://www.researchgate.net/publication/391776617_AI_Agents_vs_Agentic_AI_A_Conceptual_Taxonomy_Applications_and_Challenges
- arxiv.org, accessed June 1, 2025, https://arxiv.org/html/2505.10468v4
- Your guide to agentic AI - AI Accelerator Institute, accessed June 1, 2025, https://www.aiacceleratorinstitute.com/your-guide-to-agentic-ai/
- What is Agentic AI? Understanding AI Agents & Automation | OpenText, accessed June 1, 2025, https://www.opentext.com/what-is/agentic-ai
- AI agents components and their role in autonomous decision-making, accessed June 1, 2025, https://toloka.ai/blog/ai-agents-components-and-their-role-in-autonomous-decision-making/
- Wireless Agentic AI with Retrieval-Augmented Multimodal Semantic Perception - arXiv, accessed June 1, 2025, https://arxiv.org/html/2505.23275v1
- Multi-Agent Collaboration Mechanisms: A Survey of LLMs - arXiv, accessed June 1, 2025, https://arxiv.org/html/2501.06322v1
- What is AI Agent Learning? | IBM, accessed June 1, 2025, https://www.ibm.com/think/topics/ai-agent-learning
- What is Reinforcement Learning and How Does it Works?, accessed June 1, 2025, https://www.analyticsvidhya.com/blog/2021/02/introduction-to-reinforcement-learning-for-beginners/
- Reinforcement learning - Wikipedia, accessed June 1, 2025, https://en.wikipedia.org/wiki/Reinforcement_learning
- ojs.aaai.org, accessed June 1, 2025, https://ojs.aaai.org/index.php/AAAI/article/view/35095/37250
- journal.pandawan.id, accessed June 1, 2025, https://journal.pandawan.id/italic/article/download/661/488/3415
- Coordination Field for Agentic UAV Task Allocation In Low-altitude Urban Scenarios - arXiv, accessed June 1, 2025, https://arxiv.org/html/2505.00091v3
- arxiv.org, accessed June 1, 2025, https://arxiv.org/pdf/2502.11433
- ML-Agent: Reinforcing LLM Agents for Autonomous Machine Learning Engineering - arXiv, accessed June 1, 2025, https://arxiv.org/abs/2505.23723
- arxiv.org, accessed June 1, 2025, http://arxiv.org/pdf/2505.01441
- A Brain-Inspired Framework for Vision-Language-Action Models in Embodied Agents - arXiv, accessed June 1, 2025, https://arxiv.org/html/2505.23450v1
- (PDF) Reinforcement Learning in Robotics: A Survey - ResearchGate, accessed June 1, 2025, https://www.researchgate.net/publication/258140920_Reinforcement_Learning_in_Robotics_A_Survey
- MIT Review Explores Breakthroughs in Robotic Perception, accessed June 1, 2025, https://www.azorobotics.com/News.aspx?newsID=15689
- Wireless Agentic AI with Retrieval-Augmented Multimodal Semantic Perception - arXiv, accessed June 1, 2025, https://www.arxiv.org/pdf/2505.23275
- [2505.23450] Agentic Robot: A Brain-Inspired Framework for Vision-Language-Action Models in Embodied Agents - arXiv, accessed June 1, 2025, https://www.arxiv.org/abs/2505.23450
- [2301.02555] "No, to the Right" -- Online Language Corrections for Robotic Manipulation via Shared Autonomy - arXiv, accessed June 1, 2025, https://arxiv.org/abs/2301.02555
- Online Language Corrections for Robotic Manipulation via Shared Autonomy - Stanford NLP Group, accessed June 1, 2025, https://www-nlp.stanford.edu/pubs/cui2023lilac.pdf
- www.arxiv.org, accessed June 1, 2025, https://www.arxiv.org/pdf/2409.06369v1
- CoordField: Coordination Field for Agentic UAV Task Allocation In Low-altitude Urban Scenarios - arXiv, accessed June 1, 2025, https://arxiv.org/pdf/2505.00091
- [2505.00091] CoordField: Coordination Field for Agentic UAV Task Allocation In Low-altitude Urban Scenarios - arXiv, accessed June 1, 2025, https://www.arxiv.org/abs/2505.00091
- [2501.06322] Multi-Agent Collaboration Mechanisms: A Survey of LLMs - arXiv, accessed June 1, 2025, https://arxiv.org/abs/2501.06322
- Agentic AI: A Promising Evolution, But Not Without Limits - Blueprint, accessed June 1, 2025, https://www.blueprintsys.com/blog/agentic-ai-a-promising-evolution-but-not-without-limits
- Ethical Implications of Agentic AI in Financial Services - SG Analytics, accessed June 1, 2025, https://www.sganalytics.com/blog/agentic-ai-in-financial-services/
- Ethical Considerations in Deploying Agentic AI for AML Compliance - Lucinity, accessed June 1, 2025, https://lucinity.com/blog/ethical-considerations-in-deploying-agentic-ai-for-aml-compliance
- How Agentic AI is Reshaping the Global South: Opportunities & Risks - Modern Diplomacy, accessed June 1, 2025, https://moderndiplomacy.eu/2025/03/09/how-agentic-ai-is-reshaping-the-global-south-opportunities-risks/
- Why Explainable AI in Banking and Finance Is Critical for Compliance - Lumenova AI, accessed June 1, 2025, https://www.lumenova.ai/blog/ai-banking-finance-compliance/
- Explainable AI in Robotics: Building Trust and Transparency in Autonomous Systems, accessed June 1, 2025, https://smythos.com/ai-agents/agent-architectures/explainable-ai-in-robotics/
- Financial Fraud Detection Using Explainable AI and Stacking ... - arXiv, accessed June 1, 2025, https://arxiv.org/abs/2505.10050
- Ethical Considerations of Agentic AI - ProcessMaker, accessed June 1, 2025, https://www.processmaker.com/blog/ethical-considerations-of-agentic-ai/
- Navigating the Challenges: 5 Common Pitfalls in Agentic AI Adoption - CapTech, accessed June 1, 2025, https://www.captechconsulting.com/articles/navigating-the-challenges-5-common-pitfalls-in-agentic-ai-adoption
- (PDF) Agentic AI: Autonomous Intelligence for Complex Goals – A ..., accessed June 1, 2025, https://www.researchgate.net/publication/388313991_Agentic_AI_Autonomous_Intelligence_for_Complex_Goals_-_A_Comprehensive_Survey
- Securing Agentic AI: A Comprehensive Threat Model and Mitigation Framework for Generative AI Agents - arXiv, accessed June 1, 2025, https://arxiv.org/html/2504.19956v1
- Top 7 Concerns of Technology Leaders That Implemented Agentic AI - CTO Academy, accessed June 1, 2025, https://cto.academy/agentic-ai-concerns-and-solutions/
- The Rise of Agentic AI: Implications, Concerns, and the Path Forward, accessed June 1, 2025, https://www.computer.org/csdl/magazine/ex/2025/02/10962241/25NBj1aT8Zi
- Are AI Agents Safe? Potential Risks and Challenges - Springs, accessed June 1, 2025, https://springsapps.com/knowledge/are-ai-agents-safe-potential-risks-and-challenges
- Agentic AI Explained: Key Features, Benefits, and Real-World Impact | Splunk, accessed June 1, 2025, https://www.splunk.com/en_us/blog/learn/agentic-ai.html
- Reimagining Investment Portfolio Management with Agentic AI ..., accessed June 1, 2025, https://www.mongodb.com/blog/post/innovation/reimagining-investment-portfolio-management-with-agentic-ai
- Generative to Agentic AI: Survey, Conceptualization, and Challenges - arXiv, accessed June 1, 2025, https://arxiv.org/html/2504.18875
- Agentic AI Systems Applied to tasks in Financial Services: Modeling and model risk management crews - arXiv, accessed June 1, 2025, https://arxiv.org/html/2502.05439v2
- Towards Agentic AI Networking in 6G: A Generative Foundation Model-as-Agent Approach, accessed June 1, 2025, https://arxiv.org/html/2503.15764v2
- Scaling Laws of Scientific Discovery with AI and Robot Scientists - arXiv, accessed June 1, 2025, https://arxiv.org/html/2503.22444v1
- Collaborative Agentic AI Needs Interoperability Across Ecosystems - arXiv, accessed June 1, 2025, https://arxiv.org/html/2505.21550v1
- From Autonomous Agents to Integrated Systems, A New Paradigm: Orchestrated Distributed Intelligence - arXiv, accessed June 1, 2025, https://arxiv.org/html/2503.13754v2
- arxiv.org, accessed June 1, 2025, https://arxiv.org/pdf/2503.13657
- [2503.13657] Why Do Multi-Agent LLM Systems Fail? - arXiv, accessed June 1, 2025, https://arxiv.org/abs/2503.13657
- What Are Multi-Agent Systems in AI? - Multimodal, accessed June 1, 2025, https://www.multimodal.dev/post/what-are-multi-agent-systems-in-ai
- Multi-Agent Systems: Building the Autonomous Enterprise, accessed June 1, 2025, https://www.automationanywhere.com/rpa/multi-agent-systems
- Experts Reveal How Agentic AI Is Shaping Cybersecurity in 2025, accessed June 1, 2025, https://www.cybersecuritytribe.com/articles/how-agentic-ai-is-shaping-cybersecurity-in-2025
- arxiv.org, accessed June 1, 2025, https://arxiv.org/abs/2411.11451
- Markov Decision Process (MDP) - Iterate.ai, accessed June 1, 2025, https://www.iterate.ai/ai-glossary/mdp-markov-decision-process-explained
- Partially observable Markov decision process - Wikipedia, accessed June 1, 2025, https://en.wikipedia.org/wiki/Partially_observable_Markov_decision_process
- What is Partially Observable MDP (POMDP) - Activeloop, accessed June 1, 2025, https://www.activeloop.ai/resources/glossary/partially-observable-mdp-pomdp/
- Belief–desire–intention software model - Wikipedia, accessed June 1, 2025, https://en.wikipedia.org/wiki/Belief%E2%80%93desire%E2%80%93intention_software_model
- What is the belief-desire-intention (BDI) agent model? — Klu, accessed June 1, 2025, https://klu.ai/glossary/belief-desire-intention-agent-model
- Agentic Frameworks: A Guide to the Systems Used to Build AI ..., accessed June 1, 2025, https://www.moveworks.com/us/en/resources/blog/what-is-agentic-framework
- AI Agent Frameworks: Tools for Smarter Systems | DataCamp, accessed June 1, 2025, https://www.datacamp.com/blog/ai-agent-frameworks
- Agentic trading: Where ethics, risk, and alpha collide - KX, accessed June 1, 2025, https://kx.com/blog/agentic-trading-where-ethics-risk-and-alpha-collide/