To save content items to your account,
please confirm that you agree to abide by our usage policies.
If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account.
Find out more about saving content to .
To save content items to your Kindle, first ensure no-reply@cambridge.org
is added to your Approved Personal Document E-mail List under your Personal Document Settings
on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part
of your Kindle email address below.
Find out more about saving to your Kindle.
Note you can select to save to either the @free.kindle.com or @kindle.com variations.
‘@free.kindle.com’ emails are free but can only be saved to your device when it is connected to wi-fi.
‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.
In the past, techniques for natural language translation were not very relevant for acting and planning systems. However, with the recent advent of large language models and their various multimodal extensions into foundation models, this is no longer the case. This last part introduces large language models and their potential benefits in acting, planning, and learning. It discusses the perceiving, monitoring, and goal reasoning functions for deliberation.
Learning to act with probabilistic models is the area of reinforcement learning (RL), the topic of this chapter. RL in some ways parallels the adaptation mechanisms of natural beings to their environment, relying on feedback mechanisms and extending the homeostasis regulations to complex behaviors. With continual learning, an actor can cope with a continually changing environment.This chapter first introduces the main principles of reinforcement learning. It presents a simple Q-learning RL algorithm. It shows how to generalize a learned relation with a parametric representation. it introduces neural network methods, which play a major in learning and are needed for deep RL (Section 10.5) and policy-based RL (Section 10.6). The issues of aided reinforcement learning with shaped rewards, imitation learning, and inverse reinforcement learning are addressed next. Section 10.8 is about probabilistic planning and RL.
This part of the book is devoted to acting, planning, and learning with operational models of actions expressed with a hierarchical task-oriented representation. Operational models are valuable for acting. They allow for detailed descriptions of complex actions handling dynamic environments with exogenous events. The representation relies on hierarchical refinement methods that describe alternative ways to handle tasks and react to events. A method can be any complex algorithm, decomposing a task into subtasks and primitive actions. Subtasks are refined recursively. Actions trigger the execution of sensory-motor procedures in closed loops that query and change the world stochastically.
This paper elaborates the design and analysis of cross-aperture-coupled twin port ceramic radiator. Stimulation of alumina ceramic using a cross slot helps to produce circular waves within 7.35–7.8 GHz. The polarization diversity concept helps to improve the separation level by above 25 dB. Loading of double negative unit cell made metasurface (MS) improves the antenna gain over 11.5 dBi within the working spectrum. Machine learning (ML) techniques, i.e. Decision Tree and Random Forest are utilized to predict the |S11|/Axial ratio parameters. Experimental verification/ML prediction and optimized simulated consequences confirm that the structured radiator works efficiently between 7.21 and 8.2 GHz with over 25 dB isolation between the ports. Directive pattern and decent values of (MIMO) parameters make the radiator applicable for the 6G communication system.
Task and motion planning (TAMP) problems combine abstract causal relations from preconditions to effects with computational geometry, kinematics, and dynamics. This chapter is about the integration of planning for motion/manipulation with planning for abstract actions. It introduces the main sampling-based algorithms for motion planning. Manipulation planning is subsequently introduced. A few approaches specific to TAMP are then presented.
This chapter is about planning approaches with explicit time in the descriptive and operational models of actions, as well as in the models of the expected evolution of the world not caused by the actor. It describes a planning algorithm that handles durative and concurrent activities with respect to a predicted dynamics. Section 17.1 presents a knowledge representation for modeling actions and tasks with temporal variables using temporal refinement methods. Temporal plans and planning problems are defined as chronicles, i.e., collections of assertions and tasks with explicit temporal constraints. A planning algorithm with temporal refinement methods is developed in Section 17.2. The basic techniques for managing temporal and domain constraints are then presented in Section 17.3.
Local eddy viscosity and diffusivity models are widely used to understand and predict turbulent flows. However, the local approximations in space and time are not always valid for actual turbulent flows. Recently, a non-local eddy diffusivity model for turbulent scalar flux was proposed to improve the local model and was validated using direct numerical simulation (DNS) of homogeneous isotropic turbulence with an inhomogeneous mean scalar (Hamba 2022 J. Fluid Mech.950, A38). The model was modified using the scale-space energy density in preparation for application to inhomogeneous turbulence (Hamba 2023 J. Fluid Mech.977, A11). In this paper, the model is further improved by incorporating the effects of turbulence anisotropy, inhomogeneity and wall boundaries. The needed inputs from the flow to evaluate the model are the Reynolds stress and the energy dissipation rate. With the improved model, one- and two-dimensional profiles ofthe non-local eddy diffusivity in turbulent channel flow are evaluated and compared with the exact DNS values. The DNS results reveal a contribution to the scalar flux from the mean scalar gradient in a wide upstream region. Additionally, the temporal profile of the non-local eddy diffusivity moves downstream, diffuses anisotropically and is tilted towards the bottom wall. The model reproduces this behaviour of mean flow convection and anisotropic turbulent diffusion well. These results indicate that the non-local eddy diffusivity model is useful for gaining insights into scalar transport in inhomogeneous turbulence.
This chapter is about two key aspects of learning with deterministic models: learning heuristics to speed up the search for a solution plan and the automated synthesis of the model itself. We discuss how to learn heuristics for exploring parts of the search space that are more likely to lead to solutions. We then address the problem of how to learn a deterministic model, with a focus on learning action schemas.