Frontiers of intelligent control and information processing

This review volume encapsulates the state-of-art pioneering works in the development of intelligent control systems. Proposition and evocations of each solution is backed up with evidences from applications, could be used as references for the consideration of decision support and communication comp...

Full description

Saved in:
Bibliographic Details
Other Authors: Liu, Derong 1963- (Editor), Alippi, Cesare (Editor), Zhao, Dongbin (Editor), Huaguang, Zhang (Editor)
Format: Book
Language:English
Published: Hackensack, NJ World Scientific [2014]
Subjects:
Tags: Add Tag
No Tags, Be the first to tag this record!

MARC

LEADER 00000nam a2200000 i 4500
001 98130
003 MY-KLNDU
005 20241220022422.0
008 221104t20142015njua b 000 0 eng d
020 |a 9789814616874 
039 9 |a 202211041132  |b VLOAD  |c 201601141528  |d faezah  |c 201601141527  |d faezah  |y 201512081635  |z syarifuddin 
040 |a UPNM  |b eng  |c UPNM  |e rda 
090 |a TJ 216  |b .F766 2014 
245 0 0 |a Frontiers of intelligent control and information processing  |c edited by Derong Liu, Cesare Alippi, Dongbin Zhao, Huaguang Zhang 
264 1 |a Hackensack, NJ  |b World Scientific  |c [2014] 
264 4 |c © 2015 
300 |a xvii, 461 pages  |b illustrations  |c 25 cm 
336 |a text  |2 rdacontent 
337 |a unmediated  |2 rdamedia 
338 |a volume  |2 rdacarrier 
504 |a Includes bibliographical references 
505 0 |a 1. Dynamic Graphical Games: Online Adaptive Learning Solutions Using Approximate Dynamic Programming-- 1.1 Introduction -- 1.2 Graphs and Synchronization of Multi-Agent Dynamical Systems ;1.2.1 Graphs; 1.2.2 Synchronization and tracking error dynamics --1.3 Multiple Player CooperativeGames on Graphs; 1.3.1 Graphical games; 1.3.2 Comparison of graphical games with standard dynamic games; 1.3.3 Nash equilibrium for graphical games; 1.3.4 Hamiltonian equation for dynamic graphical games; 1.3.5 Bellman equation for dynamic graphical games. 1.3.6 Discrete Hamilton-Jacobi theory: Equivalence of Bellman and discrete-time Hamilton Jacobi equations1.3.7 Stability and Nash solution of the graphical games --1.4 Approximate Dynamic Programming for Graphical Games; 1.4.1 Heuristic dynamic programming for graphical games; 1.4.2 Dual heuristic programming for graphical games --1.5 Coupled Riccati Recursions; 1.6 Graphical Game Solutions by Actor-Critic Learning; 1.6.1 Actor-critic networks and tuning; 1.6.2 Actor-critic offline tuning with exploration; 1.6.3 Actor-critic online tuning in real-time. 1.7 Graphical Game Example and Simulation Results1.7.1 Riccati recursion offline solution; 1.7.2 Simulation results using offline actor-critic tuning; 1.7.3 Simulation results using online actor-critic tuning; 1.8 Conclusions; Acknowledgement; Reference -- 2. Reinforcement-Learning-Based Online Learning Control for Discrete-Time Unknown Nonaffine Nonlinear Systems; 2.1 Introduction; 2.2 Problem Statement and Preliminaries; 2.2.1 Dynamics of nonaffine nonlinear discrete-time systems; 2.2.2 A single-hidden layer neural network; 2.3 Controller Design via Reinforcement Learning. 2.3.1 A basic controller design approach2.3.2 Critic neural network and weight update law; 2.3.3 Action neural network and weight update law; 2.4 Stability Analysis and Performance of the Closed-Loop System; 2.5 Numerical Examples; 2.5.1 Example 1; 2.5.2 Example 2; 2.6 Conclusions; Acknowledgement; References --3. Experimental Studies on Data-Driven Heuristic Dynamic Programming for POMDP; 3.1 Introduction; 3.2 Markov Decision Process and Partially Observable Markov Decision Process; 3.2.1 Markov decision process; 3.2.2 Partially observable Markov decision process. 3.3 Problem Formulation with the State Estimator3.4 Data-Driven HDP Algorithm for POMDP; 3.4.1 Learning in the state estimator network; 3.4.2 Learning in the critic and the action network; 3.5 Simulation Study; 3.5.1 Case study one; 3.5.2 Case study two; 3.5.3 Case study three; 3.6 Conclusions and Discussion; Acknowledgement; References --4. Online Reinforcement Learning for Continuous-State Systems; 4.1 Introduction; 4.2 Background of Reinforcement Learning; 4.3 RLSPI Algorithm; 4.3.1 Policy iteration; 4.3.2 RLSPI; 4.4 Examples of RLSPI; 4.4.1 Linear discrete-time system 
520 |a This review volume encapsulates the state-of-art pioneering works in the development of intelligent control systems. Proposition and evocations of each solution is backed up with evidences from applications, could be used as references for the consideration of decision support and communication components required for today intelligent control systems 
592 |a JI 4860  |b 05/01/2016  |c RM 681.50  |h JENDELA INFORMASI 
650 0 |a Automatic control 
650 0 |a Information technology 
700 1 |a Liu, Derong  |d 1963-  |e editor 
700 1 |a Alippi, Cesare  |e editor 
700 1 |a Zhao, Dongbin  |e editor 
700 1 |a Huaguang, Zhang  |e editor 
999 |a vtls000055778  |c 98130  |d 98130