journal
MENU ▼
Read by QxMD icon Read
search

IEEE Transactions on Neural Networks and Learning Systems

journal
https://www.readbyqxmd.com/read/29771677/learning-based-adaptive-optimal-tracking-control-of-strict-feedback-nonlinear-systems
#1
Weinan Gao, Zhong-Ping Jiang
This paper proposes a novel data-driven control approach to address the problem of adaptive optimal tracking for a class of nonlinear systems taking the strict-feedback form. Adaptive dynamic programming (ADP) and nonlinear output regulation theories are integrated for the first time to compute an adaptive near-optimal tracker without any a priori knowledge of the system dynamics. Fundamentally different from adaptive optimal stabilization problems, the solution to a Hamilton-Jacobi-Bellman (HJB) equation, not necessarily a positive definite function, cannot be approximated through the existing iterative methods...
June 2018: IEEE Transactions on Neural Networks and Learning Systems
https://www.readbyqxmd.com/read/29771676/extreme-trust-region-policy-optimization-for-active-object-recognition
#2
Huaping Liu, Yupei Wu, Fuchun Sun
In this brief, we develop a deep reinforcement learning method to actively recognize objects by choosing a sequence of actions for an active camera that helps to discriminate between the objects. The method is realized using trust region policy optimization, in which the policy is realized by an extreme learning machine and, therefore, leads to efficient optimization algorithm. The experimental results on the publicly available data set show the advantages of the developed extreme trust region optimization method...
June 2018: IEEE Transactions on Neural Networks and Learning Systems
https://www.readbyqxmd.com/read/29771675/action-driven-visual-object-tracking-with-deep-reinforcement-learning
#3
Sangdoo Yun, Jongwon Choi, Youngjoon Yoo, Kimin Yun, Jin Young Choi
In this paper, we propose an efficient visual tracker, which directly captures a bounding box containing the target object in a video by means of sequential actions learned using deep neural networks. The proposed deep neural network to control tracking actions is pretrained using various training video sequences and fine-tuned during actual tracking for online adaptation to a change of target and background. The pretraining is done by utilizing deep reinforcement learning (RL) as well as supervised learning...
June 2018: IEEE Transactions on Neural Networks and Learning Systems
https://www.readbyqxmd.com/read/29771674/multisource-transfer-double-dqn-based-on-actor-learning
#4
Jie Pan, Xuesong Wang, Yuhu Cheng, Qiang Yu
Deep reinforcement learning (RL) comprehensively uses the psychological mechanisms of "trial and error" and "reward and punishment" in RL as well as powerful feature expression and nonlinear mapping in deep learning. Currently, it plays an essential role in the fields of artificial intelligence and machine learning. Since an RL agent needs to constantly interact with its surroundings, the deep Q network (DQN) is inevitably faced with the need to learn numerous network parameters, which results in low learning efficiency...
June 2018: IEEE Transactions on Neural Networks and Learning Systems
https://www.readbyqxmd.com/read/29771673/self-paced-prioritized-curriculum-learning-with-coverage-penalty-in-deep-reinforcement-learning
#5
Zhipeng Ren, Daoyi Dong, Huaxiong Li, Chunlin Chen
In this paper, a new training paradigm is proposed for deep reinforcement learning using self-paced prioritized curriculum learning with coverage penalty. The proposed deep curriculum reinforcement learning (DCRL) takes the most advantage of experience replay by adaptively selecting appropriate transitions from replay memory based on the complexity of each transition. The criteria of complexity in DCRL consist of self-paced priority as well as coverage penalty. The self-paced priority reflects the relationship between the temporal-difference error and the difficulty of the current curriculum for sample efficiency...
June 2018: IEEE Transactions on Neural Networks and Learning Systems
https://www.readbyqxmd.com/read/29771672/reusable-reinforcement-learning-via-shallow-trails
#6
Yang Yu, Shi-Yong Chen, Qing Da, Zhi-Hua Zhou
Reinforcement learning has shown great success in helping learning agents accomplish tasks autonomously from environment interactions. Meanwhile in many real-world applications, an agent needs to accomplish not only a fixed task but also a range of tasks. For this goal, an agent can learn a metapolicy over a set of training tasks that are drawn from an underlying distribution. By maximizing the total reward summed over all the training tasks, the metapolicy can then be reused in accomplishing test tasks from the same distribution...
June 2018: IEEE Transactions on Neural Networks and Learning Systems
https://www.readbyqxmd.com/read/29771671/distributed-economic-dispatch-in-microgrids-based-on-cooperative-reinforcement-learning
#7
Weirong Liu, Peng Zhuang, Hao Liang, Jun Peng, Zhiwu Huang
Microgrids incorporated with distributed generation (DG) units and energy storage (ES) devices are expected to play more and more important roles in the future power systems. Yet, achieving efficient distributed economic dispatch in microgrids is a challenging issue due to the randomness and nonlinear characteristics of DG units and loads. This paper proposes a cooperative reinforcement learning algorithm for distributed economic dispatch in microgrids. Utilizing the learning algorithm can avoid the difficulty of stochastic modeling and high computational complexity...
June 2018: IEEE Transactions on Neural Networks and Learning Systems
https://www.readbyqxmd.com/read/29771670/optimal-fault-tolerant-control-for-discrete-time-nonlinear-strict-feedback-systems-based-on-adaptive-critic-design
#8
Zhanshan Wang, Lei Liu, Yanming Wu, Huaguang Zhang
This paper investigates the problem of optimal fault-tolerant control (FTC) for a class of unknown nonlinear discrete-time systems with actuator fault in the framework of adaptive critic design (ACD). A pivotal highlight is the adaptive auxiliary signal of the actuator fault, which is designed to offset the effect of the fault. The considered systems are in strict-feedback forms and involve unknown nonlinear functions, which will result in the causal problem. To solve this problem, the original nonlinear systems are transformed into a novel system by employing the diffeomorphism theory...
June 2018: IEEE Transactions on Neural Networks and Learning Systems
https://www.readbyqxmd.com/read/29771669/suboptimal-scheduling-in-switched-systems-with-continuous-time-dynamics-a-least-squares-approach
#9
Tohid Sardarmehni, Ali Heydari
Two approximate solutions for optimal control of switched systems with autonomous subsystems and continuous-time dynamics are presented. The first solution formulates a policy iteration (PI) algorithm for the switched systems with recursive least squares. To reduce the computational burden imposed by the PI algorithm, a second solution, called single loop PI, is presented. Online and concurrent training algorithms are discussed for implementing each solution. At last, effectiveness of the presented algorithms is evaluated through numerical simulations...
June 2018: IEEE Transactions on Neural Networks and Learning Systems
https://www.readbyqxmd.com/read/29771668/approximate-dynamic-programming-combining-regional-and-local-state-following-approximations
#10
Patryk Deptula, Joel A Rosenfeld, Rushikesh Kamalapurkar, Warren E Dixon
An infinite-horizon optimal regulation problem for a control-affine deterministic system is solved online using a local state following (StaF) kernel and a regional model-based reinforcement learning (R-MBRL) method to approximate the value function. Unlike traditional methods such as R-MBRL that aim to approximate the value function over a large compact set, the StaF kernel approach aims to approximate the value function in a local neighborhood of the state that travels within a compact set. In this paper, the value function is approximated using a state-dependent convex combination of the StaF-based and the R-MBRL-based approximations...
June 2018: IEEE Transactions on Neural Networks and Learning Systems
https://www.readbyqxmd.com/read/29771667/leader-follower-output-synchronization-of-linear-heterogeneous-systems-with-active-leader-using-reinforcement-learning
#11
Yongliang Yang, Hamidreza Modares, Donald C Wunsch, Yixin Yin
This paper develops optimal control protocols for the distributed output synchronization problem of leader-follower multiagent systems with an active leader. Agents are assumed to be heterogeneous with different dynamics and dimensions. The desired trajectory is assumed to be preplanned and is generated by the leader. Other follower agents autonomously synchronize to the leader by interacting with each other using a communication network. The leader is assumed to be active in the sense that it has a nonzero control input so that it can act independently and update its control to keep the followers away from possible danger...
June 2018: IEEE Transactions on Neural Networks and Learning Systems
https://www.readbyqxmd.com/read/29771666/robust-adp-design-for-continuous-time-nonlinear-systems-with-output-constraints
#12
Bo Fan, Qinmin Yang, Xiaoyu Tang, Youxian Sun
In this paper, a novel robust adaptive dynamic programming (RADP)-based control strategy is presented for the optimal control of a class of output-constrained continuous-time unknown nonlinear systems. Our contribution includes a step forward beyond the usual optimal control result to show that the output of the plant is always within user-defined bounds. To achieve the new results, an error transformation technique is first established to generate an equivalent nonlinear system, whose asymptotic stability guarantees both the asymptotic stability and the satisfaction of the output restriction of the original system...
June 2018: IEEE Transactions on Neural Networks and Learning Systems
https://www.readbyqxmd.com/read/29771665/optimal-guaranteed-cost-sliding-mode-control-for-constrained-input-nonlinear-systems-with-matched-and-unmatched-disturbances
#13
Huaguang Zhang, Qiuxia Qu, Geyang Xiao, Yang Cui
Based on integral sliding mode and approximate dynamic programming (ADP) theory, a novel optimal guaranteed cost sliding mode control is designed for constrained-input nonlinear systems with matched and unmatched disturbances. When the system moves on the sliding surface, the optimal guaranteed cost control problem of sliding mode dynamics is transformed into the optimal control problem of a reformulated auxiliary system with a modified cost function. The ADP algorithm based on single critic neural network (NN) is applied to obtain the approximate optimal control law for the auxiliary system...
June 2018: IEEE Transactions on Neural Networks and Learning Systems
https://www.readbyqxmd.com/read/29771664/guided-policy-exploration-for-markov-decision-processes-using-an-uncertainty-based-value-of-information-criterion
#14
Isaac J Sledge, Matthew S Emigh, Jose C Principe
Reinforcement learning in environments with many action-state pairs is challenging. The issue is the number of episodes needed to thoroughly search the policy space. Most conventional heuristics address this search problem in a stochastic manner. This can leave large portions of the policy space unvisited during the early training stages. In this paper, we propose an uncertainty-based, information-theoretic approach for performing guided stochastic searches that more effectively cover the policy space. Our approach is based on the value of information, a criterion that provides the optimal tradeoff between expected costs and the granularity of the search process...
June 2018: IEEE Transactions on Neural Networks and Learning Systems
https://www.readbyqxmd.com/read/29771663/applications-of-deep-learning-and-reinforcement-learning-to-biological-data
#15
Mufti Mahmud, Mohammed Shamim Kaiser, Amir Hussain, Stefano Vassanelli
Rapid advances in hardware-based technologies during the past decades have opened up new possibilities for life scientists to gather multimodal data in various application domains, such as omics, bioimaging, medical imaging, and (brain/body)-machine interfaces. These have generated novel opportunities for development of dedicated data-intensive machine learning techniques. In particular, recent research in deep learning (DL), reinforcement learning (RL), and their combination (deep RL) promise to revolutionize the future of artificial intelligence...
June 2018: IEEE Transactions on Neural Networks and Learning Systems
https://www.readbyqxmd.com/read/29771662/optimal-and-autonomous-control-using-reinforcement-learning-a-survey
#16
Bahare Kiumarsi, Kyriakos G Vamvoudakis, Hamidreza Modares, Frank L Lewis
This paper reviews the current state of the art on reinforcement learning (RL)-based feedback control solutions to optimal regulation and tracking of single and multiagent systems. Existing RL solutions to both optimal and control problems, as well as graphical games, will be reviewed. RL methods learn the solution to optimal control and game problems online and using measured data along the system trajectories. We discuss Q-learning and the integral RL algorithm as core algorithms for discrete-time (DT) and continuous-time (CT) systems, respectively...
June 2018: IEEE Transactions on Neural Networks and Learning Systems
https://www.readbyqxmd.com/read/28103560/identifying-objective-and-subjective-words-via-topic-modeling
#17
Hanqi Wang, Fei Wu, Weiming Lu, Yi Yang, Xi Li, Xuelong Li, Yueting Zhuang
It is observed that distinct words in a given document have either strong or weak ability in delivering facts (i.e., the objective sense) or expressing opinions (i.e., the subjective sense) depending on the topics they associate with. Motivated by the intuitive assumption that different words have varying degree of discriminative power in delivering the objective sense or the subjective sense with respect to their assigned topics, a model named as dentified bjective- ubjective latent Dirichlet allocation (LDA) ( osLDA) is proposed in this paper...
March 2018: IEEE Transactions on Neural Networks and Learning Systems
https://www.readbyqxmd.com/read/28092580/determination-of-the-edge-of-criticality-in-echo-state-networks-through-fisher-information-maximization
#18
Lorenzo Livi, Filippo Maria Bianchi, Cesare Alippi
It is a widely accepted fact that the computational capability of recurrent neural networks (RNNs) is maximized on the so-called "edge of criticality." Once the network operates in this configuration, it performs efficiently on a specific application both in terms of: 1) low prediction error and 2) high short-term memory capacity. Since the behavior of recurrent networks is strongly influenced by the particular input signal driving the dynamics, a universal, application-independent method for determining the edge of criticality is still missing...
March 2018: IEEE Transactions on Neural Networks and Learning Systems
https://www.readbyqxmd.com/read/28092579/cooperative-adaptive-output-regulation-for-second-order-nonlinear-multiagent-systems-with-jointly-connected-switching-networks
#19
Wei Liu, Jie Huang
This paper studies the cooperative global robust output regulation problem for a class of heterogeneous second-order nonlinear uncertain multiagent systems with jointly connected switching networks. The main contributions consist of the following three aspects. First, we generalize the result of the adaptive distributed observer from undirected jointly connected switching networks to directed jointly connected switching networks. Second, by performing a new coordinate and input transformation, we convert our problem into the cooperative global robust stabilization problem of a more complex augmented system via the distributed internal model principle...
March 2018: IEEE Transactions on Neural Networks and Learning Systems
https://www.readbyqxmd.com/read/28092578/experienced-gray-wolf-optimization-through-reinforcement-learning-and-neural-networks
#20
E Emary, Hossam M Zawbaa, Crina Grosan
In this paper, a variant of gray wolf optimization (GWO) that uses reinforcement learning principles combined with neural networks to enhance the performance is proposed. The aim is to overcome, by reinforced learning, the common challenge of setting the right parameters for the algorithm. In GWO, a single parameter is used to control the exploration/exploitation rate, which influences the performance of the algorithm. Rather than using a global way to change this parameter for all the agents, we use reinforcement learning to set it on an individual basis...
March 2018: IEEE Transactions on Neural Networks and Learning Systems
journal
journal
48247
1
2
Fetch more papers »
Fetching more papers... Fetching...
Read by QxMD. Sign in or create an account to discover new knowledge that matter to you.
Remove bar
Read by QxMD icon Read
×

Search Tips

Use Boolean operators: AND/OR

diabetic AND foot
diabetes OR diabetic

Exclude a word using the 'minus' sign

Virchow -triad

Use Parentheses

water AND (cup OR glass)

Add an asterisk (*) at end of a word to include word stems

Neuro* will search for Neurology, Neuroscientist, Neurological, and so on

Use quotes to search for an exact phrase

"primary prevention of cancer"
(heart or cardiac or cardio*) AND arrest -"American Heart Association"