Intelligent Systems, Technologies and Applications by Unknown
Author:Unknown
Language: eng
Format: epub
ISBN: 9789811539145
Publisher: Springer Singapore
3 Background of Reinforcement Learning
Information extraction in the form of concept extraction, diagnostic inference and named-entity recognition is important for medical research as well as for automatic preliminary medical care. Our objective is to extract concepts from semi-structured data like discharge summaries of the I2B2 dataset. Deep learning in the form of reinforcement learning and Bi-LSTM models is used in the process of achieving this objective.
Active learning involves selecting highly informative samples from an un-annotated dataset to decrease the cost of annotation as well as time. This problem can be formulated as a reinforcement learning problem where the state space and actions are modeled as Markov decision process (MDP). An MDP should satisfy Markov chain property, which states: The future state depends only on the current state and no other. In reinforcement learning (RL), an agent is the decisive factor, which selects the best possible action to go to the next state from the current state by maximizing the reward. The major difference between supervised learning and RL is that in the former, there is a labeled dataset against which the model can be evaluated, whereas in the latter case, the agent learns from its experience. The agent tries every possible path to reach the destination during which it learns the best path from origin to destination. Learning in RL is incorporated using a reward mechanism. Depending on the design of the reward function, the learning happens in the best possible way. There are two types of RL: positive and negative.
Positive RL: Every time the reward is maximized, it helps in gaining strength and the frequency of the behavior.
Negative RL: The attainment of negative condition is avoided in this kind of learning, thus maximizing the behavior.
RL can be applied to any kind of problem where the source and destination are known in advance. The best path taken by the agent can be automated using deep neural networks; hence, deep RL gives an advantage over the traditional approach. Any RL internally works as MDP.
The goal of any MDP is to find the best policy that can take us to the destination state with maximum reward. Markov chain Monte Carlo (MCMC) is an advancement over Markov chain. Monte Carlo approach is to calculate the reward at the end of each episode. This algorithm uses the previous steps to generate the future steps. This strategy considerably speeds up the process.
Download
This site does not store any files on its server. We only index and link to content provided by other sites. Please contact the content providers to delete copyright contents if any and email us, we'll remove relevant links or contents immediately.
Automotive | Engineering |
Transportation |
Whiskies Galore by Ian Buxton(41529)
Introduction to Aircraft Design (Cambridge Aerospace Series) by John P. Fielding(32888)
Small Unmanned Fixed-wing Aircraft Design by Andrew J. Keane Andras Sobester James P. Scanlan & András Sóbester & James P. Scanlan(32573)
Craft Beer for the Homebrewer by Michael Agnew(17933)
Turbulence by E. J. Noyes(7700)
The Complete Stick Figure Physics Tutorials by Allen Sarah(7138)
Kaplan MCAT General Chemistry Review by Kaplan(6595)
The Thirst by Nesbo Jo(6436)
Bad Blood by John Carreyrou(6274)
Modelling of Convective Heat and Mass Transfer in Rotating Flows by Igor V. Shevchuk(6222)
Learning SQL by Alan Beaulieu(6035)
Weapons of Math Destruction by Cathy O'Neil(5829)
Man-made Catastrophes and Risk Information Concealment by Dmitry Chernov & Didier Sornette(5646)
Digital Minimalism by Cal Newport;(5389)
Life 3.0: Being Human in the Age of Artificial Intelligence by Tegmark Max(5184)
iGen by Jean M. Twenge(5161)
Secrets of Antigravity Propulsion: Tesla, UFOs, and Classified Aerospace Technology by Ph.D. Paul A. Laviolette(4990)
Design of Trajectory Optimization Approach for Space Maneuver Vehicle Skip Entry Problems by Runqi Chai & Al Savvaris & Antonios Tsourdos & Senchun Chai(4839)
Electronic Devices & Circuits by Jacob Millman & Christos C. Halkias(4748)
