Efficient Reinforcement Learning using Gaussian Processes

This book examines Gaussian processes in both model-based reinforcement learning (RL) and inference in nonlinear dynamic systems.First, we introduce PILCO, a fully Bayesian approach for efficient RL in continuous-valued state and action spaces when no expert knowledge is available. PILCO takes model...

Full description

Saved in:
Bibliographic Details
Superior document:Karlsruhe Series on Intelligent Sensor-Actuator-Systems / Karlsruher Institut für Technologie, Intelligent Sensor-Actuator-Systems Laboratory
:
Year of Publication:2010
Language:English
Series:Karlsruhe Series on Intelligent Sensor-Actuator-Systems / Karlsruher Institut für Technologie, Intelligent Sensor-Actuator-Systems Laboratory
Physical Description:1 electronic resource (IX, 205 p. p.)
Tags: Add Tag
No Tags, Be the first to tag this record!
id 993545542804498
ctrlnum (CKB)4920000000101410
(oapen)https://directory.doabooks.org/handle/20.500.12854/45907
(EXLCZ)994920000000101410
collection bib_alma
record_format marc
spelling Deisenroth, Marc Peter auth
Efficient Reinforcement Learning using Gaussian Processes
KIT Scientific Publishing 2010
1 electronic resource (IX, 205 p. p.)
text txt rdacontent
computer c rdamedia
online resource cr rdacarrier
Karlsruhe Series on Intelligent Sensor-Actuator-Systems / Karlsruher Institut für Technologie, Intelligent Sensor-Actuator-Systems Laboratory
This book examines Gaussian processes in both model-based reinforcement learning (RL) and inference in nonlinear dynamic systems.First, we introduce PILCO, a fully Bayesian approach for efficient RL in continuous-valued state and action spaces when no expert knowledge is available. PILCO takes model uncertainties consistently into account during long-term planning to reduce model bias. Second, we propose principled algorithms for robust filtering and smoothing in GP dynamic systems.
English
autonomous learning
Gaussian processes
control
machine learning
Bayesian inference
3-86644-569-5
language English
format eBook
author Deisenroth, Marc Peter
spellingShingle Deisenroth, Marc Peter
Efficient Reinforcement Learning using Gaussian Processes
Karlsruhe Series on Intelligent Sensor-Actuator-Systems / Karlsruher Institut für Technologie, Intelligent Sensor-Actuator-Systems Laboratory
author_facet Deisenroth, Marc Peter
author_variant m p d mp mpd
author_sort Deisenroth, Marc Peter
title Efficient Reinforcement Learning using Gaussian Processes
title_full Efficient Reinforcement Learning using Gaussian Processes
title_fullStr Efficient Reinforcement Learning using Gaussian Processes
title_full_unstemmed Efficient Reinforcement Learning using Gaussian Processes
title_auth Efficient Reinforcement Learning using Gaussian Processes
title_new Efficient Reinforcement Learning using Gaussian Processes
title_sort efficient reinforcement learning using gaussian processes
series Karlsruhe Series on Intelligent Sensor-Actuator-Systems / Karlsruher Institut für Technologie, Intelligent Sensor-Actuator-Systems Laboratory
series2 Karlsruhe Series on Intelligent Sensor-Actuator-Systems / Karlsruher Institut für Technologie, Intelligent Sensor-Actuator-Systems Laboratory
publisher KIT Scientific Publishing
publishDate 2010
physical 1 electronic resource (IX, 205 p. p.)
isbn 1-000-01979-9
3-86644-569-5
illustrated Not Illustrated
work_keys_str_mv AT deisenrothmarcpeter efficientreinforcementlearningusinggaussianprocesses
status_str n
ids_txt_mv (CKB)4920000000101410
(oapen)https://directory.doabooks.org/handle/20.500.12854/45907
(EXLCZ)994920000000101410
carrierType_str_mv cr
hierarchy_parent_title Karlsruhe Series on Intelligent Sensor-Actuator-Systems / Karlsruher Institut für Technologie, Intelligent Sensor-Actuator-Systems Laboratory
is_hierarchy_title Efficient Reinforcement Learning using Gaussian Processes
container_title Karlsruhe Series on Intelligent Sensor-Actuator-Systems / Karlsruher Institut für Technologie, Intelligent Sensor-Actuator-Systems Laboratory
_version_ 1787548858393296896
fullrecord <?xml version="1.0" encoding="UTF-8"?><collection xmlns="http://www.loc.gov/MARC21/slim"><record><leader>01617nam-a2200337z--4500</leader><controlfield tag="001">993545542804498</controlfield><controlfield tag="005">20231214133457.0</controlfield><controlfield tag="006">m o d </controlfield><controlfield tag="007">cr|mn|---annan</controlfield><controlfield tag="008">202102s2010 xx |||||o ||| 0|eng d</controlfield><datafield tag="020" ind1=" " ind2=" "><subfield code="a">1-000-01979-9</subfield></datafield><datafield tag="035" ind1=" " ind2=" "><subfield code="a">(CKB)4920000000101410</subfield></datafield><datafield tag="035" ind1=" " ind2=" "><subfield code="a">(oapen)https://directory.doabooks.org/handle/20.500.12854/45907</subfield></datafield><datafield tag="035" ind1=" " ind2=" "><subfield code="a">(EXLCZ)994920000000101410</subfield></datafield><datafield tag="041" ind1="0" ind2=" "><subfield code="a">eng</subfield></datafield><datafield tag="100" ind1="1" ind2=" "><subfield code="a">Deisenroth, Marc Peter</subfield><subfield code="4">auth</subfield></datafield><datafield tag="245" ind1="1" ind2="0"><subfield code="a">Efficient Reinforcement Learning using Gaussian Processes</subfield></datafield><datafield tag="260" ind1=" " ind2=" "><subfield code="b">KIT Scientific Publishing</subfield><subfield code="c">2010</subfield></datafield><datafield tag="300" ind1=" " ind2=" "><subfield code="a">1 electronic resource (IX, 205 p. p.)</subfield></datafield><datafield tag="336" ind1=" " ind2=" "><subfield code="a">text</subfield><subfield code="b">txt</subfield><subfield code="2">rdacontent</subfield></datafield><datafield tag="337" ind1=" " ind2=" "><subfield code="a">computer</subfield><subfield code="b">c</subfield><subfield code="2">rdamedia</subfield></datafield><datafield tag="338" ind1=" " ind2=" "><subfield code="a">online resource</subfield><subfield code="b">cr</subfield><subfield code="2">rdacarrier</subfield></datafield><datafield tag="490" ind1="1" ind2=" "><subfield code="a">Karlsruhe Series on Intelligent Sensor-Actuator-Systems / Karlsruher Institut für Technologie, Intelligent Sensor-Actuator-Systems Laboratory</subfield></datafield><datafield tag="520" ind1=" " ind2=" "><subfield code="a">This book examines Gaussian processes in both model-based reinforcement learning (RL) and inference in nonlinear dynamic systems.First, we introduce PILCO, a fully Bayesian approach for efficient RL in continuous-valued state and action spaces when no expert knowledge is available. PILCO takes model uncertainties consistently into account during long-term planning to reduce model bias. Second, we propose principled algorithms for robust filtering and smoothing in GP dynamic systems.</subfield></datafield><datafield tag="546" ind1=" " ind2=" "><subfield code="a">English</subfield></datafield><datafield tag="653" ind1=" " ind2=" "><subfield code="a">autonomous learning</subfield></datafield><datafield tag="653" ind1=" " ind2=" "><subfield code="a">Gaussian processes</subfield></datafield><datafield tag="653" ind1=" " ind2=" "><subfield code="a">control</subfield></datafield><datafield tag="653" ind1=" " ind2=" "><subfield code="a">machine learning</subfield></datafield><datafield tag="653" ind1=" " ind2=" "><subfield code="a">Bayesian inference</subfield></datafield><datafield tag="776" ind1=" " ind2=" "><subfield code="z">3-86644-569-5</subfield></datafield><datafield tag="906" ind1=" " ind2=" "><subfield code="a">BOOK</subfield></datafield><datafield tag="ADM" ind1=" " ind2=" "><subfield code="b">2023-12-15 05:55:13 Europe/Vienna</subfield><subfield code="f">system</subfield><subfield code="c">marc21</subfield><subfield code="a">2019-11-10 04:18:40 Europe/Vienna</subfield><subfield code="g">false</subfield></datafield><datafield tag="AVE" ind1=" " ind2=" "><subfield code="i">DOAB Directory of Open Access Books</subfield><subfield code="P">DOAB Directory of Open Access Books</subfield><subfield code="x">https://eu02.alma.exlibrisgroup.com/view/uresolver/43ACC_OEAW/openurl?u.ignore_date_coverage=true&amp;portfolio_pid=5338015360004498&amp;Force_direct=true</subfield><subfield code="Z">5338015360004498</subfield><subfield code="b">Available</subfield><subfield code="8">5338015360004498</subfield></datafield></record></collection>