Learning Control from Raw Position Measurements
Clicks: 13
ID: 283133
2023
Article Quality & Performance Metrics
Overall Quality
Improving Quality
0.0
/100
Combines engagement data with AI-assessed academic quality
Reader Engagement
Emerging Content
0.3
/100
1 views
1 readers
Trending
AI Quality Assessment
Not analyzed
Abstract
We propose a Model-Based Reinforcement Learning (MBRL) algorithm named
VF-MC-PILCO, specifically designed for application to mechanical systems where
velocities cannot be directly measured. This circumstance, if not adequately
considered, can compromise the success of MBRL approaches. To cope with this
problem, we define a velocity-free state formulation which consists of the
collection of past positions and inputs. Then, VF-MC-PILCO uses Gaussian
Process Regression to model the dynamics of the velocity-free state and
optimizes the control policy through a particle-based policy gradient approach.
We compare VF-MC-PILCO with our previous MBRL algorithm, MC-PILCO4PMS, which
handles the lack of direct velocity measurements by modeling the presence of
velocity estimators. Results on both simulated (cart-pole and UR5 robot) and
real mechanical systems (Furuta pendulum and a ball-and-plate rig) show that
the two algorithms achieve similar results. Conveniently, VF-MC-PILCO does not
require the design and implementation of state estimators, which can be a
challenging and time-consuming activity to be performed by an expert user.
Reference Key |
romeres2023learning
Use this key to autocite in the manuscript while using
SciMatic Manuscript Manager or Thesis Manager
|
---|---|
Authors | Fabio Amadio; Alberto Dalla Libera; Daniel Nikovski; Ruggero Carli; Diego Romeres |
Journal | arXiv |
Year | 2023 |
DOI | DOI not found |
URL | |
Keywords |
Citations
No citations found. To add a citation, contact the admin at info@scimatic.org
Comments
No comments yet. Be the first to comment on this article.