Nothing Special   »   [go: up one dir, main page]

De Lellis et al., 2023 - Google Patents

CT-DQN: Control-tutored deep reinforcement learning

De Lellis et al., 2023

View PDF
Document ID
14781752081995401367
Author
De Lellis F
Coraggio M
Russo G
Musolesi M
dI Bernardo M
Publication year
Publication venue
Learning for Dynamics and Control Conference

External Links

Snippet

One of the major challenges in Deep Reinforcement Learning for control is the need for extensive training to learn the policy. Motivated by this, we present the design of the Control- Tutored Deep Q-Networks (CT-DQN) algorithm, a Deep Reinforcement Learning algorithm …
Continue reading at proceedings.mlr.press (PDF) (other versions)

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06NCOMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N99/00Subject matter not provided for in other groups of this subclass
    • G06N99/005Learning machines, i.e. computer in which a programme is changed according to experience gained by the machine itself during a complete run
    • GPHYSICS
    • G05CONTROLLING; REGULATING
    • G05BCONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
    • G05B13/00Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion
    • G05B13/02Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric
    • G05B13/0265Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric the criterion being a learning criterion
    • G05B13/027Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric the criterion being a learning criterion using neural networks only
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06NCOMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computer systems based on biological models
    • G06N3/02Computer systems based on biological models using neural network models
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06NCOMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N5/00Computer systems utilising knowledge based models
    • G06N5/02Knowledge representation
    • G06N5/022Knowledge engineering, knowledge acquisition
    • GPHYSICS
    • G05CONTROLLING; REGULATING
    • G05BCONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
    • G05B13/00Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion
    • G05B13/02Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric
    • G05B13/04Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric involving the use of models or simulators

Similar Documents

Publication Publication Date Title
Yu et al. Mopo: Model-based offline policy optimization
Ni et al. Adaptive learning in tracking control based on the dual critic network design
Cao et al. Matrix measure strategies for stability and synchronization of inertial BAM neural network with time delays
Xu et al. Knowledge transfer in multi-task deep reinforcement learning for continuous control
Pan et al. Model reference composite learning control without persistency of excitation
Joshi et al. Design and flight evaluation of deep model reference adaptive controller
Ure et al. Health aware planning under uncertainty for uav missions with heterogeneous teams
Behzadan et al. The faults in our pi stars: Security issues and open challenges in deep reinforcement learning
De Lellis et al. CT-DQN: Control-tutored deep reinforcement learning
Zhang et al. Model‐Free Attitude Control of Spacecraft Based on PID‐Guide TD3 Algorithm
Valluru et al. Stabilization of nonlinear inverted pendulum system using MOGA and APSO tuned nonlinear PID controller
Krishnakumar et al. Hybrid fuzzy logic flight controller synthesis via pilot modeling
Xian et al. Control of quadrotor robot via optimized nonlinear type-2 fuzzy fractional PID with fractional filter: Theory and experiment
Alsalti et al. Data-driven nonlinear predictive control for feedback linearizable systems
Gu et al. A human-centered safe robot reinforcement learning framework with interactive behaviors
Liu et al. How to guide your learner: Imitation learning with active adaptive expert involvement
Long et al. Online Optimal Control of Robotic Systems with Single Critic NN‐Based Reinforcement Learning
Zhang et al. Research on adaptive non-singular fast terminal sliding mode control based on variable exponential power reaching law in manipulators
Uchibe Cooperative and competitive reinforcement and imitation learning for a mixture of heterogeneous learning modules
Cheng et al. Offline quantum reinforcement learning in a conservative manner
Lin et al. Adaptive fuzzy total sliding-mode control of unknown nonlinear systems
Yang et al. Variational adversarial kernel learned imitation learning
Li et al. Curricular robust reinforcement learning via GAN-based perturbation through continuously scheduled task sequence
Chawla et al. System identification of an inverted pendulum using adaptive neural fuzzy inference system
Deka et al. Arc-actor residual critic for adversarial imitation learning