Search or filter publications

Filter by type:

Filter by publication type

Filter by year:

to

Results

  • Showing results for:
  • Reset all filters

Search results

  • CONFERENCE PAPER
    Joulani P, György A, Szepesvári C, 2017,

    A Modular Analysis of Adaptive (Non-)Convex Optimization: Optimism, Composite Objectives, and Variational Bounds.

    , Publisher: PMLR, Pages: 681-720
  • CONFERENCE PAPER
    Kamthe S, Deisenroth MP, 2017,

    Data-Efficient Reinforcement Learning with Probabilistic Model Predictive Control.

  • CONFERENCE PAPER
    Kanajar P, Caldwell DG, Kormushev P, 2017,

    Climbing over Large Obstacles with a Humanoid Robot via Multi-Contact Motion Planning

  • JOURNAL ARTICLE
    Kupcsik A, Deisenroth MP, Peters J, Loh AP, Vadakkepat P, Neumann Get al., 2017,

    Model-based contextual policy search for data-efficient generalization of robot skills

    , Artificial Intelligence, Vol: 247, Pages: 415-439, ISSN: 0004-3702

    © 2014 Elsevier B.V. In robotics, lower-level controllers are typically used to make the robot solve a specific task in a fixed context. For example, the lower-level controller can encode a hitting movement while the context defines the target coordinates to hit. However, in many learning problems the context may change between task executions. To adapt the policy to a new context, we utilize a hierarchical approach by learning an upper-level policy that generalizes the lower-level controllers to new contexts. A common approach to learn such upper-level policies is to use policy search. However, the majority of current contextual policy search approaches are model-free and require a high number of interactions with the robot and its environment. Model-based approaches are known to significantly reduce the amount of robot experiments, however, current model-based techniques cannot be applied straightforwardly to the problem of learning contextual upper-level policies. They rely on specific parametrizations of the policy and the reward function, which are often unrealistic in the contextual policy search formulation. In this paper, we propose a novel model-based contextual policy search algorithm that is able to generalize lower-level controllers, and is data-efficient. Our approach is based on learned probabilistic forward models and information theoretic policy search. Unlike current algorithms, our method does not require any assumption on the parametrization of the policy or the reward function. We show on complex simulated robotic tasks and in a real robot experiment that the proposed learning framework speeds up the learning process by up to two orders of magnitude in comparison to existing methods, while learning high quality policies.

  • CONFERENCE PAPER
    Olofsson S, Mehrian M, Geris L, Calandra R, Deisenroth MP, Misener Ret al., 2017,

    Bayesian Multi-Objective Optimisation of Neotissue Growth in a Perfusion Bioreactor Set-Up

    , 27th European Symposium on Computer-Aided Process Engineering (ESCAPE), Publisher: ELSEVIER SCIENCE BV, Pages: 2155-2160, ISSN: 1570-7946
  • CONFERENCE PAPER
    Pardo F, Tavakoli A, Levdik V, Kormushev Pet al., 2017,

    Time Limits in Reinforcement Learning

    , Deep Reinforcement Learning Symposium (DRLS), 31st Conference on Neural Information Processing Systems (NIPS 2017)

    In reinforcement learning, it is common to let an agent interact with its environment for a fixed amount of time before resetting the environment and repeating the process in a series of episodes. The task that the agent has to learn can either be to maximize its performance over (i) that fixed period, or (ii) an indefinite period where time limits are only used during training to diversify experience. In this paper, we investigate theoretically how time limits could effectively be handled in each of the two cases. In the first one, we argue that the terminations due to time limits are in fact part of the environment, and propose to include a notion of the remaining time as part of the agent’s input. In the second case, the time limits are not part of the environment and are only used to facilitate learning. We argue that such terminations should not be treated as environmental ones and propose a method, specific to value-based algorithms, that incorporates this insight by continuing to bootstrap at the end of each partial episode. To illustrate the significance of our proposals, we perform several experiments on a range of environments from simple few-state transition graphs to complex control tasks, including novel and standard benchmark domains. Our results show that the proposed methods improve the performance and stability of existing reinforcement learning algorithms.

  • CONFERENCE PAPER
    Rakicevic N, Kormushev P, 2017,

    Efficient Robot Task Learning and Transfer via Informed Search in Movement Parameter Space

  • CONFERENCE PAPER
    Somuyiwa SO, Gyorgy A, Gunduz D, 2017,

    Energy-Efficient Wireless Content Delivery with Proactive Caching

    , 15th International Symposium on Modeling and Optimization in Mobile, Ad Hoc, and Wireless Networks (WiOpt), Publisher: IEEE
  • CONFERENCE PAPER
    Somuyiwa SO, Gyorgy A, Gunduz D, 2017,

    Improved Policy Representation and Policy Search for Proactive Content Caching in Wireless Networks

    , 15th International Symposium on Modeling and Optimization in Mobile, Ad Hoc, and Wireless Networks (WiOpt), Publisher: IEEE
  • CONFERENCE PAPER
    Tavakoli A, Pardo F, Kormushev P, 2017,

    Action Branching Architectures for Deep Reinforcement Learning

This data is extracted from the Web of Science and reproduced under a licence from Thomson Reuters. You may not copy or re-distribute this data in whole or in part without the written consent of the Science business of Thomson Reuters.

Request URL: http://wlsprd.imperial.ac.uk:80/respub/WEB-INF/jsp/search-t4-html.jsp Request URI: /respub/WEB-INF/jsp/search-t4-html.jsp Query String: id=954&limit=10&page=3&respub-action=search.html Current Millis: 1532366836824 Current Time: Mon Jul 23 18:27:16 BST 2018