9
Research proposal - DG 2018 GIRARD, Alexandre Physics-based learning for smarter motion controllers Abstract- The proposed research aims at making robots more autonomous, agile and versatile, by investigating ways to include learning in the low-level motion controllers so that their “motor skills” can improve over time and adapt to new situations. A reflexes-based parallel control architecture and a physics-based feature layers are going to be investigated for the control of ground vehicles. 1 Motivation In many areas, autonomous robots have the potential for improving task efficiency, safety and releasing humans of mundane repetitive tasks, for instance transportation (self-driving cars), min- ing (autonomous trucks, robotized excavation), agriculture, inspection, surveillance and even home use (autonomous vacuum cleaners and lawn mowers). The more uncertain and unstructured the environment is, higher capabilities are required for the system to be autonomous and not require a constant human supervision and frequent interventions. Achieving autonomy in complex environ- ments is a challenge at many levels: perception, cognition, controls, etc. The proposed research focus on the science of controlling the motion of dynamic systems, i.e. improving the ”motor skills” of robots, for instance developing control theory and algorithms that can make a self-driving car learn how to drive in the snow or a robotized excavator learn how to dig and improve over time. Figure 1: Many applications where autonomous robots are working in complex environments and would benefit from smarter motion controllers that can learn to adapt and improve over time. 2 Objectives Long-term (10-15 Years) The global objective of the proposed research program is to univer- salize the usage of robots through the development of smarter motor-skill algorithms in order to make them agile, smart and versatile. More specifically, this program will look for a way to include learning, planning and feedback control into a unified and cogent approach to motion control, which could lead to a big breakthrough in the fields of robotics and automation [Levine, 2016]. Short-term (5 years) The short term focus will be on three specific but key aspects: 1) Exploring parallel control architectures to include learning in low-level feedback loops. 2) Accelerating the learning and universalizing the knowledge through the use of physics-based features. Lastly, as a case study and experimental validation, 3) using small-scale autonomous vehicles to evaluate the developed control algorithms for managing uncertain and challenging terrain conditions. 3 Technical challenges and literature review 3.1 Overview of relevant fields An artificial intelligence researcher in the 80s wrote ”It is comparatively easy to make computers exhibit adult level performance on intelligence tests or playing checkers, and difficult or impossible to give them the skills of a one-year-old when it comes to perception and mobility.” [Moravec, Page 1

Physics-based learning for smarter motion controllers · parallel control architectures to include learning in low-level feedback loops. 2) Accelerating 2) Accelerating the learning

Embed Size (px)

Citation preview

Page 1: Physics-based learning for smarter motion controllers · parallel control architectures to include learning in low-level feedback loops. 2) Accelerating 2) Accelerating the learning

Research proposal - DG 2018 GIRARD, Alexandre

Physics-based learning for smarter motion controllersAbstract- The proposed research aims at making robots more autonomous, agile and versatile, by

investigating ways to include learning in the low-level motion controllers so that their “motor skills”can improve over time and adapt to new situations. A reflexes-based parallel control architectureand a physics-based feature layers are going to be investigated for the control of ground vehicles.

1 MotivationIn many areas, autonomous robots have the potential for improving task efficiency, safety and

releasing humans of mundane repetitive tasks, for instance transportation (self-driving cars), min-ing (autonomous trucks, robotized excavation), agriculture, inspection, surveillance and even homeuse (autonomous vacuum cleaners and lawn mowers). The more uncertain and unstructured theenvironment is, higher capabilities are required for the system to be autonomous and not require aconstant human supervision and frequent interventions. Achieving autonomy in complex environ-ments is a challenge at many levels: perception, cognition, controls, etc. The proposed research focuson the science of controlling the motion of dynamic systems, i.e. improving the ”motor skills”of robots, for instance developing control theory and algorithms that can make a self-driving carlearn how to drive in the snow or a robotized excavator learn how to dig and improve over time.

Figure 1: Many applications where autonomous robots are working in complex environments andwould benefit from smarter motion controllers that can learn to adapt and improve over time.

2 ObjectivesLong-term (10-15 Years) The global objective of the proposed research program is to univer-salize the usage of robots through the development of smarter motor-skill algorithms in order tomake them agile, smart and versatile. More specifically, this program will look for a way to includelearning, planning and feedback control into a unified and cogent approach to motion control, whichcould lead to a big breakthrough in the fields of robotics and automation [Levine, 2016].

Short-term (5 years) The short term focus will be on three specific but key aspects: 1) Exploringparallel control architectures to include learning in low-level feedback loops. 2) Acceleratingthe learning and universalizing the knowledge through the use of physics-based features. Lastly,as a case study and experimental validation, 3) using small-scale autonomous vehicles to evaluatethe developed control algorithms for managing uncertain and challenging terrain conditions.

3 Technical challenges and literature review3.1 Overview of relevant fields

An artificial intelligence researcher in the 80s wrote ”It is comparatively easy to make computersexhibit adult level performance on intelligence tests or playing checkers, and difficult or impossibleto give them the skills of a one-year-old when it comes to perception and mobility.” [Moravec,

Page 1

Page 2: Physics-based learning for smarter motion controllers · parallel control architectures to include learning in low-level feedback loops. 2) Accelerating 2) Accelerating the learning

Research proposal - DG 2018 GIRARD, Alexandre

1988]. Since then, we had a breakthrough with deep learning, that can solve challenging perceptionproblems such as classifying objects in images [Krizhevsky et al., 2012]. However, applying thistechnique to mobility problems is not as straightforward, mainly because it is a type of problemwhere we cannot have a big database of “questions and answers” to train a neuralnetwork. Also, pure reinforcement learning approaches for controlling the motion of robots areusually intractable without approximations, due to the infinitely large state-space of robots withmany continuous degree of freedom, the so-called curse of dimensionality [Bellman, 1957].

On the other hand, model-based control approaches have been highly successful for control-ling the motion of complex dynamic systems. State-of-the-art robots have feedback laws andtrajectory generation schemes relying heavily on dynamic models. A typical architecture is gen-erating a feasible/optimal trajectory using either nonlinear programming [Betts, 2010] or searchalgorithms [Lavalle, 2006], then stabilizing the trajectory with a feedback law using a linearizedversion of a dynamic model (ex: darpa challenge humanoid robots [Kuindersma et al., 2016] andself-driving cars [Paden et al., 2016]). Using this approach, robotic systems have been able to ac-complish incredible feats. However, with this type of controllers, robots don’t improve their skillsover time: they would not get better even after doing the same task a thousand time. Hence, simplyby comparing to our own brain, it is clear that control algorithms could be much smarter.

Figure 2: The proposed research aims at exploring hybrid controllers that would leverage the best ofboth worlds and bridge the gap between model-based and data-driven approaches to motion control.

3.2 Specific Research Efforts

One field that successfully combined traditional model-based approaches and the concept oflearning is adaptive control [Astrom and Wittenmark, 2013]. With an adaptive controller, thelearning takes place over unknown model parameters. When the learning is done offline, it is usu-ally called identification. An interesting success of the approach was conducted at Stanford withan aerobatic helicopter, where a local model is identified while an expert fly the helicopter andthen a trajectory optimization scheme is used to conduct autonomous maneuvers [Abbeel et al.,2006]. Regarding autonomous vehicle managing complex ground contacts, state of the art con-trollers generally use model predictive control (MPC) where complex tire adherence models can beconsidered [Beal and Gerdes, 2013]. Interesting explored learning-based approaches consist of esti-mating on-line terrain parameters [Iagnemma et al., 2002] and disturbances [Ostafew et al., 2014].While those approaches are very good steps toward integrating learning in controllers, they arehowever limited by the chosen model and controller structure. With adaptive techniques onlysome parameters can evolve and the improvement potential is limited. Alternatively,iterative learning control is an approach where the control signal is learned directly without anyconstraints, but addresses only sequences of identical repeated tasks [Bristow et al., 2006]. The AI

Page 2

Page 3: Physics-based learning for smarter motion controllers · parallel control architectures to include learning in low-level feedback loops. 2) Accelerating 2) Accelerating the learning

Research proposal - DG 2018 GIRARD, Alexandre

community made some recent effort to use deep neural networks to enable reinforcement learningfor high-dimensional problems. For instance, Google DeepMind created algorithms that can learn toplay simple video games [Mnih et al., 2015] and control simple simulated physical systems [Lillicrapet al., 2015] directly from raw pixel data. There has also been some recent success regarding the useof neural networks for manipulation tasks [Levine et al., 2016]. While those works are very promis-ing, pure reinforcement learning is still highly limited due to the very large number ofrequired training episodes, especially when tackling complex high-dimensional systems and/orreal physical systems with uncertainty. With some applications, the neural network can be trainedwith simulations [Tobin et al., 2017], but for many problems the physics is too complex and accuratesimulations are too computationally heavy. Also, the idea of fusioning machine learning and tradi-tional control is generating a lot of interest recently and a new conference has even be created inthe Fall 2017 to bring together researchers working on robotics and machine learning [CoRL, 2017].While there is some very promising works using a wide range of techniques, we are still missing aunified and generalizable approach to motion control including planning, feedback and learning; andthere is still many open questions. Compared to adaptive controllers, the proposed research aim atgiving more freedom to the policy learning for exploring new maneuvers and strate-gies. Compared to deep reinforcement learning the propose approach is aimed at continuouslyimproving controllers that can initially work with no data at all. More specifically, the twokey novel aspects this program aims to explore 1) a parallel architecture where the learning takeplace in the form of situation-specific reflexes, and 2) a physic-based layer approach for fast learningin scenarios with limited data availability.

3.3 My expertise and recent progress

I first explored a learning approach to motion control during my MS studies, while I was workingwith soft robots made of polymer and very hard to model accurately analytically. I developed acontroller that learn experimentally the behavior of the soft robot and then uses this knowledge torobustly control its position. [Girard and Plante, 2013] [Girard and Plante, 2014]. Then during myPhD at MIT, in addition to studying traditional robotics (dynamics, non-linear control design, etc),I did my minor in the machine learning field. I then explored both AI and traditional techniques forcontrolling hybrid dynamic systems. More specifically, I developed controllers for robots equippedwith multiple (discrete) gear-ratio actuators using dynamic programming [Girard and Asada, 2016],a model-based Lyapunov scheme [Girard and Asada, 2017] and compared the two approaches [Gi-rard, 2017]. I also worked on identifying simple dynamic models (i.e. offline learning) based onexperiments with novel data-driven techniques [Asada et al., 2016]. I thus have developed a uniqueexpertise needed to conduct the proposed research at the frontier of these fields.

4 MethodologyThe proposed program is divided into three projects. The first two are oriented toward fundamen-

tal contributions to control engineering, while the last aims at validating the developed approacheswith a practical case study. The program will train two doctoral students (PhD-1 and PhD-2), twomaster students (MS-1 and MS-2) and four undergraduate interns.

4.1 Parallel control architectures development (Years 1-4, MS-1 and PhD-1)

The first project will explore using reflexes-based control architectures to integrate machinelearning, planning algorithms and feedback laws. The base concept, illustrated at Fig. 3, consist ina baseline model-based controller, and a learning controller overriding the torque command or the

Page 3

Page 4: Physics-based learning for smarter motion controllers · parallel control architectures to include learning in low-level feedback loops. 2) Accelerating 2) Accelerating the learning

Research proposal - DG 2018 GIRARD, Alexandre

reference trajectories in specific situations (i.e. reflexes). MS-1 will 1) review existing architectures,2) explore novel reflexes-based designs and 3) evaluate them using simple simulated robots. Systemstypically used for benchmarking in the literature (ex.: inverted pendulum) will be used. MS-1 goal isto provide quantitative performance results (learning rate, ability to find non-trivial solutions, etc)for comparing multiple architectures. PhD-1 will pursue the theoretical development of the mostpromising reflexes-based architecture, and be tasked with developing a methodology to provide 1)stability guarantees in the context of event-based reflexes and 2) learning convergence conditions.Both MS-1 and PhD-1 will also participate in the experimental validation of the developed controllerson small scale autonomous vehicle prototypes, see section 4.3.

Figure 3: A parallel control architecture for leveraging the advantage of traditional model-basedapproaches and the power of learning schemes. Here reflexes would override the planner or thefeedback controller in specific situations based on past experiences.

4.2 Physics-based features for learning control laws (Years 3-5, PhD-2)

While deep neural networks can learn good features by themselves, they require huge amountsof data to do so and the knowledge is very specific and hard to transfer. The second projectaims at investigating if a physics-based features layer (see Fig. 4) can lead to faster learning andalso to knowledge representations that are more universal, i.e. independent of the sensor typeand transposable between similar situations and similar systems. Instead of using general purposefeature approximations for parameterizing the state-space and control-inputs spaces, PhD-2 willwork on designing features based on dimensionless physically-relevant quantities. For instance,simple intuitive dimensionless number that are relevant for maneuver selection are the frictioncoefficient for a vehicle, the Froude number for a walking robot or the Mach/Reynolds numberfor an airplane. Higher-order relevant quantities are not trivial however. PhD-2 will 1) proposefeature designs for ground vehicles and then 2) use small-scale autonomous car (see section 4.3)to compare experimentally the physics-based filter to alternative approaches, like pure end-to-endneural-networks and general purpose features. PhD-2 objective is to demonstrate the effectivenessof physics-based filters to learn with limited data, by showing experimentally faster learningrates, and make the learned policy independent of the sensing modality, by making cars withdifferent sensors and sizes share policy improvements.

4.3 Learning to maneuver on uncertain grounds with small scale-autonomous vehicleprototypes (Years 2-5, Interns and MS-2)

The third project is aimed at both be an experimental proof-of-concept of the concepts developedby the first two projects and addressing a relevant need (especially in Canada) of the fast growingautonomous vehicle industry. The goal is to make autonomous cars learn to adapt their driving

Page 4

Page 5: Physics-based learning for smarter motion controllers · parallel control architectures to include learning in low-level feedback loops. 2) Accelerating 2) Accelerating the learning

Research proposal - DG 2018 GIRARD, Alexandre

Figure 4: Physics-based features for accelerating the learning and universalizing the knowledge

depending on the ground conditions (gravel, ice, snow, etc). Interns, working in collaboration withthe graduate students, will develop a small-scale autonomous car platform (about 1/10th scalesimilarly to MIT’s racecar platform [MIT, 2016]). The car will be based on a off-the-shelf remotecontrolled electric car, and retrofitted with an embedded computer and sensors such as inertialmeasurement units (IMUs), scanning laser range-finder and cameras. Interns will be tasked withmechanical design, electrical design and development of a modular software architecture (usingROS [Quigley et al., 2009]). MS-2 will study specifically the task of driving on uncertain grounds.He will 1) integrate and adapt the developed control schemes (in collaboration with by PhD-1 andPhD-2) to the driving context, 2) lead the field tests with the platform and 3) aims at demonstratingan autonomous car learning situation-specific maneuvers using both the reflexes-based architectureand the physics-based feature layer.

Risk mitigation strategies: The program is designed such that each sub-project is not dependenton the success of another. The reflex-based architecture and the physics-based features could beinvestigated independently, and could focus on using simulations for evaluation if necessary. Also,the experimental work of MS-2 with the small-scale autonomous car could focus on demonstratingonly the single most successful technique developed in the program.

5 ImpactsAdvancement of knowledge: The proposed research aim at solving fundamental problems

in robotics and automation by investigating new ways to bridge the fields of machine learning andcontrol theory, and could thus lead to major advancements in those two applied sciences. Societal:The society could greatly benefit from robotic systems playing a more direct role in our life: self-driving car making our roads safer, robots releasing humans of many dangerous and repetitivetasks, etc. The proposed program will develop enabling technologies in terms of robotics mobility.Industrial: For the industry, especially in Canada, having autonomous vehicles that can manage allkinds of terrain and weather would be a great advantage for industries ranging from mining, farming,surveillance and transportation. This research program aims at be the catalyst for future appliedresearch project in partnership with those industries. Student Training: Because of the relevanceof the applications for the industry and the growing academic interest, I expect graduating studentsof this research program to have many interesting opportunities in both industrial and academicsetting. Students with advanced degrees on the topics of robotics, machine learningand/or autonomous cars are bound for great careers, rarely has the industry been sointerested at recruiting researchers

Page 5

Page 6: Physics-based learning for smarter motion controllers · parallel control architectures to include learning in low-level feedback loops. 2) Accelerating 2) Accelerating the learning

Research proposal - DG 2018 GIRARD, Alexandre

6 Budget6.1 Salaries and benefits ($244k - 5 Years)

While everyone is expected to work as a team toward an experimental demonstration of physics-based smart motion controllers, each students will be responsible with its own sub-project:

PhD-1: Development of the reflex-based control architecture (Years 2-4, $21k/year)- Review of the adaptive control and reinforcement learning fields- Selection (with MS-1) and optimization of the most promising control architecture- Development of the mathematical framework to analyze learning rates and stability- Experimental validation of the proposed approach with the autonomous car platform

PhD-2: Design of physics-based features to accelerate learning (Years 3-5, $21k/year)- Review of feature design (machine learning field) and dimensional analysis- Design physics-based features for ground vehicles- Experimental validation of faster learning rates with the autonomous platform- Platform independence experimental tests with car using different sensors

MS-1: Comparison of parallel architectures for including learning (Years 1-2, $17.5k/year)- Review of learning schemes and control architectures for motion control- Propose novel parallel control architectures for including learning loops- Systematic evaluation of all approaches using simulation of simple physical systems

MS-2: Experimental field tests (Years 4-5, $17.5k/year)- Review of control scheme to manage uncertain ground conditions- Adaptation of PhD-1 and PhD-2 algorithms to the context of driving- Field tests with the autonomous car platform (with PhD-1 and PhD-2)

Interns (four): Autonomous platform development (Years 1-3, 4 x $12k)- Hands-on experimental work: design, fabrication, software development and field tests- Internships will be used to recruit top undergraduates student for pursuing graduate studies- Internships will be 4-months long to fit in Sherbrooke cooperative program

Figure 5: Time-line of the projects and students

Page 6

Page 7: Physics-based learning for smarter motion controllers · parallel control architectures to include learning in low-level feedback loops. 2) Accelerating 2) Accelerating the learning

Research proposal - DG 2018 GIRARD, Alexandre

6.2 Equipment and facility ($18.5k - 5 Years)

The research will be conducted at the Interdisciplinary Institute for Technological Innovation(3IT) which hosts a machine shop, prototyping machines and test equipments. An amount of1.5K/year is budgeted to pay for the user fees of these resources. Also, an amount of 6k is budgeted,for years 1 to 3 of the program, for supporting equipments for the autonomous cars (three laptopcomputers, network equipments, battery chargers, etc.) Furthermore, an amount of 5k, for years4 to 5 of the program, is budgeted for miscellaneous fees regarding the preparation of a suitableoutdoor terrain for field tests.

6.3 Materials and supplies ($30k - 5 Years)

This proposal requests the budget for building a total of five autonomous car platforms. The costof materials for state-of-the-art 1/10th autonomous platforms is estimated to be about $6k per car,using the MIT racecar platform [MIT, 2016] as a reference. Multiple cars are budgeted since overthe course of this 5 year program, the hardware will be evolving to accommodate new technologyand experimental needs. Also, multiple cars will accommodate students working simultaneouslyon the programming and/or doing experimental work. Furthermore, multiple vehicles will enableaccelerated large scale data collection.

Figure 6: MIT’s Racecar Autonomous Plat-form (based on a 1/10th scale Traxxas remotecontrolled electric car chassis)

Estimated car price breakdown:- Traxxas car chassis : $550- Nvidia Jetson computer : $600- Hokuyo Lidar : $2500- ZED stereo camera : $600- Stucture camera : $500- Sparkfun 9-axis IMU : $100- Misc. electronics : $300- Custom parts : $850

6.4 Travel ($12k - 5 Years)

An amount of $3k/year, for years 2 to 5 of theprogram, is budgeted for traveling to internationalconferences. I aim for each of my HQPs to travelat least once to a leading international conferenceon robotics such as ICRA (International conferenceon robotics and automation) or IROS ( InternationalConference on Intelligent Robots and Systems).

6.5 Dissemination costs ($6k - 5 Years)

An amount of $1k/year, for years 2 to 5 of the program, is budgeted for professional proofreadingand editing services, to help producing high quality papers. An amount of $500/year, also for years 2to 5, is budgeted to cover miscellaneous costs related to publishing such as additional pages charges,memberships to IEEE for students, etc.

Page 7

Page 8: Physics-based learning for smarter motion controllers · parallel control architectures to include learning in low-level feedback loops. 2) Accelerating 2) Accelerating the learning

Research proposal - DG 2018 GIRARD, Alexandre

References[Abbeel et al., 2006] Abbeel, P., Coates, A., Quigley, M., and Ng, A. Y. (2006). An Application of

Reinforcement Learning to Aerobatic Helicopter Flight. In Proceedings of the 19th InternationalConference on Neural Information Processing Systems, NIPS’06, pages 1–8, Cambridge, MA,USA. MIT Press.

[Asada et al., 2016] Asada, H. H., Wu, F., Girard, A., and Mayalu, M. (2016). A data-drivenapproach to precise linearization of nonlinear dynamical systems in augmented latent space. In2016 American Control Conference (ACC), pages 1838–1844, Boston.

[Beal and Gerdes, 2013] Beal, C. E. and Gerdes, J. C. (2013). Model Predictive Control for VehicleStabilization at the Limits of Handling. IEEE Transactions on Control Systems Technology,21(4):1258–1269.

[Bellman, 1957] Bellman, R. (1957). Dynamic Programming. Princeton University Press. Google-Books-ID: wdtoPwAACAAJ.

[Betts, 2010] Betts, J. (2010). Practical Methods for Optimal Control and Estimation Using Non-linear Programming. Advances in Design and Control. Society for Industrial and Applied Math-ematics.

[Bristow et al., 2006] Bristow, D. A., Tharayil, M., and Alleyne, A. G. (2006). A survey of iterativelearning control. IEEE Control Systems, 26(3):96–114.

[CoRL, 2017] CoRL (2017). 1st Annual Conference on Robot Learning : http://www.robot-learning.org/home.

[Girard, 2017] Girard, A. (2017). Fast and Strong Lightweight Robots based on Variable Gear RatioActuators and Control Algorithms Leveraging the Natural Dynamics. PhD thesis, MassachusettsInstitute of Technology.

[Girard and Asada, 2016] Girard, A. and Asada, H. H. (2016). A practical optimal control approachfor two-speed actuators. In IEEE International Conference on Robotics and Automation (ICRA),pages 4572–4577, Stockholm.

[Girard and Asada, 2017] Girard, A. and Asada, H. H. (2017). Leveraging Natural Load Dynamicswith Variable Gear-ratio Actuators. IEEE Robotics and Automation Letters, 2(2):741–748.

[Girard and Plante, 2013] Girard, A. and Plante, J. S. (2013). Applied control of binary robotsbased on influence vectors. In 2013 IEEE International Conference on Robotics and Automation(ICRA), pages 523–529, Karlsruhe.

[Girard and Plante, 2014] Girard, A. and Plante, J. S. (2014). Influence Vectors Control for RobotsUsing Cellular-Like Binary Actuators. IEEE Transactions on Robotics, 30(3):642–651.

[Iagnemma et al., 2002] Iagnemma, K., Shibly, H., and Dubowsky, S. (2002). On-line terrain pa-rameter estimation for planetary rovers. In Proceedings 2002 IEEE International Conference onRobotics and Automation (Cat. No.02CH37292), volume 3, pages 3142–3147.

Page 8

Page 9: Physics-based learning for smarter motion controllers · parallel control architectures to include learning in low-level feedback loops. 2) Accelerating 2) Accelerating the learning

Research proposal - DG 2018 GIRARD, Alexandre

[Krizhevsky et al., 2012] Krizhevsky, A., Sutskever, I., and Hinton, G. E. (2012). ImageNet Classi-fication with Deep Convolutional Neural Networks. pages 1097–1105. Curran Associates, Inc.

[Kuindersma et al., 2016] Kuindersma, S., Deits, R., Fallon, M., Valenzuela, A., Dai, H., Permenter,F., Koolen, T., Marion, P., and Tedrake, R. (2016). Optimization-based locomotion planning,estimation, and control design for the atlas humanoid robot. Autonomous Robots, 40(3):429–455.

[Lavalle, 2006] Lavalle, S. M. (2006). Planning Algorithms. Cambridge University Press, Cambridge; New York, 1 edition edition.

[Levine, 2016] Levine, S. (2016). Deep Learning for Robots: Learning from Large-Scale Interaction.

[Levine et al., 2016] Levine, S., Pastor, P., Krizhevsky, A., and Quillen, D. (2016). Learning Hand-Eye Coordination for Robotic Grasping with Deep Learning and Large-Scale Data Collection.arXiv:1603.02199 [cs]. arXiv: 1603.02199.

[Lillicrap et al., 2015] Lillicrap, T. P., Hunt, J. J., Pritzel, A., Heess, N., Erez, T., Tassa, Y.,Silver, D., and Wierstra, D. (2015). Continuous control with deep reinforcement learning.arXiv:1509.02971 [cs, stat]. arXiv: 1509.02971.

[MIT, 2016] MIT (2016). RACECAR – A Powerful Platform for Robotics Research and Teaching: http://fast.scripts.mit.edu/racecar/.

[Mnih et al., 2015] Mnih, V., Kavukcuoglu, K., Silver, D., Rusu, A. A., Veness, J., Bellemare, M. G.,Graves, A., Riedmiller, M., Fidjeland, A. K., Ostrovski, G., Petersen, S., Beattie, C., Sadik, A.,Antonoglou, I., King, H., Kumaran, D., Wierstra, D., Legg, S., and Hassabis, D. (2015). Human-level control through deep reinforcement learning. Nature, 518(7540):529–533.

[Moravec, 1988] Moravec, H. (1988). Mind Children: The Future of Robot and Human Intelligence.Harvard University Press. Google-Books-ID: 56mb7XuSx3QC.

[Ostafew et al., 2014] Ostafew, C. J., Schoellig, A. P., and Barfoot, T. D. (2014). Learning-basednonlinear model predictive control to improve vision-based mobile robot path-tracking in challeng-ing outdoor environments. In 2014 IEEE International Conference on Robotics and Automation(ICRA), pages 4029–4036.

[Paden et al., 2016] Paden, B., Cap, M., Yong, S. Z., Yershov, D., and Frazzoli, E. (2016). A Surveyof Motion Planning and Control Techniques for Self-Driving Urban Vehicles. IEEE Transactionson Intelligent Vehicles, 1(1):33–55.

[Quigley et al., 2009] Quigley, M., Conley, K., Gerkey, B., Faust, J., Foote, T., Leibs, J., Wheeler,R., and Ng, A. (2009). ROS: an open-source Robot Operating System.

[Tobin et al., 2017] Tobin, J., Fong, R., Ray, A., Schneider, J., Zaremba, W., and Abbeel, P. (2017).Domain Randomization for Transferring Deep Neural Networks from Simulation to the RealWorld. arXiv:1703.06907 [cs]. arXiv: 1703.06907.

[Astrom and Wittenmark, 2013] Astrom, K. J. and Wittenmark, B. (2013). Adaptive Control: Sec-ond Edition. Courier Corporation. Google-Books-ID: 4CLCAgAAQBAJ.

Page 9