Skip to main content
guest
Join
|
Help
|
Sign In
chessprogramming
Home
guest
|
Join
|
Help
|
Sign In
Wiki Home
Recent Changes
Pages and Files
Members
Home
Basics
Getting Started
Board Representation
Search
Evaluation
Principle Topics
Chess
Programming
Artificial Intelligence
Knowledge
Learning
Testing
Tuning
User Interface
Protocols
Dictionary
Lists
Arts
Cartoons
CC Forums
Conferences
Dedicated CC
Engines
Games
Hardware
History
Organizations
Papers
People
Periodical
Samples
Software
Timeline
Tournaments and Matches
Videos
Misc
Acknowledgments
On New Pages
Recommended Reading
Wikispaces Help
Timothy Lillicrap
Edit
0
2
…
2
Tags
people
researcher
Notify
RSS
Backlinks
Source
Print
Export (PDF)
Home
*
People
* Timothy Lillicrap
Timothy P. (Tim) Lillicrap
,
a Canadian
neuroscientist
an
AI
researcher, adjunct professor at
University College London
, and staff research scientist at
Google
,
DeepMind
, where he is involved in the
AlphaGo
and
AlphaZero
projects mastering the games of
Go
,
chess
and
Shogi
. He holds a B.Sc. in
cognitive science
and
artificial intelligence
from
University of Toronto
in 2005, and a Ph.D. in
systems neuroscience
from
Queen's University
in 2014 under
Stephen H. Scott
[1]
[2]
. His research focuses on
machine learning
and
statistics
for
optimal control
and
decision making
, as well as using these mathematical frameworks to understand how the
brain
learns. He has developed algorithms and approaches for exploiting
deep neural networks
in the context of
reinforcement learning
, and new
recurrent memory architectures
for
one-shot learning
[3]
.
Timothy Lillicrap
[4]
Table of Contents
Selected Publications
2014
2015 ...
External Links
References
What links here?
Selected Publications
[5]
2014
Timothy Lillicrap
(
2014
).
Modelling Motor Cortex using Neural Network Controls Laws
. Ph.D. Systems Neuroscience Thesis, Centre for Neuroscience Studies,
Queen's University
, advisor:
Stephen H. Scott
2015 ...
Timothy Lillicrap
,
Jonathan J. Hunt
,
Alexander Pritzel
,
Nicolas Heess
,
Tom Erez
,
Yuval Tassa
,
David Silver
,
Daan Wierstra
(
2015
).
Continuous Control with Deep Reinforcement Learning
.
arXiv:1509.02971
Nicolas Heess
,
Jonathan J. Hunt
,
Timothy Lillicrap
,
David Silver
(
2015
).
Memory-based control with recurrent neural networks
.
arXiv:1512.04455
2016
David Silver
,
Aja Huang
,
Chris J. Maddison
,
Arthur Guez
,
Laurent Sifre
,
George van den Driessche
,
Julian Schrittwieser
,
Ioannis Antonoglou
,
Veda Panneershelvam
,
Marc Lanctot
,
Sander Dieleman
,
Dominik Grewe
,
John Nham
,
Nal Kalchbrenner
,
Ilya Sutskever
,
Timothy Lillicrap
,
Madeleine Leach
,
Koray Kavukcuoglu
,
Thore Graepel
,
Demis Hassabis
(
2016
).
Mastering the game of Go with deep neural networks and tree search
.
Nature
, Vol. 529 »
AlphaGo
Volodymyr Mnih
,
Adrià Puigdomènech Badia
,
Mehdi Mirza
,
Alex Graves
,
Timothy Lillicrap
,
Tim Harley
,
David Silver
,
Koray Kavukcuoglu
(
2016
).
Asynchronous Methods for Deep Reinforcement Learning
.
arXiv:1602.01783v2
Shixiang Gu
,
Timothy Lillicrap
,
Ilya Sutskever
,
Sergey Levine
(
2016
).
Continuous Deep Q-Learning with Model-based Acceleration
.
arXiv:1603.00748
[6]
Shixiang Gu
,
Ethan Holly
,
Timothy Lillicrap
,
Sergey Levine
(
2016
).
Deep Reinforcement Learning for Robotic Manipulation with Asynchronous Off-Policy Updates
.
arXiv:1610.00633
Shixiang Gu
,
Timothy Lillicrap
,
Zoubin Ghahramani
,
Richard E. Turner
,
Sergey Levine
(
2016
).
Q-Prop: Sample-Efficient Policy Gradient with An Off-Policy Critic
.
arXiv:1611.02247
2017
Yutian Chen
,
Matthew W. Hoffman
,
Sergio Gomez Colmenarejo
,
Misha Denil
,
Timothy Lillicrap
,
Matthew Botvinick
,
Nando de Freitas
(
2017
).
Learning to Learn without Gradient Descent by Gradient Descent
.
arXiv:1611.03824v6
,
ICML 2017
David Silver
,
Julian Schrittwieser
,
Karen Simonyan
,
Ioannis Antonoglou
,
Aja Huang
,
Arthur Guez
,
Thomas Hubert
,
Lucas Baker
,
Matthew Lai
,
Adrian Bolton
,
Yutian Chen
,
Timothy Lillicrap
,
Fan Hui
,
Laurent Sifre
,
George van den Driessche
,
Thore Graepel
,
Demis Hassabis
(
2017
).
Mastering the game of Go without human knowledge
.
Nature
, Vol. 550
[7]
David Silver
,
Thomas Hubert
,
Julian Schrittwieser
,
Ioannis Antonoglou
,
Matthew Lai
,
Arthur Guez
,
Marc Lanctot
,
Laurent Sifre
,
Dharshan Kumaran
,
Thore Graepel
,
Timothy Lillicrap
,
Karen Simonyan
,
Demis Hassabis
(
2017
).
Mastering Chess and Shogi by Self-Play with a General Reinforcement Learning Algorithm
.
arXiv:1712.01815
»
AlphaZero
External Links
homepage of timothy lillicrap
Timothy P. Lillicrap - Google Scholar Citations
Tim Lillicrap - Data efficient Deep Reinforcement Learning for Continuous Control
,
YouTube
Video
References
^
Timothy Lillicrap
(
2014
).
Modelling Motor Cortex using Neural Network Controls Laws
. Ph.D. Systems Neuroscience Thesis, Centre for Neuroscience Studies,
Queen's University
, advisor:
Stephen H. Scott
^
Curriculum Vitae - Timothy P. Lillicrap
(pdf)
^
timothy lillicrap - research
^
Image captured from the
Data efficient Deep Reinforcement Learning for Continuous Control - Video
at 20:21
^
dblp: Timothy P. Lillicrap
^
Q-learning from Wikipedia
^
AlphaGo Zero: Learning from scratch
by
Demis Hassabis
and
David Silver
,
DeepMind
, October 18, 2017
What links here?
Page
Date Edited
AlphaZero
Feb 10, 2018
Arthur Guez
Dec 6, 2017
Chess
Jan 21, 2018
Chris J. Maddison
Dec 8, 2017
David Silver
Feb 11, 2018
Deep Learning
Feb 12, 2018
DeepMind
Dec 9, 2017
Demis Hassabis
Dec 8, 2017
Dharshan Kumaran
Dec 9, 2017
Gian-Carlo Pascutto
Jan 16, 2018
Go
Jan 24, 2018
Ilya Sutskever
Jan 28, 2017
Ioannis Antonoglou
Dec 6, 2017
Julian Schrittwieser
Dec 7, 2017
Karen Simonyan
Dec 10, 2017
Koray Kavukcuoglu
Dec 10, 2017
Laurent Sifre
Dec 7, 2017
LCZero
Apr 18, 2018
Learning
Feb 20, 2018
Marc Lanctot
Jan 10, 2018
Matthew Lai
Dec 6, 2017
Monte-Carlo Tree Search
Apr 26, 2018
Neural Networks
Mar 12, 2018
People
Feb 28, 2018
Reinforcement Learning
Feb 12, 2018
Robots
Feb 20, 2018
Shih-Chieh Huang
Oct 18, 2017
Shogi
Feb 19, 2018
Thomas Hubert
Dec 7, 2017
Thore Graepel
Jan 10, 2018
Timothy Lillicrap
Dec 9, 2017
University of Toronto
Dec 8, 2017
Volodymyr Mnih
Dec 8, 2017
Up one level
Javascript Required
You need to enable Javascript in your browser to edit pages.
help on how to format text
Turn off "Getting Started"
Home
...
Loading...
a Canadian neuroscientist an AI researcher, adjunct professor at University College London, and staff research scientist at Google, DeepMind, where he is involved in the AlphaGo and AlphaZero projects mastering the games of Go, chess and Shogi. He holds a B.Sc. in cognitive science and artificial intelligence from University of Toronto in 2005, and a Ph.D. in systems neuroscience from Queen's University in 2014 under Stephen H. Scott [1] [2]. His research focuses on machine learning and statistics for optimal control and decision making, as well as using these mathematical frameworks to understand how the brain learns. He has developed algorithms and approaches for exploiting deep neural networks in the context of reinforcement learning, and new recurrent memory architectures for one-shot learning [3].
Table of Contents
Selected Publications
[5]2014
2015 ...
- Timothy Lillicrap, Jonathan J. Hunt, Alexander Pritzel, Nicolas Heess, Tom Erez, Yuval Tassa, David Silver, Daan Wierstra (2015). Continuous Control with Deep Reinforcement Learning. arXiv:1509.02971
- Nicolas Heess, Jonathan J. Hunt, Timothy Lillicrap, David Silver (2015). Memory-based control with recurrent neural networks. arXiv:1512.04455
2016- David Silver, Aja Huang, Chris J. Maddison, Arthur Guez, Laurent Sifre, George van den Driessche, Julian Schrittwieser, Ioannis Antonoglou, Veda Panneershelvam, Marc Lanctot, Sander Dieleman, Dominik Grewe, John Nham, Nal Kalchbrenner, Ilya Sutskever, Timothy Lillicrap, Madeleine Leach, Koray Kavukcuoglu, Thore Graepel, Demis Hassabis (2016). Mastering the game of Go with deep neural networks and tree search. Nature, Vol. 529 » AlphaGo
- Volodymyr Mnih, Adrià Puigdomènech Badia, Mehdi Mirza, Alex Graves, Timothy Lillicrap, Tim Harley, David Silver, Koray Kavukcuoglu (2016). Asynchronous Methods for Deep Reinforcement Learning. arXiv:1602.01783v2
- Shixiang Gu, Timothy Lillicrap, Ilya Sutskever, Sergey Levine (2016). Continuous Deep Q-Learning with Model-based Acceleration. arXiv:1603.00748 [6]
- Shixiang Gu, Ethan Holly, Timothy Lillicrap, Sergey Levine (2016). Deep Reinforcement Learning for Robotic Manipulation with Asynchronous Off-Policy Updates. arXiv:1610.00633
- Shixiang Gu, Timothy Lillicrap, Zoubin Ghahramani, Richard E. Turner, Sergey Levine (2016). Q-Prop: Sample-Efficient Policy Gradient with An Off-Policy Critic. arXiv:1611.02247
2017External Links
References
What links here?
Up one level