Skip to main content

Graduate Program

NSF Research Traineeship (NRT)

Practicum Projects

A replication of AlphaGo

In this course, Practicum In Data-Enabled Research Into Human Behavior And Its Cognitive & Neural Mechanisms, we started off with a discussion of a variety of projects/artifacts the students are interested in. One interesting publication [1] caught our eyes. Many had thought that computer master of the game of go was decades away, so the sudden success of Google DeepMind's AlphaGo system came as a shock. The paper used deep reinforcement learning, which has influences from the area of Brain and Cognitive Science. 

[1] Silver D, Huang A, Maddison C J, et al. Mastering the game of Go with deep neural networks and tree search[J]. Nature, 2016, 529(7587): 484-489.

The project can be found here: https://github.com/Rochester-NRT/AlphaGo/wiki

Bios of the class members

Chris Bates

Chris started his PhD work at University of Rochester's Brain and Cognitive Sciences department in 2015, under Professor Robert Jacobs. He is broadly interested in computational modeling of cognition and perception, including physical reasoning and multisensory perception. He received his B.S. in mechanical engineering from Purdue University.

Project contributions. His main contributions in the AlphaGo replication project pertain to the reinforcement learning policy, value network, and programming the game-play logic. 

What did I learn from the course? "This course project has exposed me to coding best-practices and use of Github as a collaborative coding tool. It has also exposed me to programming with neural networks on GPUs and use of computing clusters. I have been surprised at how doable the project is and the eagerness of community collaborators on Github."

Iris Yuping Ren

Iris Ren is a PhD student in the Audio Information Retrival lab, Electrical and Computer Engineering department. She received her B.S. in Statistics, B.M.S in Culture Industry Management from Shandong University in 2013, double Masters in Complex System Science from the University of Warwick and Ecole Polytechnique in 2015. She plays the violin and piano, sings in choirs, and learning ukulele and percussion. Her website can be found here.

Project contributions. She worked on parts of the game logic, benchmarking, hdf5 file concatenation, game utils, etc. 

What did I learn from the course? The framework of the AlphaGo system. Brainstorming and teamwork. Github version control. Techniques in python. 

How might this course affect my career? "I am now more interested in a career relevant to machine learning." 

What was surprising or the greatest challenge? Writing quality code and optimisation. 

Richard Lange

Richard is a second-year PhD student pursuing a dual degree in Computer Science and Brain and Cognitive Science. His interests span neural coding, machine learning, cognitive science, and more.

What did I learn from the course? "In this course, I took on my first significant project leadership role. In addition to getting valuable hands-on experience with modern neural network techniques, I learned how to better explain technical concepts, and how to delegate work effectively."

How might this course affect my caree? "There is a lot that brain scientists can learn from neural networks, and there is no subsitute for hands-on experience. I hope to take what I learned from this project as a springboard into using them in my research."

What was surprising or the greatest challenge? "This course challenged us to break our habits formed working on software alone, taking time to blueprint and write a good open-source project."

Louis Marti

Louis graduated from the University of Maryland, College Park with a double degree in Computer Science and Psychology. He worked as a professional software developer for six years before deciding to pursue a PhD in Cognitive Science. He is now studying meta-cognition and certainty with the goal of learning how humans come to believe they know what they believe they know.

Project contributions

Louis worked on bench-marking and optimization, along with the web server

Tyler Trine

Tyler is an undergraduate data science student. His plans for next year are to join the workforce. Longer term, he is passionate about furthering the field of artificial intelligence and using it to improve people's lives.

Project contributions. He helped interpret and implement parts of AlphaGo's training pipeline, based on DeepMind's originial paper in Nature. 

What did I learn from the course? "I've learned more from this class than any other I've taken in college! In hindsight, we took on a very ambitious project. But throughout the process of learning how AlphaGo works, I became familiar with state-of-the-art deep learning techniques. If we hadn't reached for such a lofty goal, I may not have learned as much."

How might this course affect my career? "Very positively! This class has developed my understanding of machine learning immensely. The techniques we used are general; they have a broad range of applications. Their use is so widespread, in fact, that I feel I would have an edge in both industry and academia."

What was surprising or the greatest challenge? "The greatest challenge was managing complexity. AlphaGo has so many moving parts, each with a distinct purpose. Furthermore, these parts combine to form an idiosyncratic and, at times, unintuitive whole. Thankfully Richard, who studied computer science in his undergraduate, showed us how to build the project scalably from the ground up!"

Yue Wang

University of Rochester, CS MS student (entered 2015).

Email: ywang214@ur.rochester.edu

Project contributions. She replicated AlphaGo's Monte Carlo tree search.

What did I learn from the course? Learned about many things, including: game tree search, convolutional neural networks, reinforcement learning, supervised training, data structures, and algorithms. 

How might this project affect my career? 

"The biggest takeaway of this course is that it inspired me to learn lots of new things, such as reinforcement learning, convnets, ResNet, GoogleNet, Deep Dream, recurrent neural networks, LSTM, SGD, momentum, Adagrad, and so on. I find them very interesting, and enjoy studying them. I will keep learning about CNN even after the course ends. I find it most helpful on a personal interest level."

What was surprising or the greatest challenge?

"Replicating AlphaGo is a great project to work on in terms of combining the knowledge I learned from my machine learning course and other related courses, and having an idea of how they can be used in the real world. It is pretty impressive that AlphaGo brings convnets to go-playing, by training all those layers of this large network with human players' data and self-play data where the network plays against itself. The project is very challenging, but it is worth working on, regardless of whether we are able to finish the entire project within a few months or not."