Ask a Librarian

Threre are lots of ways to contact a librarian. Choose what works best for you.

HOURS TODAY

Closed

Reference Desk

CONTACT US BY PHONE

(802) 656-2022

Voice

(802) 503-1703

Text

MAKE AN APPOINTMENT OR EMAIL A QUESTION

Schedule an Appointment

Meet with a librarian or subject specialist for in-depth help.

Email a Librarian

Submit a question for reply by e-mail.

WANT TO TALK TO SOMEONE RIGHT AWAY?

Library Hours for Saturday, November 23rd

All of the hours for today can be found below. We look forward to seeing you in the library.
HOURS TODAY
Closed
MAIN LIBRARY

SEE ALL LIBRARY HOURS
WITHIN HOWE LIBRARY

MapsM-Th by appointment, email govdocs@uvm.edu

Media ServicesClosed

Reference DeskClosed

OTHER DEPARTMENTS

Special CollectionsClosed

Dana Health Sciences Library10:00 am - 6:00 pm

 

CATQuest

Search the UVM Libraries' collections

UVM Theses and Dissertations

Browse by Department
Format:
Online
Author:
Felag, Jack Michael
Dept./Program:
Complex Systems and Data Science
Year:
2020
Degree:
M.S.
Abstract:
Agents are often trained to perform a task via optimization algorithms. One class of algorithms used is evolution, which is "survival of the fitness" used to pick the best agents for the objective, and slowly changing the best over time to find a good solution. Evolution, or evolutionary algorithms, have been commonly used to automatically select for a better body of the agent, which can outperform hand-designed models. Another class of algorithms used is reinforcement learning. Through this strategy, agents learn from prior experiences in order to maximize some reward. Generally, this reward is how close the objective is to being complete, or otherwise some stepping stone towards its completion. Evolution and reinforcement learning can both train agents to the point where a task can be successfully completed, or an objective met. In this thesis, we outline a framework for combining evolution and reinforcement learning, and outline experimental designs to test this method against traditional reinforcement learning. Finally, we show preliminary results as a proof of concept for the methods described.