Date of Degree

2-2018

Document Type

Dissertation

Degree Name

Ph.D.

Program

Computer Science

Advisor

Felisa Vázquez-Abad

Committee Members

Robert Haralick

Chao Chen

Vikram Krishnamurthy

Subject Categories

Artificial Intelligence and Robotics | Computer Sciences | Dynamical Systems | Theory and Algorithms

Keywords

optimization, stochastic processes, machine learning

Abstract

It has been hypothesized that neural network models with cyclic connectivity may be more powerful than their feed-forward counterparts. This thesis investigates this hypothesis in several ways. We study the gradient estimation and optimization procedures for several variants of these networks. We show how the convergence of the gradient estimation procedures are related to the properties of the networks. Then we consider how to tune the relative rates of gradient estimation and parameter adaptation to ensure successful optimization in these models. We also derive new gradient estimators for stochastic models. First, we port the forward sensitivity analysis method to the stochastic setting. Secondly, we show how to apply measure valued differentiation in order to calculate derivatives of long-term costs in general models on a discrete state space. Throughout, we emphasize how the proper geometric framework can simplify and generalize the analysis of these problems.

This work is embargoed and will be available for download on Friday, February 01, 2019

Graduate Center users:
To read this work, log in to your GC ILL account and place a thesis request.

Non-GC Users:
See the GC’s lending policies to learn more.

Share

COinS