[iasmath-semru] Theoretical Computer Science/Discrete Math Seminars--Week of December 10, 2018
Kristina Phillips
kphillips at ias.edu
Tue Dec 4 16:04:11 EST 2018
INSTITUTE FOR ADVANCED STUDY
School of Mathematics
Princeton, NJ 08540
Theoretical Computer Science/Discrete Math Seminars
Week of December 10, 2018
--------------
To view mathematics in titles and abstracts, please click on the talk's
link.
--------------
Monday, December 10
Computer Science/Discrete Mathematics Seminar I
Topic: A matrix expander Chernoff bound
Speaker: Ankit Garg, Microsoft Research
Time/Room: 11:15am - 12:15pm/Simonyi Hall 101
Abstract Link: <http://www.math.ias.edu/seminars/abstract?event=128879>
http://www.math.ias.edu/seminars/abstract?event=128879
Monday, December 10
Seminar on Theoretical Machine Learning
Topic: On Expressiveness and Optimization in Deep
Learning
Speaker: Nadav Cohen, Member, School of Mathematics
Time/Room: 12:15pm - 1:45pm/White Levy Room
Abstract Link: <http://www.math.ias.edu/seminars/abstract?event=139466>
http://www.math.ias.edu/seminars/abstract?event=139466
Tuesday, December 11
Computer Science/Discrete Mathematics Seminar II
Topic: An invitation to tensor networks
Speaker: Michael Walter, University of Amsterdam
Time/Room: 10:30am - 12:30pm/Simonyi Hall 101
Abstract Link: <http://www.math.ias.edu/seminars/abstract?event=129115>
http://www.math.ias.edu/seminars/abstract?event=129115
1 A matrix expander Chernoff bound
Ankit Garg
Chernoff-type bounds study concentration of sums of independent random
variables and are extremely useful in various settings. In many settings,
the random variables may not be completely independent but only have limited
independence. One such setting, which turns out to be useful in
derandomization and theoretical computer science, in general, involves
random walks on expanders. I will talk about a Chernoff-type bound for sums
of matrix-valued random variables sampled via a random walk on an expander.
Our proof is based on a new multi-matrix extension of the Golden-Thompson
inequality which improves in some ways a recent inequality due to Sutter et
al. and may be of independent interest, as well as an adaptation of an
argument for the scalar case due to Healy. Based on joint work with Yin Tat
Lee, Zhao Song and Nikhil Srivastava.
http://www.math.ias.edu/seminars/abstract?event=128879
2 On Expressiveness and Optimization in Deep Learning
Nadav Cohen
Understanding deep learning calls for addressing three fundamental
questions: expressiveness, optimization and generalization. Expressiveness
refers to the ability of compactly sized deep neural networks to represent
functions capable of solving real-world problems. Optimization concerns the
effectiveness of simple gradient-based algorithms in solving non-convex
neural network training programs. Generalization treats the phenomenon of
deep learning models not overfitting despite having much more parameters
than examples to learn from. This talk will describe a series of works aimed
at unraveling some of the mysteries behind expressiveness and optimization.
I will begin by establishing an equivalence between convolutional and
recurrent networks --- the most successful deep learning architectures to
date --- and hierarchical tensor decompositions. The equivalence will be
used to answer various questions concerning expressiveness, resulting in new
theoretically-backed tools for deep network design. I will then turn to
discuss a recent line of work analyzing optimization of deep linear neural
networks. By studying the trajectories of gradient descent, we will derive
the most general guarantee to date for efficient convergence to global
minimum of a gradient-based algorithm training a deep network. Moreover, in
stark contrast with conventional wisdom, we will see that sometimes,
gradient descent can train a deep linear network faster than a classic
linear model. In other words, depth can accelerate optimization, even
without any gain in expressiveness, and despite introducing non-convexity to
a formerly convex problem.
The talk will be of a high-level nature, suitable for a diverse crowd. It
will cover works done in collaboration with Sanjeev Arora, Noah Golowich,
Elad Hazan, Wei Hu, Yoav Levine, Or Sharir, Amnon Shashua, Ronen Tamari and
David Yakira.
http://www.math.ias.edu/seminars/abstract?event=139466
3 An invitation to tensor networks
Michael Walter
Tensor networks describe high-dimensional tensors as the contraction of a
network (or graph) of low-dimensional tensors. Many interesting tensor can
be succinctly represented in this fashion -- from many-body ground states in
quantum physics to the matrix multiplication tensors in algebraic
complexity. I will give a mathematical introduction to the formalism, give
several examples, and sketch some of the most important results. We will
discuss the role of the network, how symmetries are encoded, tensor networks
as a computational model, and survey some recent algorithmic results.
http://www.math.ias.edu/seminars/abstract?event=129115
Computer Science/Discrete Math Seminars can be found on our web page:
http://www.math.ias.edu/csdm
http://www.math.ias.edu
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://imap.math.ias.edu/pipermail/iasmathsemrutgers/attachments/20181204/fc0fe4d8/attachment.html>
More information about the Iasmathsemrutgers
mailing list