Nash, Nyquist, Networks, Neural Networks, and… Next?

From data to robust intelligent systems, sample by sample—an engineering research seminar with Peter Chin, Research Professor, Boston College

February 17, 2021
12 pm - 1 pm
Location
Videoconference
Sponsored by
Thayer School of Engineering
Audience
Public
More information
Ashley Parker

ZOOM LINK
Meeting ID: 967 0890 8163    
Passcode: 300932

Our world is becoming increasingly more data-driven and more data-intensive by the day, and the need to handle such deluge of data (images, videos, texts, etc.) and make sense of them is more urgent than ever. Resurgence of AI (artificial intelligence) and ML (machine learning) research of the past decade after the years of the so-called AI winter is in part motivated by such a need, though output of such ML algorithms themselves often generate even more data (chatbots, Tik-Tok videos, etc.) with more and more intelligent agents (human and otherwise) interacting with each other either cooperatively or adversarially over such data. 

Understanding the shape of data (along with information and even intelligence therein contained) may help. Standing tall among the major mathematical achievements of 20th century are some theorems whose subsequent impacts have far outweighed their original intent. One such theorem is due to John Nash, whose proof of the existence of equilibrium in a non-cooperative game gave rise to the concept of the eponymous Nash Equilibrium, which in many ways revolutionized the field of economics. Less known but equally impactful also by Nash is Nash embedding theorem, which states that every Riemannian manifold can be isometrically embedded into some high dimensional Euclidean space. Yet another is due to Harry Nyquist, whose Nyquist-Shannon sampling theorem, which states that every time-varying band-limited signal can be perfectly reconstructed from a sequence of samples acquired at the twice rate of its maximum frequency, laid the foundation of the modern information and communication theory.  Furthermore, when the data can be described as a large-scale dense graph, Szemeredi’s regularity lemma has proved to be a powerful tool for revealing the structures of such graphs (in terms of partitions), despite being called just a “lemma”...

I will describe some recent progress in extending these results, for example, in dynamic game theory, where the rules of the game change over time, and in the theory of compressive sensing, which guarantees perfect reconstruction of signals from far fewer number of samples than required by the Nyquist theorem, if the signals are sparse in some appropriate domain. I will then describe some applications of these extensions for machine learning, brain imaging, community detection, etc. and finally hint at some potential connection between some of these these important theorems.

Location
Videoconference
Sponsored by
Thayer School of Engineering
Audience
Public
More information
Ashley Parker