
10th Annual Machine Learning Symposium
Friday, March 4, 2016
The New York Academy of Sciences
Machine Learning, a subfield of computer science, involves the development of mathematical algorithms that discover knowledge from specific data sets, and then "learn" from the data in an iterative fashion that allows predictions to be made. Today, Machine Learning has a wide range of applications, including natural language processing, search engine functionality, medical diagnosis, credit card fraud detection, and stock market analysis.
This symposium — the tenth in an ongoing series presented by the Machine Learning Discussion Group at the New York Academy of Sciences — will feature Keynote Presentations from leading scientists in both applied and theoretical Machine Learning.
Registration Pricing
Member | $60 |
Member (Student / Postdoc / Resident / Fellow) | $25 |
Nonmember (Academia) | $105 |
Nonmember (Corporate) | $160 |
Nonmember (Non-profit) | $105 |
Nonmember (Student / Postdoc / Resident / Fellow) | $70 |
Livestream
Three of the keynote addresses for this meeting will be presented via Livestream. For full details, and to view the Livestreams, use the link below:
Agenda
* Presentation titles and times are subject to change.
March 4, 2016 | |
9:00 AM | Registration, Continental Breakfast, and Poster Set-up |
10:00 AM | Welcome Remarks |
10:10 AM | Keynote Address 1: Online Methods for Learning in Social Networks |
10:50 AM | Audience Q&A |
Spotlight Talks: Session 1A series of short, early career investigator presentations across a variety of topics at the frontier of machine learning. Selected from Poster Abstracts. | |
11:05 AM | Kernel Extraction via Voted Risk Minimization |
11:10 AM | Multitask Matrix Completion for Learning Protein Interactions across Diseases |
11:15 AM | Learning Theory and Algorithms for Time Series Prediction |
11:20 AM | Continuous-time Limit of Stochastic Gradient Descent Revisited |
11:25 AM | Temporal Convolutional Neural Networks for Diagnosis from Lab Tests |
11:30 AM | Networking Break and Poster Viewing |
12:20 PM | Keynote Address 2: Personalized Mobile Health Interventions |
1:00 PM | Audience Q&A |
1:15 PM | Networking Lunch and Poster Viewing |
Spotlight Talks: Session 2 | |
2:30 PM | Infinite Factorial Dynamical Model |
2:35 PM | Efficient Algorithms for Adversarial Contextual Learning |
2:40 PM | Robust Bayesian Modeling via Generalized Empirical Bayes |
2:45 PM | Another Look at DWD: Thrifty Algorithm and Bayes Risk Consistency in RKHS |
2:50 PM | Multitask Learning Improves Prediction of Cancer Drug Sensitivity |
2:55 PM | Keynote Address 3: New Results at the Crossroads of Convexity, Learning and Information Theory |
3:35 PM | Audience Q&A |
3:50 PM | Networking Break |
4:05 PM | Keynote Address 4: Smooth Operators: The Rise of Differentiable Attention in Deep Learning |
4:45 PM | Audience Q&A |
5:00 PM | Closing Remarks and AwardsMachine Learning Challenge Award Presentation Awardees: Best Poster Presentation Awardees: Spotlight Talk Award Presentation Awardees: |
5:10 PM | Networking Reception |
6:30 PM | Symposium Adjourns |
Keynote Speakers
Sébastien Bubeck, PhD
Microsoft Research
Sebastien Bubeck is a researcher in the Theory Group at Microsoft Research. Prior to MSR he was an assistant professor in the ORFE department at Princeton University. His research has won a few awards, including the 2010 Jacques Neveu prize (best French Ph.D. in probability/statistics) and a 2015 Sloan Fellowship in Computer Science.
Alex Graves, PhD
Google DeepMind
Alex Graves is a research scientist at Google DeepMind. His work focuses on developing recurrent neural networks for sequence learning, and now features prominently in areas such as speech recognition, handwriting synthesis, and generative sequence modelling. Alex has done a BSc in Theoretical Physics at Edinburgh, Part III Maths at Cambridge, a PhD in AI at IDSIA with Juergen Schmidhuber, followed by postdocs at TU–Munich and with Geoff Hinton at the University of Toronto. Most recently he has been spearheading DeepMind's development of Neural Turing Machines.
Alexander Rakhlin, PhD
University of Pennsylvania
Alexander (Sasha) Rakhlin is an Associate Professor at the Department of Statistics at the University of Pennsylvania. He received his bachelor’s degrees in Mathematics and Computer Science from Cornell University, and a doctoral degree from MIT. He was a postdoc at UC Berkeley before joining UPenn. His research is in machine learning, with an emphasis on online methods, statistics, and computation. Alexander is a recipient of the NSF CAREER award, IBM Research Best Paper award, Machine Learning Journal award, and COLT Best Paper Award.
Ambuj Tewari, PhD
University of Michigan
Ambuj Tewari is an assistant professor in the Department of Statistics and the Department of EECS at the University of Michigan, Ann Arbor. His is also affiliated with the Michigan Institute of Data Science (MIDAS). His research interests lie in machine learning including statistical learning theory, online learning, reinforcement learning and control theory, network analysis, and optimization for machine learning. He collaborates with scientists to seek novel applications of machine learning in mobile health, learning analytics, and computational chemistry. His research has been recognized with paper awards at COLT 2005, COLT 2011, and AISTATS 2015. He was the recipient of an NSF CAREER award in 2015.
Scientific Organizing Committee
Naoki Abe, PhD
IBM Research
Naoki Abe is a principal research staff member and senior manager of “Data Analytics” within the Solutions and Mathematical Sciences Department. He obtained his B.S. and M.S. degrees in computer science from MIT in 1984, and Ph. D. in computer science from University of Pennsylvania in 1989. He has been with IBM Research since 2001, conducting research in the development of novel machine learning methodologies that open up new applications in business analytics and optimization. His research activities range from applications of reinforcement learning to business analytics, methods for anomaly detection and temporal causal modeling. Methodologies that he co-developed through these research efforts have made their ways into a number of IBM offerings, including the Tax Collections Optimizer and signature solutions “Next Best Action” and “CFO-dashboard.”
Corinna Cortes, PhD
Google Research
Corinna Cortes is the Head of Google Research, New York, where she works on a broad range of theoretical and applied large-scale machine learning problems. Prior to Google, Cortes spent more than ten years at AT&T Labs - Research, formerly AT&T Bell Labs, where she held a distinguished research position. Her research work is well-known in particular for her contributions to the theoretical foundations of support vector machines (SVMs), for which she jointly with Vladimir Vapnik received the 2008 Paris Kanellakis Theory and Practice Award, and her work on data-mining in very large data sets for which she was awarded the AT&T Science and Technology Medal in the year 2000. She received her master’s degree in physics from the University of Copenhagen and joined AT&T Bell Labs as a researcher in 1989. She received her PhD in computer science from the University of Rochester in 1993. Cortes is also a competitive runner, and a mother of two.
Jennifer L. Costley, PhD
The New York Academy of Sciences
Patrick Haffner, PhD
Interactions Corporation
Tony Jebara, PhD
Columbia University
Tony Jebara is Associate Professor of Computer Science at Columbia University and Director of Machine Learning at Netflix. His research intersects computer science and statistics to develop new frameworks for learning from data with applications in recommendation, social networks, spatio-temporal data, vision and text. Jebara has founded and advised several startups including Sense Networks (acquired by yp.com), Evidation Health, Agolo, Ufora, MagikEye, and Bookt (acquired by RealPage NASDAQ:RP). He has published over 100 peer-reviewed papers in conferences, workshops and journals (such as NIPS, ICML, UAI, COLT, JMLR, CVPR, ICCV, and AISTAT) as well as the book Machine Learning: Discriminative and Generative. Jebara was the recipient of the Career award from the National Science Foundation, a best paper award at the 26th International Conference on Machine Learning, a best student paper award at the 20th International Conference on Machine Learning as well as an outstanding contribution award from the Pattern Recognition Society in 2001. Jebara's research has been featured on ABC, BBC, the New York Times, Slash Dot, Wired, Businessweek, IEEE Spectrum and more. He obtained his PhD in 2002 from MIT. Esquire magazine named him one of their Best and Brightest of 2008.
John Langford, PhD
Microsoft Research
Mehryar Mohri, PhD
Courant Institute of Mathematical Sciences, New York University
Gunnar Rätsch, PhD
Memorial Sloan-Kettering Cancer Center
Robert Schapire, PhD
Microsoft Research
Rob Schapire is a Principal Researcher at Microsoft Research in New York City. He received his PhD from MIT in 1991. After a short post-doc at Harvard, he joined the technical staff at AT&T Labs (formerly AT&T Bell Laboratories) in 1991. In 2002, he became a Professor of Computer Science at Princeton University. He joined Microsoft Research in 2014. His awards include the 1991 ACM Doctoral Dissertation Award, the 2003 Gödel Prize, and the 2004 Kanelakkis Theory and Practice Award (both of the last two with Yoav Freund). He is a fellow of the AAAI, and a member of the National Academy of Engineering. His main research interest is in theoretical and applied machine learning, with particular focus on boosting, online learning, game theory, and maximum entropy. For more information, see http://rob.schapire.net.
Di Xu, PhD
American Express Corporation
Sponsors
For sponsorship opportunities please contact Dr. Jennifer Costley at jcostley@nyas.org or 212.298.8675.
Gold Sponsor
Bronze Sponsors
IBM Watson
Academy Friends
Presented by
Abstracts
Keynote Speakers
Online Methods for Learning in Social Networks
Alexander Rakhlin, PhD, University of Pennsylvania
We consider the problem of predicting a binary label for an individual given the information about the person and her position within a network. Such a fusion of the two sources of information naturally arises in a variety of applications, including recommendation systems, ad placement, and personalized medical care.
When formalizing this problem, one faces a computationally intractable combinatorial objective. We present an unexpected phenomenon: it is possible to develop poly-time (and statistically near-optimal) online prediction methods even when the offline problem is provably hard. These prediction methods arise in a systematic way from a new online relaxation framework. Our approach naturally extends to the contextual multi-armed bandit setting with large sets of policies—a notoriously difficult problem, which is often encountered in real-world applications.
Joint work with K. Sridharan
Personalized Mobile Health Interventions
Ambuj Tewari, PhD, University of Michigan, Ann Arbor
Mobile health (mHealth) refers to the use of mobile devices in accessing and delivering healthcare services. The growing use of mobile phones worldwide along with increasing barriers to receiving quality healthcare is sparking tremendous interest in mHealth both in academia as well as in industry. However, to realize the true potential of mHealth, we need interdisciplinary collaborations between specialists in behavioral sciences, human computer interaction, machine learning and statistics. I will describe Just-in-time Adaptive Intervention (JITAI), a conceptual framework designed to facilitate interdisciplinary communication in mHealth. I will also highlight the role of reinforcement learning algorithms in personalizing mHealth interventions to suit the needs of individuals at particular moments in their daily lives.
New Results at the Crossroads of Convexity, Learning and Information Theory
Sebastien Bubeck, PhD, Theory Group, Microsoft Research
I will present three new results: (i) the Cramer transform of the uniform measure on a convex body is a universal self-concordant barrier; (ii) projected gradient descent with Gaussian noise allows to sample from a log-concave measure in polynomial time; and (iii) Thompson sampling combined with a multi-scale exploration solves the Bayesian convex bandit problem. The unifying theme in these results is the interplay between concepts from convex geometry, learning and information theory.
Smooth Operators: The Rise of Differentiable Attention in Deep Learning
Alex Graves, PhD, Google DeepMind
The ability to focus on one thing and ignore others plays a vital role in guiding human cognition. As well as allowing us to pick out salient information from noisy data, selective attention makes it possible to pursue one thought at a time, to remember one event rather than all events. This talk charts the recent introduction of smooth attention mechanisms into deep learning, which has made it feasible for neural networks to learn how to filter both external data and their own internal representations and manipulations of prior data. Crucially, because these mechanisms are differentiable, they can be trained with gradient descent, using only input and output examples. The talk will review a range of range of neural networks incorporating various forms of textual, visual and internal attention.
Travel & Lodging
Our Location
The New York Academy of Sciences
7 World Trade Center
250 Greenwich Street, 40th floor
New York, NY 10007-2157
212.298.8600
Hotels Near 7 World Trade Center
Recommended partner hotel
Club Quarters, World Trade Center
140 Washington Street
New York, NY 10006
Phone: 212.577.1133
The New York Academy of Sciences is a member of the Club Quarters network, which offers significant savings on hotel reservations to member organizations. Located opposite Memorial Plaza on the south side of the World Trade Center, Club Quarters, World Trade Center is just a short walk to the Academy.
Use Club Quarters Reservation Password NYAS to reserve your discounted accommodations online.
Other nearby hotels
212.945.0100 | |
212.693.2001 | |
212.385.4900 | |
212.269.6400 | |
212.742.0003 | |
212.232.7700 | |
212.747.1500 | |
212.344.0800 |