BEGIN:VCALENDAR
VERSION:2.0
PRODID:-//IFDS - ECPv6.0.1.1//NONSGML v1.0//EN
CALSCALE:GREGORIAN
METHOD:PUBLISH
X-ORIGINAL-URL:https://ifds.info
X-WR-CALDESC:Events for IFDS
REFRESH-INTERVAL;VALUE=DURATION:PT1H
X-Robots-Tag:noindex
X-PUBLISHED-TTL:PT1H
BEGIN:VTIMEZONE
TZID:America/Chicago
BEGIN:DAYLIGHT
TZOFFSETFROM:-0600
TZOFFSETTO:-0500
TZNAME:CDT
DTSTART:20230312T080000
END:DAYLIGHT
BEGIN:STANDARD
TZOFFSETFROM:-0500
TZOFFSETTO:-0600
TZNAME:CST
DTSTART:20231105T070000
END:STANDARD
BEGIN:DAYLIGHT
TZOFFSETFROM:-0600
TZOFFSETTO:-0500
TZNAME:CDT
DTSTART:20240310T080000
END:DAYLIGHT
BEGIN:STANDARD
TZOFFSETFROM:-0500
TZOFFSETTO:-0600
TZNAME:CST
DTSTART:20241103T070000
END:STANDARD
TZID:America/Los_Angeles
BEGIN:DAYLIGHT
TZOFFSETFROM:-0800
TZOFFSETTO:-0700
TZNAME:PDT
DTSTART:20240310T100000
END:DAYLIGHT
BEGIN:STANDARD
TZOFFSETFROM:-0700
TZOFFSETTO:-0800
TZNAME:PST
DTSTART:20241103T090000
END:STANDARD
END:VTIMEZONE
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20231115T123000
DTEND;TZID=America/Chicago:20231115T133000
DTSTAMP:20260407T033447
CREATED:20240315T164000Z
LAST-MODIFIED:20240315T173311Z
UID:2823-1700051400-1700055000@ifds.info
SUMMARY:SILO: A Theory of Universal Learning
DESCRIPTION:Purdue\, Steve Hanneke
URL:https://ifds.info/event/silo-a-theory-of-universal-learning/
LOCATION:Orchard View Room\, 330 N. Orchard Street\, 3rd Floor NE\, Madison\, Wisconsin\, 53715\, United States
CATEGORIES:SILO
ATTACH;FMTTYPE=image/png:https://ifds.info/wp-content/uploads/2022/10/SILO-1024x683-1-e1665597390709.png
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20231120T123000
DTEND;TZID=America/Chicago:20231120T123000
DTSTAMP:20260407T033447
CREATED:20231024T213142Z
LAST-MODIFIED:20231024T215229Z
UID:2705-1700483400-1700483400@ifds.info
SUMMARY:IFDS Ideas Forum
DESCRIPTION:TBD\nSpeakers: Karan Srivastava / Chenghui Li
URL:https://ifds.info/event/short-talk-exact-title-tbd-3/
LOCATION:Orchard View Room\, 330 N. Orchard Street\, 3rd Floor NE\, Madison\, Wisconsin\, 53715\, United States
CATEGORIES:IFDS Ideas Forum
ATTACH;FMTTYPE=image/jpeg:https://ifds.info/wp-content/uploads/2023/10/iStock-1415961311.jpg
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20231122T123000
DTEND;TZID=America/Chicago:20231122T133000
DTSTAMP:20260407T033447
CREATED:20240315T164000Z
LAST-MODIFIED:20240315T173404Z
UID:2824-1700656200-1700659800@ifds.info
SUMMARY:SILO: World Knowledge in the Time of Large Models
DESCRIPTION:DeepMind\, Kenneth Marino
URL:https://ifds.info/event/silo-world-knowledge-in-the-time-of-large-models/
LOCATION:Orchard View Room\, 330 N. Orchard Street\, 3rd Floor NE\, Madison\, Wisconsin\, 53715\, United States
CATEGORIES:SILO
ATTACH;FMTTYPE=image/png:https://ifds.info/wp-content/uploads/2022/10/SILO-1024x683-1-e1665597390709.png
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20231127T123000
DTEND;TZID=America/Chicago:20231127T123000
DTSTAMP:20260407T033447
CREATED:20231024T212227Z
LAST-MODIFIED:20231024T215121Z
UID:2686-1701088200-1701088200@ifds.info
SUMMARY:IFDS Ideas Forum
DESCRIPTION:TBD\nSpeaker: Zarifis Nikos
URL:https://ifds.info/event/ifds-ideas-forum-47/
CATEGORIES:IFDS Ideas Forum
ATTACH;FMTTYPE=image/jpeg:https://ifds.info/wp-content/uploads/2023/10/iStock-1415961311.jpg
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20231129T123000
DTEND;TZID=America/Chicago:20231129T133000
DTSTAMP:20260407T033447
CREATED:20240315T164000Z
LAST-MODIFIED:20240315T173459Z
UID:2825-1701261000-1701264600@ifds.info
SUMMARY:SILO: Understanding deep nets: on local Lipschitz functions and learned proximal networks
DESCRIPTION:Jeremias Sulam\, JHU
URL:https://ifds.info/event/silo-understanding-deep-nets-on-local-lipschitz-functions-and-learned-proximal-networks/
LOCATION:Orchard View Room\, 330 N. Orchard Street\, 3rd Floor NE\, Madison\, Wisconsin\, 53715\, United States
CATEGORIES:SILO
ATTACH;FMTTYPE=image/png:https://ifds.info/wp-content/uploads/2022/10/SILO-1024x683-1-e1665597390709.png
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20231204T123000
DTEND;TZID=America/Chicago:20231204T123000
DTSTAMP:20260407T033447
CREATED:20231024T213142Z
LAST-MODIFIED:20231024T215321Z
UID:2706-1701693000-1701693000@ifds.info
SUMMARY:IFDS Ideas Forum
DESCRIPTION:TBD\nSpeakers: Jitian Zhao / Zhexuan Liu
URL:https://ifds.info/event/short-talk-exact-title-tbd-4/
LOCATION:Orchard View Room\, 330 N. Orchard Street\, 3rd Floor NE\, Madison\, Wisconsin\, 53715\, United States
CATEGORIES:IFDS Ideas Forum
ATTACH;FMTTYPE=image/jpeg:https://ifds.info/wp-content/uploads/2023/10/iStock-1415961311.jpg
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20231206T123000
DTEND;TZID=America/Chicago:20231206T133000
DTSTAMP:20260407T033447
CREATED:20240315T164000Z
LAST-MODIFIED:20240315T173546Z
UID:2826-1701865800-1701869400@ifds.info
SUMMARY:SILO: Invariant Low-Dimensional Subspaces in Gradient Descent for Learning Deep Networks
DESCRIPTION:Qing Qu\, UMich
URL:https://ifds.info/event/silo-invariant-low-dimensional-subspaces-in-gradient-descent-for-learning-deep-networks/
LOCATION:Orchard View Room\, 330 N. Orchard Street\, 3rd Floor NE\, Madison\, Wisconsin\, 53715\, United States
CATEGORIES:SILO
ATTACH;FMTTYPE=image/png:https://ifds.info/wp-content/uploads/2022/10/SILO-1024x683-1-e1665597390709.png
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20231211T123000
DTEND;TZID=America/Chicago:20231211T123000
DTSTAMP:20260407T033447
CREATED:20231024T213143Z
LAST-MODIFIED:20231024T215415Z
UID:2707-1702297800-1702297800@ifds.info
SUMMARY:IFDS Ideas Forum
DESCRIPTION:TBD\nSpeaker: Thanasis Pittas
URL:https://ifds.info/event/tbd-2/
LOCATION:Orchard View Room\, 330 N. Orchard Street\, 3rd Floor NE\, Madison\, Wisconsin\, 53715\, United States
CATEGORIES:IFDS Ideas Forum
ATTACH;FMTTYPE=image/jpeg:https://ifds.info/wp-content/uploads/2023/10/iStock-1415961311.jpg
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20231218T123000
DTEND;TZID=America/Chicago:20231218T123000
DTSTAMP:20260407T033447
CREATED:20231024T213157Z
LAST-MODIFIED:20231024T215501Z
UID:2708-1702902600-1702902600@ifds.info
SUMMARY:IFDS Ideas Forum
DESCRIPTION:TBD\nSpeaker: Sushrut Karmalkar
URL:https://ifds.info/event/tbd-3/
LOCATION:Orchard View Room\, 330 N. Orchard Street\, 3rd Floor NE\, Madison\, Wisconsin\, 53715\, United States
CATEGORIES:IFDS Ideas Forum
ATTACH;FMTTYPE=image/jpeg:https://ifds.info/wp-content/uploads/2023/10/iStock-1415961311.jpg
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20240124T123000
DTEND;TZID=America/Chicago:20240124T133000
DTSTAMP:20260407T033447
CREATED:20240315T164000Z
LAST-MODIFIED:20240315T173629Z
UID:2827-1706099400-1706103000@ifds.info
SUMMARY:SILO: Funsearch: a novel approach to using large language models as part of mathematical practice
DESCRIPTION:Jordan Ellenberg
URL:https://ifds.info/event/silo-funsearch-a-novel-approach-to-using-large-language-models-as-part-of-mathematical-practice/
LOCATION:Orchard View Room\, 330 N. Orchard Street\, 3rd Floor NE\, Madison\, Wisconsin\, 53715\, United States
CATEGORIES:SILO
ATTACH;FMTTYPE=image/png:https://ifds.info/wp-content/uploads/2022/10/SILO-1024x683-1-e1665597390709.png
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Los_Angeles:20240126T133000
DTEND;TZID=America/Los_Angeles:20240126T143000
DTSTAMP:20260407T033447
CREATED:20240318T212134Z
LAST-MODIFIED:20240318T212230Z
UID:2879-1706275800-1706279400@ifds.info
SUMMARY:How do neural networks learn features from data?
DESCRIPTION:Speaker Bio: Adit is currently the George F. Carrier Postdoctoral Fellow in the School of Engineering and Applied Sciences at Harvard. He completed his Ph.D. in electrical engineering and computer science (EECS) at MIT advised by Caroline Uhler and was a Ph.D. fellow at the Eric and Wendy Schmidt Center at the Broad Institute of MIT and Harvard. His research focuses on advancing theoretical foundations of machine learning and developing new methods for tackling biomedical problems. \n\n\n\nAbstract: Understanding how neural networks learn features\, or relevant patterns in data\, for prediction is necessary for their reliable use in technological and scientific applications. We propose a unifying mechanism that characterizes feature learning in neural network architectures. Namely\, we show that features learned by neural networks are captured by a statistical operator known as the average gradient outer product (AGOP). Empirically\, we show that the AGOP captures features across a broad class of network architectures including convolutional networks and large language models. Moreover\, we use AGOP to enable feature learning in general machine learning models through an algorithm we call Recursive Feature Machine (RFM). We show that RFM automatically identifies sparse subsets of features relevant for prediction and explicitly connects feature learning in neural networks with classical sparse recovery and low rank matrix factorization algorithms. Overall\, this line of work advances our fundamental understanding of how neural networks extract features from data\, leading to the development of novel\, interpretable\, and effective models for use in scientific applications.
URL:https://ifds.info/event/how-do-neural-networks-learn-features-from-data/
LOCATION:CSE (Allen) 403
CATEGORIES:MLOpt@UWash
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20240131T123000
DTEND;TZID=America/Chicago:20240131T133000
DTSTAMP:20260407T033447
CREATED:20240315T164015Z
LAST-MODIFIED:20240315T172723Z
UID:2828-1706704200-1706707800@ifds.info
SUMMARY:SILO: A scalable method to exploit screening in Gaussian process models with noise
DESCRIPTION:Chris Geoga
URL:https://ifds.info/event/silo-a-scalable-method-to-exploit-screening-in-gaussian-process-models-with-noise/
LOCATION:Orchard View Room\, 330 N. Orchard Street\, 3rd Floor NE\, Madison\, Wisconsin\, 53715\, United States
CATEGORIES:SILO
ATTACH;FMTTYPE=image/png:https://ifds.info/wp-content/uploads/2022/10/SILO-1024x683-1-e1665597390709.png
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20240205T123000
DTEND;TZID=America/Chicago:20240205T133000
DTSTAMP:20260407T033447
CREATED:20240318T213252Z
LAST-MODIFIED:20240318T213252Z
UID:2892-1707136200-1707139800@ifds.info
SUMMARY:Towards a new toolbox of optimal statistical primitives
DESCRIPTION:Abstract: Given society’s increasing reliance on data\, its collection and processing into useful information is a technical problem of growing focus\, and perhaps paradoxically\, a critical bottleneck in many data science and machine learning applications. My research focuses on designing algorithms that push the limits of both statistical efficiency and computational efficiency. In particular\, my work tackles the divide between the theory and practice of data science\, which exists even for the most basic statistical problems including mean and (co)variance estimation. Conventional methods such as the sample mean\, while supported by theoretical results under strong assumptions\, are often brittle in the presence of extreme data points. To counter such deficiencies\, practitioners often use ad-hoc and unprincipled “outlier removal” heuristics\, revealing a marked gap between the theory and practice even for these fundamental problems. \nIn this talk\, I will describe my work towards building a new toolbox of optimal statistical primitives\, bridging the theory-practice divide. I will specifically highlight 3 works: A) constructing a statistically-optimal and computationally-efficient 1-dimensional mean estimator\, whose estimation error is optimal even in the leading multiplicative constant\, under bare minimum distributional assumptions\, B) a rather different but also optimal mean estimator for the “very high-dimensional” regime\, and C) a recent result on robustly clustering Gaussian mixtures based on their covariances even in the presence of adversarial data corruption. To conclude the talk\, I will discuss my vision for the new theory and toolbox\, serving as a blueprint for my long-term future research. \nBio: Jasper Lee is a postdoctoral research associate at the University of Wisconsin-Madison\, mentored by Ilias Diakonikolas in the Department of Computer Sciences\, and also affiliated with the Institute for Foundations of Data Science. He completed his PhD at Brown University\, advised by Paul Valiant. \nHis research interests are broadly in the foundations of data science\, aiming to design practical\, data-efficient and computationally-efficient algorithms for a variety of statistical applications. \nHis work is partially supported by a Croucher Fellowship for Postdoctoral Research.
URL:https://ifds.info/event/towards-a-new-toolbox-of-optimal-statistical-primitives/
LOCATION:Orchard View Room\, 330 N. Orchard Street\, 3rd Floor NE\, Madison\, Wisconsin\, 53715\, United States
CATEGORIES:IFDS Ideas Forum
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20240207T123000
DTEND;TZID=America/Chicago:20240207T133000
DTSTAMP:20260407T033447
CREATED:20240315T164015Z
LAST-MODIFIED:20240315T172944Z
UID:2829-1707309000-1707312600@ifds.info
SUMMARY:SILO: Universality in High-Dimensional Statistics
DESCRIPTION:Rishabh Dudeja
URL:https://ifds.info/event/silo-universality-in-high-dimensional-statistics/
LOCATION:Orchard View Room\, 330 N. Orchard Street\, 3rd Floor NE\, Madison\, Wisconsin\, 53715\, United States
CATEGORIES:SILO
ATTACH;FMTTYPE=image/png:https://ifds.info/wp-content/uploads/2022/10/SILO-1024x683-1-e1665597390709.png
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Los_Angeles:20240209T133000
DTEND;TZID=America/Los_Angeles:20240209T143000
DTSTAMP:20260407T033447
CREATED:20240318T212503Z
LAST-MODIFIED:20240318T212503Z
UID:2882-1707485400-1707489000@ifds.info
SUMMARY:Policy Optimization with Compatible Mirror Approximation
DESCRIPTION:Speaker Bio: Zhihan is a fourth-year PhD student in the Paul G. Allen School of Computer Science & Engineering at University of Washington\, advised by Prof. Maryam Fazel. His research interests are broadly in statistics\, optimization and machine learning. \n\n\nAbstract: We propose Compatible Mirror Policy Optimization (CoMPO)\, a framework that incorporates general function approximation into policy mirror descent methods. In contrast to the popular approach of using the $L_2$ norm to measure function approximation errors (regardless of the mirror map)\, CoMPO uses the Bregman divergence induced by the specific mirror map for policy projection. Such a compatibility bridges the gap between theory and practice: not only does it achieve fast linear convergence with general function approximation\, but it also includes several well-known practical methods as special cases\, immediately providing them strong convergence guarantees.
URL:https://ifds.info/event/policy-optimization-with-compatible-mirror-approximation/
LOCATION:Zoom
CATEGORIES:MLOpt@UWash
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20240212T123000
DTEND;TZID=America/Chicago:20240212T133000
DTSTAMP:20260407T033447
CREATED:20240318T213417Z
LAST-MODIFIED:20240318T213417Z
UID:2894-1707741000-1707744600@ifds.info
SUMMARY:Theoretical exploration of foundation model adaption methods
DESCRIPTION:Speaker: Kangwook Lee
URL:https://ifds.info/event/theoretical-exploration-of-foundation-model-adaption-methods/
LOCATION:Orchard View Room\, 330 N. Orchard Street\, 3rd Floor NE\, Madison\, Wisconsin\, 53715\, United States
CATEGORIES:IFDS Ideas Forum
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20240214T123000
DTEND;TZID=America/Chicago:20240214T133000
DTSTAMP:20260407T033447
CREATED:20240315T164015Z
LAST-MODIFIED:20240315T173037Z
UID:2830-1707913800-1707917400@ifds.info
SUMMARY:SILO: Theoretical Exploration of Foundation Model Adaptation Methods
DESCRIPTION:Kangwook Lee
URL:https://ifds.info/event/silo-theoretical-exploration-of-foundation-model-adaptation-methods/
LOCATION:Orchard View Room\, 330 N. Orchard Street\, 3rd Floor NE\, Madison\, Wisconsin\, 53715\, United States
CATEGORIES:SILO
ATTACH;FMTTYPE=image/png:https://ifds.info/wp-content/uploads/2022/10/SILO-1024x683-1-e1665597390709.png
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Los_Angeles:20240216T133000
DTEND;TZID=America/Los_Angeles:20240216T143000
DTSTAMP:20260407T033447
CREATED:20240318T212625Z
LAST-MODIFIED:20240318T212625Z
UID:2884-1708090200-1708093800@ifds.info
SUMMARY:Offline Multi-task Transfer RL with Representational Penalization
DESCRIPTION:Speaker Bio: Avinandan is a second year PhD student\, advised by Maryam Fazel and Lillian Ratliff. His interests are in sequential learning and game theory. \n\n\nAbstract: We study the problem of representational transfer in offline Reinforcement Learning (RL)\, where a learner has access to episodic data from a number of source tasks collected a priori\, and aims to learn a shared representation to be used in finding a good policy for a target task. Unlike in online RL where the agent interacts with the environment while learning a policy\, in the offline setting there cannot be such interactions in either the source tasks or the target task; thus multi-task offline RL can suffer from incomplete coverage.We propose an algorithm to compute pointwise uncertainty measures for the learnt representation\, and establish a data-dependent upper bound for the suboptimality of the learnt policy for the target task. Our algorithm leverages the collective exploration done by source tasks to mitigate poor coverage at some points by a few tasks\, thus overcoming the limitation of needing uniformly good coverage for a meaningful transfer by existing offline algorithms. We complement our theoretical results with empirical evaluation on a rich-observation MDP which requires many samples for complete coverage. Our findings illustrate the benefits of penalizing and quantifying the uncertainty in the learnt representation.
URL:https://ifds.info/event/offline-multi-task-transfer-rl-with-representational-penalization/
LOCATION:CSE (Allen) 403
CATEGORIES:MLOpt@UWash
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20240219T123000
DTEND;TZID=America/Chicago:20240219T133000
DTSTAMP:20260407T033447
CREATED:20240318T213630Z
LAST-MODIFIED:20240318T213656Z
UID:2897-1708345800-1708349400@ifds.info
SUMMARY:A good score does not lead to a good generative model
DESCRIPTION:Speaker: Sixu Li \nAbstract: Score-based Generative Models (SGMs) is one leading method in generative modeling\, renowned for their ability to generate high-quality samples from complex\, high-dimensional data distributions. The method enjoys empirical success and is supported by rigorous theoretical convergence properties. In particular\, it has been shown that SGMs can generate samples from a distribution that is close to the ground-truth if the underlying score function is learned well\, suggesting the success of SGM as a generative model. We provide a counter-example in this paper. Through the sample complexity argument\, we provide one specific setting where the score function is learned well. Yet\, SGMs in this setting can only output samples that are Gaussian blurring of training data points\, mimicking the effects of kernel density estimation. The finding resonates a series of recent finding that reveal that SGMs can demonstrate strong memorization effect and fail to generate. This is joint with Shi Chen and Qin Li. 
URL:https://ifds.info/event/a-good-score-does-not-lead-to-a-good-generative-model/
LOCATION:WID 1145\, 330 N Orchard Street\, Madison\, WI\, 53715\, United States
CATEGORIES:IFDS Ideas Forum
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20240221T123000
DTEND;TZID=America/Chicago:20240221T133000
DTSTAMP:20260407T033447
CREATED:20240315T164015Z
LAST-MODIFIED:20240315T173126Z
UID:2831-1708518600-1708522200@ifds.info
SUMMARY:SILO: Foundations of Real-World Reinforcement Learning
DESCRIPTION:Jeongyeol Kwon
URL:https://ifds.info/event/silo-foundations-of-real-world-reinforcement-learning/
LOCATION:Orchard View Room\, 330 N. Orchard Street\, 3rd Floor NE\, Madison\, Wisconsin\, 53715\, United States
CATEGORIES:SILO
ATTACH;FMTTYPE=image/png:https://ifds.info/wp-content/uploads/2022/10/SILO-1024x683-1-e1665597390709.png
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Los_Angeles:20240223T133000
DTEND;TZID=America/Los_Angeles:20240223T143000
DTSTAMP:20260407T033447
CREATED:20240318T212752Z
LAST-MODIFIED:20240318T212752Z
UID:2886-1708695000-1708698600@ifds.info
SUMMARY:GumbelSpec Sampling for Accelerating LLM Inference
DESCRIPTION:Bio: Tianxiao Shen is a postdoctoral scholar at the University of Washington\, working with Yejin Choi and Zaid Harchaoui. Her research interests lie in natural language processing and machine learning\, in particular developing models and algorithms for efficient\, accurate\, diverse\, flexible and controllable text generation. She received her PhD from MIT\, advised by Regina Barzilay and Tommi Jaakkola. Before that\, she did her undergrad at Tsinghua University. \n\n\n\n\nAbstract: We propose GumbelSpec sampling\, a novel algorithm that leverages smaller language models to accelerate inference of large language models without changing their output distribution. Central to our approach is the application of the Gumbel-Softmax technique to convert the stochastic decoding process into a deterministic process by integrating independently sampled Gumbel noise. Employing the same set of Gumbel noise\, we perform beam search on the smaller model to generate multiple candidate short continuations\, and then utilize tree-based attention to efficiently verify them in parallel using the larger model. GumbelSpec sampling significantly improves upon previous rejection sampling based speculative decoding methods by increasing the token acceptance rate by 1.7x-2.2x and achieving an additional speedup of 1.2x-1.5x. This results in a total speedup of 1.5x-2.6x compared to traditional autoregressive decoding.
URL:https://ifds.info/event/gumbelspec-sampling-for-accelerating-llm-inference/
LOCATION:CSE (Allen) 403
CATEGORIES:MLOpt@UWash
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20240226T123000
DTEND;TZID=America/Chicago:20240226T133000
DTSTAMP:20260407T033447
CREATED:20240318T213825Z
LAST-MODIFIED:20240318T213825Z
UID:2900-1708950600-1708954200@ifds.info
SUMMARY:Prelimit coupling and steady-state convergence of constant-stepsize nonsmooth contractive SA
DESCRIPTION:Speaker: Yixuan Zhang \n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nAbstract:  \nMotivated by Q-learning\, we study nonsmooth contractive stochastic approximation (SA) with constant stepsize. We focus on two important classes of dynamics: 1) nonsmooth contractive SA with additive noise\, and 2) synchronous and asynchronous Q-learning\, which features both additive and multiplicative noise. For both dynamics\, we establish weak convergence of the iterates to a stationary limit distribution in Wasserstein distance. Furthermore\, we propose a prelimit coupling technique for establishing steady-state convergence and characterize the limit of the stationary distribution as the stepsize goes to zero. Using this result\, we derive that the asymptotic bias of nonsmooth SA is proportional to the square root of the stepsize\, which stands in sharp contrast to smooth SA. This bias characterization allows for the use of Richardson-Romberg extrapolation for bias reduction in nonsmooth SA.
URL:https://ifds.info/event/prelimit-coupling-and-steady-state-convergence-of-constant-stepsize-nonsmooth-contractive-sa/
LOCATION:WID 1145\, 330 N Orchard Street\, Madison\, WI\, 53715\, United States
CATEGORIES:IFDS Ideas Forum
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20240228T123000
DTEND;TZID=America/Chicago:20240228T133000
DTSTAMP:20260407T033447
CREATED:20240315T164015Z
LAST-MODIFIED:20240315T173213Z
UID:2832-1709123400-1709127000@ifds.info
SUMMARY:SILO: Reinforcement Learning with Robustness and Safety Guarantees
DESCRIPTION:Dileep Kalathil\, TAMU
URL:https://ifds.info/event/silo-reinforcement-learning-with-robustness-and-safety-guarantees/
LOCATION:Orchard View Room\, 330 N. Orchard Street\, 3rd Floor NE\, Madison\, Wisconsin\, 53715\, United States
CATEGORIES:SILO
ATTACH;FMTTYPE=image/png:https://ifds.info/wp-content/uploads/2022/10/SILO-1024x683-1-e1665597390709.png
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20240304T123000
DTEND;TZID=America/Chicago:20240304T133000
DTSTAMP:20260407T033447
CREATED:20240318T214117Z
LAST-MODIFIED:20240318T214117Z
UID:2902-1709555400-1709559000@ifds.info
SUMMARY:IFDS Ideas Forum Double Header
DESCRIPTION:Speaker 1: Puqian Wang \nRobustly Learning Single-Index Models via Alignment Sharpness \nAbstract: We study the problem of learning Single-Index Models under the L_2^2 loss in the agnostic model. We give an efficient learning algorithm\, achieving a constant factor approximation to the optimal loss\, that succeeds under a range of distributions (including log-concave distributions) and a broad class of monotone and Lipschitz link functions. This is the first efficient constant factor approximate agnostic learner\, even for Gaussian data and for any nontrivial class of link functions. Prior work for the case of unknown link function either works in the realizable setting or does not attain constant factor approximation. The main technical ingredient enabling our algorithm and analysis is a novel notion of a local error bound in optimization that we term alignment sharpness and that may be of broader interest. \n  \nSpeaker 2: Lisheng Ren \nSQ Lower Bounds for Non-Gaussian Component Analysis with Weaker Assumptions \nAbstract: The talk focuses on the complexity of Non-Gaussian Component Analysis (NGCA) in the Statistical Query (SQ) model. The NGCA has been a useful problem framework for obtaining SQ hardness for a variety of statistical problems. In particular\, it was previously known that for any univariate distribution $A$ satisfying certain conditions\, distinguishing between a standard multivariate Gaussian and a distribution that behaves like $A$ in a random hidden direction and like a standard Gaussian in the orthogonal complement\, is SQ-hard. This required 1) $A$ matches many low-order moments with a standard Gaussian\, and (2) the chi-squared norm of $A$ with respect to the standard Gaussian is finite\, where the chi-squared restriction is needed for technical reasons. In this talk\, we will present the new result that shows the Condition (2) above is indeed not necessary.
URL:https://ifds.info/event/ifds-ideas-forum-double-header/
LOCATION:Orchard View Room\, 330 N. Orchard Street\, 3rd Floor NE\, Madison\, Wisconsin\, 53715\, United States
CATEGORIES:IFDS Ideas Forum
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20240306T123000
DTEND;TZID=America/Chicago:20240306T123000
DTSTAMP:20260407T033447
CREATED:20240315T164215Z
LAST-MODIFIED:20240315T171618Z
UID:2833-1709728200-1709728200@ifds.info
SUMMARY:SILO: Online Learning Guided Quasi-Newton Methods: Improved Global Non-asymptotic Guarantees
DESCRIPTION:Aryan Mokhtari\, UT
URL:https://ifds.info/event/silo-online-learning-guided-quasi-newton-methods-improved-global-non-asymptotic-guarantees/
LOCATION:Orchard View Room\, 330 N. Orchard Street\, 3rd Floor NE\, Madison\, Wisconsin\, 53715\, United States
CATEGORIES:SILO
ATTACH;FMTTYPE=image/png:https://ifds.info/wp-content/uploads/2022/10/SILO-1024x683-1-e1665597390709.png
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20240308T133000
DTEND;TZID=America/Chicago:20240308T143000
DTSTAMP:20260407T033447
CREATED:20240318T213102Z
LAST-MODIFIED:20240318T213102Z
UID:2890-1709904600-1709908200@ifds.info
SUMMARY:Low-Rank Structures in Optimal Transport
DESCRIPTION:Bio: Meyer Scetbon is currently a Research Scientist at Microsoft Research. He completed his PhD at Institut Polytechnique de Paris\, advised by M. Cuturi. He did\, as a visiting student\, his MS theses UW and Technion\, on kernel-based viewpoints on deep neural networks advised by Z. Harchaoui\, and end-to-end signal and image denoising advised by M. Elad\, respectively. \n\n\n\n\n\nAbstract: Optimal transport (OT) plays an increasingly important role in machine learning (ML) to compare probability distributions. Yet\, it poses\, in its original form\, several challenges when used for applied problems: (i) computing OT between discrete distributions amounts to solving a large and expensive network flow problem which requires a supercubic complexity in the number of points; (ii) estimating OT using sampled measures is doomed by the curse of dimensionality. These issues can be mitigated using an entropic regularization\, solved with the Sinkhorn algorithm\, which improves on both statistical and computational aspects. While much faster\, entropic OT still requires a quadratic complexity with respect to the number of points and therefore remains prohibitive for large-scale problems. In this talk\, I will present new regularization approaches for the OT problem\, as well as its quadratic extension\, the Gromov-Wasserstein (GW) problem\, which impose low-rank structures on the admissible couplings. This results in the development of new algorithms that enjoy a linear complexity both in time and memory with respect to the number of points\, enabling their applications in the large-scale setting where millions of points need to be compared. Additionally\, I will show that these new regularization schemes have better statistical performances compared to the entropic approach\, that they naturally interpolate between the Maximum Mean Discrepancy (MMD) and OT\, and that they offer general clustering methods for arbitrary geometry.Website: <https://meyerscetbon.github.io/_pages/publications/>
URL:https://ifds.info/event/low-rank-structures-in-optimal-transport/
LOCATION:CSE (Allen) 403
CATEGORIES:MLOpt@UWash
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20240311T123000
DTEND;TZID=America/Chicago:20240311T133000
DTSTAMP:20260407T033447
CREATED:20240318T214257Z
LAST-MODIFIED:20240318T214257Z
UID:2904-1710160200-1710163800@ifds.info
SUMMARY:Towards Plurality: Learning from Diverse Human Preferences
DESCRIPTION:Speaker: Ramya Vinayak \n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nAbstract: \nLarge pre-trained models trained on internet-scale data are often not ready for safe deployment out of the box. They are heavily fine-tuned and aligned using large quantities of human preference data. When we want to align an AI/ML model to human preference or values\,  it is worthwhile to ask whose preference and values we are aligning it to? Recently\, the limitations of current approaches due to their inherent uniformity assumption have been highlighted and the need for plurality – capturing the diversity in human preferences and values – is getting recognized as an important challenge to address. While alignment from human preferences has currently become a very active area of research\, it is worthwhile to note that there is rich literature on learning preferences from human judgements using comparison queries. It plays a crucial role in several applications ranging from cognitive and behavioral psychology\, crowdsourcing democracy\, surveys in social science applications\, and recommendation systems. However\, the models in the literature often focus on learning average preference over the population due to the limitations on the amount of data available per individual and also assume the knowledge of the metric or way humans judge similarity and dissimilarity. \nIn this talk\, I will discuss some recent results that focus on how we can reliably capture diversity in preferences while pooling together data from individuals. In particular\, I will talk about fundamental questions in two directions: (1) Simultaneous metric and preference learning where the goal is to learn an unknown shared metric from preference queries while the preferences are diverse and also unknown. (2) Learning distribution of preferences over a population with a single comparison query per individual. \nBio: \nRamya Korlakai Vinayak is an assistant professor in the Dept. of ECE and affiliated faculty in the Dept. of Computer Science and the Dept. of Statistics at the UW-Madison. Her research interests span the areas of machine learning\, statistical inference\, and crowdsourcing. Her work focuses on addressing theoretical and practical challenges that arise when learning from heterogeneous data from people. Prior to joining UW-Madison\, Ramya was a postdoctoral researcher in the Paul G. Allen School of Computer Science and Engineering at the University of Washington. She received her Ph.D. in Electrical Engineering from Caltech. She obtained her Masters from Caltech and Bachelors from IIT Madras. She is a recipient of the Schlumberger Foundation Faculty of the Future fellowship from 2013-15\, and an invited participant at the Rising Stars in EECS workshop in 2019. She is the recipient of NSF CAREER Award 2023-2028.
URL:https://ifds.info/event/towards-plurality-learning-from-diverse-human-preferences/
LOCATION:Orchard View Room\, 330 N. Orchard Street\, 3rd Floor NE\, Madison\, Wisconsin\, 53715\, United States
CATEGORIES:IFDS Ideas Forum
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20240313T123000
DTEND;TZID=America/Chicago:20240313T123000
DTSTAMP:20260407T033447
CREATED:20240315T164215Z
LAST-MODIFIED:20240315T172045Z
UID:2834-1710333000-1710333000@ifds.info
SUMMARY:SILO: Unsupervised Learning: Validation beyond Visualization
DESCRIPTION:Marina Meila\, UWash
URL:https://ifds.info/event/silo-unsupervised-learning-validation-beyond-visualization/
LOCATION:Researchers’ Link\, 330 N Orchard St\, Madison\, WI\, 53715\, United States
CATEGORIES:SILO
ATTACH;FMTTYPE=image/png:https://ifds.info/wp-content/uploads/2022/10/SILO-1024x683-1-e1665597390709.png
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Los_Angeles:20240315T133000
DTEND;TZID=America/Los_Angeles:20240315T143000
DTSTAMP:20260407T033447
CREATED:20240318T212943Z
LAST-MODIFIED:20240318T212943Z
UID:2888-1710509400-1710513000@ifds.info
SUMMARY:Optimized Decision Making via Active Learning of Stochastic Hamiltonians
DESCRIPTION:Speaker: Prof. Chandrajit Bajaj \, UT Austin \nAbstract: A Hamiltonian represents the energy of a dynamical system in phase space with coordinates of position and momentum. The Hamilton’s equations of motion are obtainable as coupled symplectic differential equations.  In this talk I shall show how optimized decision making (action sequences) can be obtained via a reinforcement learning problem wherein the agent interacts with the unknown environment to simultaneously learn a Hamiltonian surrogate and the optimal action sequences using Hamilton dynamics\, by invoking the Pontryagin Maximum Principle. We use optimal control theory to define an optimal control gradient flow\, which guides the reinforcement learning process of the agent to progressively optimize the Hamiltonian while simultaneously converging to the optimal action sequence. Extensions to stochastic Hamiltonians leading to stochastic action sequences and the free-energy principle shall also be discussed. This is joint work with  Taemin Heo\, Minh Nguyen.
URL:https://ifds.info/event/optimized-decision-making-via-active-learning-of-stochastic-hamiltonians/
LOCATION:CSE (Allen) 403
CATEGORIES:MLOpt@UWash
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20240318T123000
DTEND;TZID=America/Chicago:20240318T133000
DTSTAMP:20260407T033447
CREATED:20240318T214430Z
LAST-MODIFIED:20240318T214430Z
UID:2906-1710765000-1710768600@ifds.info
SUMMARY:IFDS Ideas Forum Double Header
DESCRIPTION:Karan Srivastava\n\n\n\n\nHarit Vishwakarma: “Confidence functions for auto-labeling”
URL:https://ifds.info/event/ifds-ideas-forum-double-header-2/
LOCATION:WID 1145\, 330 N Orchard Street\, Madison\, WI\, 53715\, United States
CATEGORIES:IFDS Ideas Forum
END:VEVENT
END:VCALENDAR