BEGIN:VCALENDAR
VERSION:2.0
PRODID:-//IFDS - ECPv6.0.1.1//NONSGML v1.0//EN
CALSCALE:GREGORIAN
METHOD:PUBLISH
X-WR-CALNAME:IFDS
X-ORIGINAL-URL:https://ifds.info
X-WR-CALDESC:Events for IFDS
REFRESH-INTERVAL;VALUE=DURATION:PT1H
X-Robots-Tag:noindex
X-PUBLISHED-TTL:PT1H
BEGIN:VTIMEZONE
TZID:America/Chicago
BEGIN:DAYLIGHT
TZOFFSETFROM:-0600
TZOFFSETTO:-0500
TZNAME:CDT
DTSTART:20230312T080000
END:DAYLIGHT
BEGIN:STANDARD
TZOFFSETFROM:-0500
TZOFFSETTO:-0600
TZNAME:CST
DTSTART:20231105T070000
END:STANDARD
BEGIN:DAYLIGHT
TZOFFSETFROM:-0600
TZOFFSETTO:-0500
TZNAME:CDT
DTSTART:20240310T080000
END:DAYLIGHT
BEGIN:STANDARD
TZOFFSETFROM:-0500
TZOFFSETTO:-0600
TZNAME:CST
DTSTART:20241103T070000
END:STANDARD
TZID:America/Los_Angeles
BEGIN:DAYLIGHT
TZOFFSETFROM:-0800
TZOFFSETTO:-0700
TZNAME:PDT
DTSTART:20240310T100000
END:DAYLIGHT
BEGIN:STANDARD
TZOFFSETFROM:-0700
TZOFFSETTO:-0800
TZNAME:PST
DTSTART:20241103T090000
END:STANDARD
END:VTIMEZONE
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20240306T123000
DTEND;TZID=America/Chicago:20240306T123000
DTSTAMP:20260409T102728
CREATED:20240315T164215Z
LAST-MODIFIED:20240315T171618Z
UID:2833-1709728200-1709728200@ifds.info
SUMMARY:SILO: Online Learning Guided Quasi-Newton Methods: Improved Global Non-asymptotic Guarantees
DESCRIPTION:Aryan Mokhtari\, UT
URL:https://ifds.info/event/silo-online-learning-guided-quasi-newton-methods-improved-global-non-asymptotic-guarantees/
LOCATION:Orchard View Room\, 330 N. Orchard Street\, 3rd Floor NE\, Madison\, Wisconsin\, 53715\, United States
CATEGORIES:SILO
ATTACH;FMTTYPE=image/png:https://ifds.info/wp-content/uploads/2022/10/SILO-1024x683-1-e1665597390709.png
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20240304T123000
DTEND;TZID=America/Chicago:20240304T133000
DTSTAMP:20260409T102728
CREATED:20240318T214117Z
LAST-MODIFIED:20240318T214117Z
UID:2902-1709555400-1709559000@ifds.info
SUMMARY:IFDS Ideas Forum Double Header
DESCRIPTION:Speaker 1: Puqian Wang \nRobustly Learning Single-Index Models via Alignment Sharpness \nAbstract: We study the problem of learning Single-Index Models under the L_2^2 loss in the agnostic model. We give an efficient learning algorithm\, achieving a constant factor approximation to the optimal loss\, that succeeds under a range of distributions (including log-concave distributions) and a broad class of monotone and Lipschitz link functions. This is the first efficient constant factor approximate agnostic learner\, even for Gaussian data and for any nontrivial class of link functions. Prior work for the case of unknown link function either works in the realizable setting or does not attain constant factor approximation. The main technical ingredient enabling our algorithm and analysis is a novel notion of a local error bound in optimization that we term alignment sharpness and that may be of broader interest. \n  \nSpeaker 2: Lisheng Ren \nSQ Lower Bounds for Non-Gaussian Component Analysis with Weaker Assumptions \nAbstract: The talk focuses on the complexity of Non-Gaussian Component Analysis (NGCA) in the Statistical Query (SQ) model. The NGCA has been a useful problem framework for obtaining SQ hardness for a variety of statistical problems. In particular\, it was previously known that for any univariate distribution $A$ satisfying certain conditions\, distinguishing between a standard multivariate Gaussian and a distribution that behaves like $A$ in a random hidden direction and like a standard Gaussian in the orthogonal complement\, is SQ-hard. This required 1) $A$ matches many low-order moments with a standard Gaussian\, and (2) the chi-squared norm of $A$ with respect to the standard Gaussian is finite\, where the chi-squared restriction is needed for technical reasons. In this talk\, we will present the new result that shows the Condition (2) above is indeed not necessary.
URL:https://ifds.info/event/ifds-ideas-forum-double-header/
LOCATION:Orchard View Room\, 330 N. Orchard Street\, 3rd Floor NE\, Madison\, Wisconsin\, 53715\, United States
CATEGORIES:IFDS Ideas Forum
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20240228T123000
DTEND;TZID=America/Chicago:20240228T133000
DTSTAMP:20260409T102728
CREATED:20240315T164015Z
LAST-MODIFIED:20240315T173213Z
UID:2832-1709123400-1709127000@ifds.info
SUMMARY:SILO: Reinforcement Learning with Robustness and Safety Guarantees
DESCRIPTION:Dileep Kalathil\, TAMU
URL:https://ifds.info/event/silo-reinforcement-learning-with-robustness-and-safety-guarantees/
LOCATION:Orchard View Room\, 330 N. Orchard Street\, 3rd Floor NE\, Madison\, Wisconsin\, 53715\, United States
CATEGORIES:SILO
ATTACH;FMTTYPE=image/png:https://ifds.info/wp-content/uploads/2022/10/SILO-1024x683-1-e1665597390709.png
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20240226T123000
DTEND;TZID=America/Chicago:20240226T133000
DTSTAMP:20260409T102728
CREATED:20240318T213825Z
LAST-MODIFIED:20240318T213825Z
UID:2900-1708950600-1708954200@ifds.info
SUMMARY:Prelimit coupling and steady-state convergence of constant-stepsize nonsmooth contractive SA
DESCRIPTION:Speaker: Yixuan Zhang \n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\nAbstract:  \nMotivated by Q-learning\, we study nonsmooth contractive stochastic approximation (SA) with constant stepsize. We focus on two important classes of dynamics: 1) nonsmooth contractive SA with additive noise\, and 2) synchronous and asynchronous Q-learning\, which features both additive and multiplicative noise. For both dynamics\, we establish weak convergence of the iterates to a stationary limit distribution in Wasserstein distance. Furthermore\, we propose a prelimit coupling technique for establishing steady-state convergence and characterize the limit of the stationary distribution as the stepsize goes to zero. Using this result\, we derive that the asymptotic bias of nonsmooth SA is proportional to the square root of the stepsize\, which stands in sharp contrast to smooth SA. This bias characterization allows for the use of Richardson-Romberg extrapolation for bias reduction in nonsmooth SA.
URL:https://ifds.info/event/prelimit-coupling-and-steady-state-convergence-of-constant-stepsize-nonsmooth-contractive-sa/
LOCATION:WID 1145\, 330 N Orchard Street\, Madison\, WI\, 53715\, United States
CATEGORIES:IFDS Ideas Forum
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Los_Angeles:20240223T133000
DTEND;TZID=America/Los_Angeles:20240223T143000
DTSTAMP:20260409T102728
CREATED:20240318T212752Z
LAST-MODIFIED:20240318T212752Z
UID:2886-1708695000-1708698600@ifds.info
SUMMARY:GumbelSpec Sampling for Accelerating LLM Inference
DESCRIPTION:Bio: Tianxiao Shen is a postdoctoral scholar at the University of Washington\, working with Yejin Choi and Zaid Harchaoui. Her research interests lie in natural language processing and machine learning\, in particular developing models and algorithms for efficient\, accurate\, diverse\, flexible and controllable text generation. She received her PhD from MIT\, advised by Regina Barzilay and Tommi Jaakkola. Before that\, she did her undergrad at Tsinghua University. \n\n\n\n\nAbstract: We propose GumbelSpec sampling\, a novel algorithm that leverages smaller language models to accelerate inference of large language models without changing their output distribution. Central to our approach is the application of the Gumbel-Softmax technique to convert the stochastic decoding process into a deterministic process by integrating independently sampled Gumbel noise. Employing the same set of Gumbel noise\, we perform beam search on the smaller model to generate multiple candidate short continuations\, and then utilize tree-based attention to efficiently verify them in parallel using the larger model. GumbelSpec sampling significantly improves upon previous rejection sampling based speculative decoding methods by increasing the token acceptance rate by 1.7x-2.2x and achieving an additional speedup of 1.2x-1.5x. This results in a total speedup of 1.5x-2.6x compared to traditional autoregressive decoding.
URL:https://ifds.info/event/gumbelspec-sampling-for-accelerating-llm-inference/
LOCATION:CSE (Allen) 403
CATEGORIES:MLOpt@UWash
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20240221T123000
DTEND;TZID=America/Chicago:20240221T133000
DTSTAMP:20260409T102728
CREATED:20240315T164015Z
LAST-MODIFIED:20240315T173126Z
UID:2831-1708518600-1708522200@ifds.info
SUMMARY:SILO: Foundations of Real-World Reinforcement Learning
DESCRIPTION:Jeongyeol Kwon
URL:https://ifds.info/event/silo-foundations-of-real-world-reinforcement-learning/
LOCATION:Orchard View Room\, 330 N. Orchard Street\, 3rd Floor NE\, Madison\, Wisconsin\, 53715\, United States
CATEGORIES:SILO
ATTACH;FMTTYPE=image/png:https://ifds.info/wp-content/uploads/2022/10/SILO-1024x683-1-e1665597390709.png
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20240219T123000
DTEND;TZID=America/Chicago:20240219T133000
DTSTAMP:20260409T102728
CREATED:20240318T213630Z
LAST-MODIFIED:20240318T213656Z
UID:2897-1708345800-1708349400@ifds.info
SUMMARY:A good score does not lead to a good generative model
DESCRIPTION:Speaker: Sixu Li \nAbstract: Score-based Generative Models (SGMs) is one leading method in generative modeling\, renowned for their ability to generate high-quality samples from complex\, high-dimensional data distributions. The method enjoys empirical success and is supported by rigorous theoretical convergence properties. In particular\, it has been shown that SGMs can generate samples from a distribution that is close to the ground-truth if the underlying score function is learned well\, suggesting the success of SGM as a generative model. We provide a counter-example in this paper. Through the sample complexity argument\, we provide one specific setting where the score function is learned well. Yet\, SGMs in this setting can only output samples that are Gaussian blurring of training data points\, mimicking the effects of kernel density estimation. The finding resonates a series of recent finding that reveal that SGMs can demonstrate strong memorization effect and fail to generate. This is joint with Shi Chen and Qin Li. 
URL:https://ifds.info/event/a-good-score-does-not-lead-to-a-good-generative-model/
LOCATION:WID 1145\, 330 N Orchard Street\, Madison\, WI\, 53715\, United States
CATEGORIES:IFDS Ideas Forum
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Los_Angeles:20240216T133000
DTEND;TZID=America/Los_Angeles:20240216T143000
DTSTAMP:20260409T102728
CREATED:20240318T212625Z
LAST-MODIFIED:20240318T212625Z
UID:2884-1708090200-1708093800@ifds.info
SUMMARY:Offline Multi-task Transfer RL with Representational Penalization
DESCRIPTION:Speaker Bio: Avinandan is a second year PhD student\, advised by Maryam Fazel and Lillian Ratliff. His interests are in sequential learning and game theory. \n\n\nAbstract: We study the problem of representational transfer in offline Reinforcement Learning (RL)\, where a learner has access to episodic data from a number of source tasks collected a priori\, and aims to learn a shared representation to be used in finding a good policy for a target task. Unlike in online RL where the agent interacts with the environment while learning a policy\, in the offline setting there cannot be such interactions in either the source tasks or the target task; thus multi-task offline RL can suffer from incomplete coverage.We propose an algorithm to compute pointwise uncertainty measures for the learnt representation\, and establish a data-dependent upper bound for the suboptimality of the learnt policy for the target task. Our algorithm leverages the collective exploration done by source tasks to mitigate poor coverage at some points by a few tasks\, thus overcoming the limitation of needing uniformly good coverage for a meaningful transfer by existing offline algorithms. We complement our theoretical results with empirical evaluation on a rich-observation MDP which requires many samples for complete coverage. Our findings illustrate the benefits of penalizing and quantifying the uncertainty in the learnt representation.
URL:https://ifds.info/event/offline-multi-task-transfer-rl-with-representational-penalization/
LOCATION:CSE (Allen) 403
CATEGORIES:MLOpt@UWash
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20240214T123000
DTEND;TZID=America/Chicago:20240214T133000
DTSTAMP:20260409T102728
CREATED:20240315T164015Z
LAST-MODIFIED:20240315T173037Z
UID:2830-1707913800-1707917400@ifds.info
SUMMARY:SILO: Theoretical Exploration of Foundation Model Adaptation Methods
DESCRIPTION:Kangwook Lee
URL:https://ifds.info/event/silo-theoretical-exploration-of-foundation-model-adaptation-methods/
LOCATION:Orchard View Room\, 330 N. Orchard Street\, 3rd Floor NE\, Madison\, Wisconsin\, 53715\, United States
CATEGORIES:SILO
ATTACH;FMTTYPE=image/png:https://ifds.info/wp-content/uploads/2022/10/SILO-1024x683-1-e1665597390709.png
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20240212T123000
DTEND;TZID=America/Chicago:20240212T133000
DTSTAMP:20260409T102728
CREATED:20240318T213417Z
LAST-MODIFIED:20240318T213417Z
UID:2894-1707741000-1707744600@ifds.info
SUMMARY:Theoretical exploration of foundation model adaption methods
DESCRIPTION:Speaker: Kangwook Lee
URL:https://ifds.info/event/theoretical-exploration-of-foundation-model-adaption-methods/
LOCATION:Orchard View Room\, 330 N. Orchard Street\, 3rd Floor NE\, Madison\, Wisconsin\, 53715\, United States
CATEGORIES:IFDS Ideas Forum
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Los_Angeles:20240209T133000
DTEND;TZID=America/Los_Angeles:20240209T143000
DTSTAMP:20260409T102728
CREATED:20240318T212503Z
LAST-MODIFIED:20240318T212503Z
UID:2882-1707485400-1707489000@ifds.info
SUMMARY:Policy Optimization with Compatible Mirror Approximation
DESCRIPTION:Speaker Bio: Zhihan is a fourth-year PhD student in the Paul G. Allen School of Computer Science & Engineering at University of Washington\, advised by Prof. Maryam Fazel. His research interests are broadly in statistics\, optimization and machine learning. \n\n\nAbstract: We propose Compatible Mirror Policy Optimization (CoMPO)\, a framework that incorporates general function approximation into policy mirror descent methods. In contrast to the popular approach of using the $L_2$ norm to measure function approximation errors (regardless of the mirror map)\, CoMPO uses the Bregman divergence induced by the specific mirror map for policy projection. Such a compatibility bridges the gap between theory and practice: not only does it achieve fast linear convergence with general function approximation\, but it also includes several well-known practical methods as special cases\, immediately providing them strong convergence guarantees.
URL:https://ifds.info/event/policy-optimization-with-compatible-mirror-approximation/
LOCATION:Zoom
CATEGORIES:MLOpt@UWash
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20240207T123000
DTEND;TZID=America/Chicago:20240207T133000
DTSTAMP:20260409T102728
CREATED:20240315T164015Z
LAST-MODIFIED:20240315T172944Z
UID:2829-1707309000-1707312600@ifds.info
SUMMARY:SILO: Universality in High-Dimensional Statistics
DESCRIPTION:Rishabh Dudeja
URL:https://ifds.info/event/silo-universality-in-high-dimensional-statistics/
LOCATION:Orchard View Room\, 330 N. Orchard Street\, 3rd Floor NE\, Madison\, Wisconsin\, 53715\, United States
CATEGORIES:SILO
ATTACH;FMTTYPE=image/png:https://ifds.info/wp-content/uploads/2022/10/SILO-1024x683-1-e1665597390709.png
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20240205T123000
DTEND;TZID=America/Chicago:20240205T133000
DTSTAMP:20260409T102728
CREATED:20240318T213252Z
LAST-MODIFIED:20240318T213252Z
UID:2892-1707136200-1707139800@ifds.info
SUMMARY:Towards a new toolbox of optimal statistical primitives
DESCRIPTION:Abstract: Given society’s increasing reliance on data\, its collection and processing into useful information is a technical problem of growing focus\, and perhaps paradoxically\, a critical bottleneck in many data science and machine learning applications. My research focuses on designing algorithms that push the limits of both statistical efficiency and computational efficiency. In particular\, my work tackles the divide between the theory and practice of data science\, which exists even for the most basic statistical problems including mean and (co)variance estimation. Conventional methods such as the sample mean\, while supported by theoretical results under strong assumptions\, are often brittle in the presence of extreme data points. To counter such deficiencies\, practitioners often use ad-hoc and unprincipled “outlier removal” heuristics\, revealing a marked gap between the theory and practice even for these fundamental problems. \nIn this talk\, I will describe my work towards building a new toolbox of optimal statistical primitives\, bridging the theory-practice divide. I will specifically highlight 3 works: A) constructing a statistically-optimal and computationally-efficient 1-dimensional mean estimator\, whose estimation error is optimal even in the leading multiplicative constant\, under bare minimum distributional assumptions\, B) a rather different but also optimal mean estimator for the “very high-dimensional” regime\, and C) a recent result on robustly clustering Gaussian mixtures based on their covariances even in the presence of adversarial data corruption. To conclude the talk\, I will discuss my vision for the new theory and toolbox\, serving as a blueprint for my long-term future research. \nBio: Jasper Lee is a postdoctoral research associate at the University of Wisconsin-Madison\, mentored by Ilias Diakonikolas in the Department of Computer Sciences\, and also affiliated with the Institute for Foundations of Data Science. He completed his PhD at Brown University\, advised by Paul Valiant. \nHis research interests are broadly in the foundations of data science\, aiming to design practical\, data-efficient and computationally-efficient algorithms for a variety of statistical applications. \nHis work is partially supported by a Croucher Fellowship for Postdoctoral Research.
URL:https://ifds.info/event/towards-a-new-toolbox-of-optimal-statistical-primitives/
LOCATION:Orchard View Room\, 330 N. Orchard Street\, 3rd Floor NE\, Madison\, Wisconsin\, 53715\, United States
CATEGORIES:IFDS Ideas Forum
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20240131T123000
DTEND;TZID=America/Chicago:20240131T133000
DTSTAMP:20260409T102728
CREATED:20240315T164015Z
LAST-MODIFIED:20240315T172723Z
UID:2828-1706704200-1706707800@ifds.info
SUMMARY:SILO: A scalable method to exploit screening in Gaussian process models with noise
DESCRIPTION:Chris Geoga
URL:https://ifds.info/event/silo-a-scalable-method-to-exploit-screening-in-gaussian-process-models-with-noise/
LOCATION:Orchard View Room\, 330 N. Orchard Street\, 3rd Floor NE\, Madison\, Wisconsin\, 53715\, United States
CATEGORIES:SILO
ATTACH;FMTTYPE=image/png:https://ifds.info/wp-content/uploads/2022/10/SILO-1024x683-1-e1665597390709.png
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Los_Angeles:20240126T133000
DTEND;TZID=America/Los_Angeles:20240126T143000
DTSTAMP:20260409T102728
CREATED:20240318T212134Z
LAST-MODIFIED:20240318T212230Z
UID:2879-1706275800-1706279400@ifds.info
SUMMARY:How do neural networks learn features from data?
DESCRIPTION:Speaker Bio: Adit is currently the George F. Carrier Postdoctoral Fellow in the School of Engineering and Applied Sciences at Harvard. He completed his Ph.D. in electrical engineering and computer science (EECS) at MIT advised by Caroline Uhler and was a Ph.D. fellow at the Eric and Wendy Schmidt Center at the Broad Institute of MIT and Harvard. His research focuses on advancing theoretical foundations of machine learning and developing new methods for tackling biomedical problems. \n\n\n\nAbstract: Understanding how neural networks learn features\, or relevant patterns in data\, for prediction is necessary for their reliable use in technological and scientific applications. We propose a unifying mechanism that characterizes feature learning in neural network architectures. Namely\, we show that features learned by neural networks are captured by a statistical operator known as the average gradient outer product (AGOP). Empirically\, we show that the AGOP captures features across a broad class of network architectures including convolutional networks and large language models. Moreover\, we use AGOP to enable feature learning in general machine learning models through an algorithm we call Recursive Feature Machine (RFM). We show that RFM automatically identifies sparse subsets of features relevant for prediction and explicitly connects feature learning in neural networks with classical sparse recovery and low rank matrix factorization algorithms. Overall\, this line of work advances our fundamental understanding of how neural networks extract features from data\, leading to the development of novel\, interpretable\, and effective models for use in scientific applications.
URL:https://ifds.info/event/how-do-neural-networks-learn-features-from-data/
LOCATION:CSE (Allen) 403
CATEGORIES:MLOpt@UWash
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20240124T123000
DTEND;TZID=America/Chicago:20240124T133000
DTSTAMP:20260409T102728
CREATED:20240315T164000Z
LAST-MODIFIED:20240315T173629Z
UID:2827-1706099400-1706103000@ifds.info
SUMMARY:SILO: Funsearch: a novel approach to using large language models as part of mathematical practice
DESCRIPTION:Jordan Ellenberg
URL:https://ifds.info/event/silo-funsearch-a-novel-approach-to-using-large-language-models-as-part-of-mathematical-practice/
LOCATION:Orchard View Room\, 330 N. Orchard Street\, 3rd Floor NE\, Madison\, Wisconsin\, 53715\, United States
CATEGORIES:SILO
ATTACH;FMTTYPE=image/png:https://ifds.info/wp-content/uploads/2022/10/SILO-1024x683-1-e1665597390709.png
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20231218T123000
DTEND;TZID=America/Chicago:20231218T123000
DTSTAMP:20260409T102728
CREATED:20231024T213157Z
LAST-MODIFIED:20231024T215501Z
UID:2708-1702902600-1702902600@ifds.info
SUMMARY:IFDS Ideas Forum
DESCRIPTION:TBD\nSpeaker: Sushrut Karmalkar
URL:https://ifds.info/event/tbd-3/
LOCATION:Orchard View Room\, 330 N. Orchard Street\, 3rd Floor NE\, Madison\, Wisconsin\, 53715\, United States
CATEGORIES:IFDS Ideas Forum
ATTACH;FMTTYPE=image/jpeg:https://ifds.info/wp-content/uploads/2023/10/iStock-1415961311.jpg
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20231211T123000
DTEND;TZID=America/Chicago:20231211T123000
DTSTAMP:20260409T102728
CREATED:20231024T213143Z
LAST-MODIFIED:20231024T215415Z
UID:2707-1702297800-1702297800@ifds.info
SUMMARY:IFDS Ideas Forum
DESCRIPTION:TBD\nSpeaker: Thanasis Pittas
URL:https://ifds.info/event/tbd-2/
LOCATION:Orchard View Room\, 330 N. Orchard Street\, 3rd Floor NE\, Madison\, Wisconsin\, 53715\, United States
CATEGORIES:IFDS Ideas Forum
ATTACH;FMTTYPE=image/jpeg:https://ifds.info/wp-content/uploads/2023/10/iStock-1415961311.jpg
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20231206T123000
DTEND;TZID=America/Chicago:20231206T133000
DTSTAMP:20260409T102728
CREATED:20240315T164000Z
LAST-MODIFIED:20240315T173546Z
UID:2826-1701865800-1701869400@ifds.info
SUMMARY:SILO: Invariant Low-Dimensional Subspaces in Gradient Descent for Learning Deep Networks
DESCRIPTION:Qing Qu\, UMich
URL:https://ifds.info/event/silo-invariant-low-dimensional-subspaces-in-gradient-descent-for-learning-deep-networks/
LOCATION:Orchard View Room\, 330 N. Orchard Street\, 3rd Floor NE\, Madison\, Wisconsin\, 53715\, United States
CATEGORIES:SILO
ATTACH;FMTTYPE=image/png:https://ifds.info/wp-content/uploads/2022/10/SILO-1024x683-1-e1665597390709.png
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20231204T123000
DTEND;TZID=America/Chicago:20231204T123000
DTSTAMP:20260409T102728
CREATED:20231024T213142Z
LAST-MODIFIED:20231024T215321Z
UID:2706-1701693000-1701693000@ifds.info
SUMMARY:IFDS Ideas Forum
DESCRIPTION:TBD\nSpeakers: Jitian Zhao / Zhexuan Liu
URL:https://ifds.info/event/short-talk-exact-title-tbd-4/
LOCATION:Orchard View Room\, 330 N. Orchard Street\, 3rd Floor NE\, Madison\, Wisconsin\, 53715\, United States
CATEGORIES:IFDS Ideas Forum
ATTACH;FMTTYPE=image/jpeg:https://ifds.info/wp-content/uploads/2023/10/iStock-1415961311.jpg
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20231129T123000
DTEND;TZID=America/Chicago:20231129T133000
DTSTAMP:20260409T102728
CREATED:20240315T164000Z
LAST-MODIFIED:20240315T173459Z
UID:2825-1701261000-1701264600@ifds.info
SUMMARY:SILO: Understanding deep nets: on local Lipschitz functions and learned proximal networks
DESCRIPTION:Jeremias Sulam\, JHU
URL:https://ifds.info/event/silo-understanding-deep-nets-on-local-lipschitz-functions-and-learned-proximal-networks/
LOCATION:Orchard View Room\, 330 N. Orchard Street\, 3rd Floor NE\, Madison\, Wisconsin\, 53715\, United States
CATEGORIES:SILO
ATTACH;FMTTYPE=image/png:https://ifds.info/wp-content/uploads/2022/10/SILO-1024x683-1-e1665597390709.png
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20231127T123000
DTEND;TZID=America/Chicago:20231127T123000
DTSTAMP:20260409T102728
CREATED:20231024T212227Z
LAST-MODIFIED:20231024T215121Z
UID:2686-1701088200-1701088200@ifds.info
SUMMARY:IFDS Ideas Forum
DESCRIPTION:TBD\nSpeaker: Zarifis Nikos
URL:https://ifds.info/event/ifds-ideas-forum-47/
CATEGORIES:IFDS Ideas Forum
ATTACH;FMTTYPE=image/jpeg:https://ifds.info/wp-content/uploads/2023/10/iStock-1415961311.jpg
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20231122T123000
DTEND;TZID=America/Chicago:20231122T133000
DTSTAMP:20260409T102728
CREATED:20240315T164000Z
LAST-MODIFIED:20240315T173404Z
UID:2824-1700656200-1700659800@ifds.info
SUMMARY:SILO: World Knowledge in the Time of Large Models
DESCRIPTION:DeepMind\, Kenneth Marino
URL:https://ifds.info/event/silo-world-knowledge-in-the-time-of-large-models/
LOCATION:Orchard View Room\, 330 N. Orchard Street\, 3rd Floor NE\, Madison\, Wisconsin\, 53715\, United States
CATEGORIES:SILO
ATTACH;FMTTYPE=image/png:https://ifds.info/wp-content/uploads/2022/10/SILO-1024x683-1-e1665597390709.png
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20231120T123000
DTEND;TZID=America/Chicago:20231120T123000
DTSTAMP:20260409T102728
CREATED:20231024T213142Z
LAST-MODIFIED:20231024T215229Z
UID:2705-1700483400-1700483400@ifds.info
SUMMARY:IFDS Ideas Forum
DESCRIPTION:TBD\nSpeakers: Karan Srivastava / Chenghui Li
URL:https://ifds.info/event/short-talk-exact-title-tbd-3/
LOCATION:Orchard View Room\, 330 N. Orchard Street\, 3rd Floor NE\, Madison\, Wisconsin\, 53715\, United States
CATEGORIES:IFDS Ideas Forum
ATTACH;FMTTYPE=image/jpeg:https://ifds.info/wp-content/uploads/2023/10/iStock-1415961311.jpg
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20231115T123000
DTEND;TZID=America/Chicago:20231115T133000
DTSTAMP:20260409T102728
CREATED:20240315T164000Z
LAST-MODIFIED:20240315T173311Z
UID:2823-1700051400-1700055000@ifds.info
SUMMARY:SILO: A Theory of Universal Learning
DESCRIPTION:Purdue\, Steve Hanneke
URL:https://ifds.info/event/silo-a-theory-of-universal-learning/
LOCATION:Orchard View Room\, 330 N. Orchard Street\, 3rd Floor NE\, Madison\, Wisconsin\, 53715\, United States
CATEGORIES:SILO
ATTACH;FMTTYPE=image/png:https://ifds.info/wp-content/uploads/2022/10/SILO-1024x683-1-e1665597390709.png
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20231113T123000
DTEND;TZID=America/Chicago:20231113T123000
DTSTAMP:20260409T102728
CREATED:20231024T213142Z
LAST-MODIFIED:20231024T215011Z
UID:2704-1699878600-1699878600@ifds.info
SUMMARY:IFDS Ideas Forum
DESCRIPTION:TBD\nSpeakers: William Powell / Andrew Lowy
URL:https://ifds.info/event/short-talk-exact-title-tbd-2/
LOCATION:Orchard View Room\, 330 N. Orchard Street\, 3rd Floor NE\, Madison\, Wisconsin\, 53715\, United States
CATEGORIES:IFDS Ideas Forum
ATTACH;FMTTYPE=image/jpeg:https://ifds.info/wp-content/uploads/2023/10/iStock-1415961311.jpg
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20231108T123000
DTEND;TZID=America/Chicago:20231108T133000
DTSTAMP:20260409T102728
CREATED:20240315T163945Z
LAST-MODIFIED:20240315T173949Z
UID:2822-1699446600-1699450200@ifds.info
SUMMARY:SILO: Leveraging Reviews: Learning to Price with Buyer and Seller Uncertainty
DESCRIPTION:Ellen Vitercik\, Stanford
URL:https://ifds.info/event/silo-leveraging-reviews-learning-to-price-with-buyer-and-seller-uncertainty/
LOCATION:Orchard View Room\, 330 N. Orchard Street\, 3rd Floor NE\, Madison\, Wisconsin\, 53715\, United States
CATEGORIES:SILO
ATTACH;FMTTYPE=image/png:https://ifds.info/wp-content/uploads/2022/10/SILO-1024x683-1-e1665597390709.png
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20231106T123000
DTEND;TZID=America/Chicago:20231106T123000
DTSTAMP:20260409T102728
CREATED:20231024T213128Z
LAST-MODIFIED:20231024T214918Z
UID:2703-1699273800-1699273800@ifds.info
SUMMARY:IFDS Ideas Forum
DESCRIPTION:TBD\nSpeakers: Joe Shenouda/Matthew Zurek
URL:https://ifds.info/event/short-talk-exact-title-tbd/
LOCATION:Orchard View Room\, 330 N. Orchard Street\, 3rd Floor NE\, Madison\, Wisconsin\, 53715\, United States
CATEGORIES:IFDS Ideas Forum
ATTACH;FMTTYPE=image/jpeg:https://ifds.info/wp-content/uploads/2023/10/iStock-1415961311.jpg
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20231101T123000
DTEND;TZID=America/Chicago:20231101T133000
DTSTAMP:20260409T102728
CREATED:20240315T163945Z
LAST-MODIFIED:20240315T173902Z
UID:2821-1698841800-1698845400@ifds.info
SUMMARY:SILO: Balanced Filtering via Disclosure-Controlled Proxies
DESCRIPTION:Emily Diana\, TTIC
URL:https://ifds.info/event/silo-balanced-filtering-via-disclosure-controlled-proxies/
LOCATION:Orchard View Room\, 330 N. Orchard Street\, 3rd Floor NE\, Madison\, Wisconsin\, 53715\, United States
CATEGORIES:SILO
ATTACH;FMTTYPE=image/png:https://ifds.info/wp-content/uploads/2022/10/SILO-1024x683-1-e1665597390709.png
END:VEVENT
BEGIN:VEVENT
DTSTART;TZID=America/Chicago:20231030T123000
DTEND;TZID=America/Chicago:20231030T123000
DTSTAMP:20260409T102728
CREATED:20231024T213128Z
LAST-MODIFIED:20231024T214414Z
UID:2702-1698669000-1698669000@ifds.info
SUMMARY:IFDS Ideas Forum
DESCRIPTION:Human-in-the-Loop Out of Distribution Detection with False Positive Rate Guarantees\nSpeakers: Harit Vishwakarma/Ramya Vinayak
URL:https://ifds.info/event/human-in-the-loop-out-of-distribution-detection-with-false-positive-rate-guarantees/
LOCATION:Orchard View Room\, 330 N. Orchard Street\, 3rd Floor NE\, Madison\, Wisconsin\, 53715\, United States
CATEGORIES:IFDS Ideas Forum
ATTACH;FMTTYPE=image/jpeg:https://ifds.info/wp-content/uploads/2023/10/iStock-1415961311.jpg
END:VEVENT
END:VCALENDAR