Search Results - "Kale, Satyen"
-
1
Logarithmic regret algorithms for online convex optimization
Published in Machine learning (01-12-2007)“…Issue Title: Special Issue on COLT 2006; Guest Editors: Avrim Blum, Gabor Lugosi and Hans Ulrich Simon In an online convex optimization problem a…”
Get full text
Conference Proceeding Journal Article -
2
Extracting certainty from uncertainty: regret bounded by variation in costs
Published in Machine learning (01-09-2010)“…Prediction from expert advice is a fundamental problem in machine learning. A major pillar of the field is the existence of learning algorithms whose average…”
Get full text
Journal Article -
3
AN ONLINE PORTFOLIO SELECTION ALGORITHM WITH REGRET LOGARITHMIC IN PRICE VARIATION
Published in Mathematical finance (01-04-2015)“…We present a novel efficient algorithm for portfolio selection which theoretically attains two desirable properties: Worst‐case guarantee: the algorithm is…”
Get full text
Journal Article -
4
Learning rotations with little regret
Published in Machine learning (01-07-2016)“…We describe online algorithms for learning a rotation from pairs of unit vectors in R n . We show that the expected regret of our online algorithm compared to…”
Get full text
Journal Article -
5
Extracting certainty from uncertainty: regret bounded byvariation incosts
Published in Machine learning (01-09-2010)“…Prediction from expert advice is a fundamental problem in machine learning. A major pillar of the field is the existence of learning algorithms whose average…”
Get full text
Journal Article -
6
Extracting certainty from uncertainty: regret bounded by variation in costs: Special Issue on Learning Theory
Published in Machine learning (2010)Get full text
Journal Article -
7
A variation on SVD based image compression
Published in Image and vision computing (01-06-2007)“…We present a variation to the well studied SVD based image compression technique. Our variation can be viewed as a preprocessing step in which the input image…”
Get full text
Journal Article -
8
Stacking as Accelerated Gradient Descent
Published 07-03-2024“…Stacking, a heuristic technique for training deep residual networks by progressively increasing the number of layers and initializing new layers by copying…”
Get full text
Journal Article -
9
Improved Differentially Private and Lazy Online Convex Optimization
Published 15-12-2023“…We study the task of $(\epsilon, \delta)$-differentially private online convex optimization (OCO). In the online setting, the release of each distinct decision…”
Get full text
Journal Article -
10
Pushing the Efficiency-Regret Pareto Frontier for Online Learning of Portfolios and Quantum States
Published 06-02-2022“…We revisit the classical online portfolio selection problem. It is widely assumed that a trade-off between computational complexity and regret is unavoidable,…”
Get full text
Journal Article -
11
Almost Tight Bounds for Differentially Private Densest Subgraph
Published 20-08-2023“…We study the Densest Subgraph (DSG) problem under the additional constraint of differential privacy. DSG is a fundamental theoretical question which plays a…”
Get full text
Journal Article -
12
Efficient Methods for Online Multiclass Logistic Regression
Published 06-10-2021“…Multiclass logistic regression is a fundamental task in machine learning with applications in classification and boosting. Previous work (Foster et al., 2018)…”
Get full text
Journal Article -
13
SGD: The Role of Implicit Regularization, Batch-size and Multiple-epochs
Published 11-07-2021“…Multi-epoch, small-batch, Stochastic Gradient Descent (SGD) has been the method of choice for learning with large over-parameterized models. A popular theory…”
Get full text
Journal Article -
14
A Multiclass Boosting Framework for Achieving Fast and Provable Adversarial Robustness
Published 01-03-2021“…Alongside the well-publicized accomplishments of deep neural networks there has emerged an apparent bug in their success on tasks such as object recognition:…”
Get full text
Journal Article -
15
Beyond Uniform Lipschitz Condition in Differentially Private Optimization
Published 21-06-2022“…Most prior results on differentially private stochastic gradient descent (DP-SGD) are derived under the simplistic assumption of uniform Lipschitzness, i.e.,…”
Get full text
Journal Article -
16
Efficient Stagewise Pretraining via Progressive Subnetworks
Published 08-02-2024“…Recent developments in large language models have sparked interest in efficient pretraining methods. Stagewise training approaches to improve efficiency, like…”
Get full text
Journal Article -
17
Agnostic Learnability of Halfspaces via Logistic Loss
Published 31-01-2022“…We investigate approximation guarantees provided by logistic regression for the fundamental problem of agnostic learning of homogeneous halfspaces. Previously,…”
Get full text
Journal Article -
18
A Deep Conditioning Treatment of Neural Networks
Published 04-02-2020“…We study the role of depth in training randomly initialized overparameterized neural networks. We give a general result showing that depth improves…”
Get full text
Journal Article -
19
Multiarmed Bandits With Limited Expert Advice
Published 19-06-2013“…We solve the COLT 2013 open problem of \citet{SCB} on minimizing regret in the setting of advice-efficient multiarmed bandits with expert advice. We give an…”
Get full text
Journal Article -
20
Asynchronous Local-SGD Training for Language Modeling
Published 17-01-2024“…Local stochastic gradient descent (Local-SGD), also referred to as federated averaging, is an approach to distributed optimization where each device performs…”
Get full text
Journal Article