Search Results - "Balles, Lukas"
-
1
Competitive Collaboration: Joint Unsupervised Learning of Depth, Camera Motion, Optical Flow and Motion Segmentation
Published in 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (01-06-2019)“…We address the unsupervised learning of several interconnected problems in low-level vision: single view depth prediction, camera motion estimation, optical…”
Get full text
Conference Proceeding -
2
Cyclic (Amino)(aryl)carbenes Enter the Field of Chromophore Ligands: Expanded π System Leads to Unusually Deep Red Emitting Cu I Compounds
Published in Journal of the American Chemical Society (13-05-2020)“…A series of copper(I) complexes bearing a cyclic (amino)(aryl)carbene (CAArC) ligand with various complex geometries have been investigated in great detail…”
Get full text
Journal Article -
3
A Negative Result on Gradient Matching for Selective Backprop
Published 08-12-2023“…With increasing scale in model and dataset size, the training of deep neural networks becomes a massive computational burden. One approach to speed up the…”
Get full text
Journal Article -
4
Choice of PEFT Technique in Continual Learning: Prompt Tuning is Not All You Need
Published 05-06-2024“…Recent Continual Learning (CL) methods have combined pretrained Transformers with prompt tuning, a parameter-efficient fine-tuning (PEFT) technique. We argue…”
Get full text
Journal Article -
5
Continual Learning with Low Rank Adaptation
Published 29-11-2023“…Recent work using pretrained transformers has shown impressive performance when fine-tuned with data from the downstream problem of interest. However, they…”
Get full text
Journal Article -
6
Gradient-Matching Coresets for Rehearsal-Based Continual Learning
Published 28-03-2022“…The goal of continual learning (CL) is to efficiently update a machine learning model with new data without forgetting previously-learned knowledge. Most…”
Get full text
Journal Article -
7
Gradient-matching coresets for continual learning
Published 09-12-2021“…We devise a coreset selection method based on the idea of gradient matching: The gradients induced by the coreset should match, as closely as possible, those…”
Get full text
Journal Article -
8
Renate: A Library for Real-World Continual Learning
Published 24-04-2023“…Continual learning enables the incremental training of machine learning models on non-stationary data streams.While academic interest in the topic is high,…”
Get full text
Journal Article -
9
Dissecting Adam: The Sign, Magnitude and Variance of Stochastic Gradients
Published 22-05-2017“…The ADAM optimizer is exceedingly popular in the deep learning community. Often it works very well, sometimes it doesn't. Why? We interpret ADAM as a…”
Get full text
Journal Article -
10
The Geometry of Sign Gradient Descent
Published 19-02-2020“…Sign-based optimization methods have become popular in machine learning due to their favorable communication cost in distributed optimization and their…”
Get full text
Journal Article -
11
Cyclic (Amino)(aryl)carbenes Enter the Field of Chromophore Ligands: Expanded π System Leads to Unusually Deep Red Emitting CuI Compounds
Published in Journal of the American Chemical Society (13-05-2020)“…A series of copper(I) complexes bearing a cyclic (amino)(aryl)carbene (CAArC) ligand with various complex geometries have been investigated in great detail…”
Get full text
Journal Article -
12
u-$\mu$P: The Unit-Scaled Maximal Update Parametrization
Published 24-07-2024“…The Maximal Update Parametrization ($\mu$P) aims to make the optimal hyperparameters (HPs) of a model independent of its size, allowing them to be swept using…”
Get full text
Journal Article -
13
Limitations of the Empirical Fisher Approximation for Natural Gradient Descent
Published 29-05-2019“…Natural gradient descent, which preconditions a gradient descent update with the Fisher information matrix of the underlying statistical model, is a way to…”
Get full text
Journal Article -
14
DeepOBS: A Deep Learning Optimizer Benchmark Suite
Published 13-03-2019“…Because the choice and tuning of the optimizer affects the speed, and ultimately the performance of deep learning, there is significant past and recent…”
Get full text
Journal Article -
15
PASHA: Efficient HPO and NAS with Progressive Resource Allocation
Published 14-07-2022“…Hyperparameter optimization (HPO) and neural architecture search (NAS) are methods of choice to obtain the best-in-class machine learning models, but in…”
Get full text
Journal Article -
16
Self-Tuning Stochastic Optimization with Curvature-Aware Gradient Filtering
Published 09-11-2020“…Standard first-order stochastic optimization algorithms base their updates solely on the average mini-batch gradient, and it has been shown that tracking…”
Get full text
Journal Article -
17
Coupling Adaptive Batch Sizes with Learning Rates
Published 15-12-2016“…Mini-batch stochastic gradient descent and variants thereof have become standard for large-scale empirical risk minimization like the training of neural…”
Get full text
Journal Article -
18
Early Stopping without a Validation Set
Published 28-03-2017“…Early stopping is a widely used technique to prevent poor generalization performance when training an over-expressive model by means of gradient-based…”
Get full text
Journal Article -
19
Competitive Collaboration: Joint Unsupervised Learning of Depth, Camera Motion, Optical Flow and Motion Segmentation
Published 24-05-2018“…We address the unsupervised learning of several interconnected problems in low-level vision: single view depth prediction, camera motion estimation, optical…”
Get full text
Journal Article