Search Results - "Launay, Julien"
-
1
Effect of viscous dissipation in the prediction of thermal behavior of an elastomer cylindrical flow
Published in Journal of materials processing technology (01-02-2018)“…In this work, the thermal behavior of an elastomer flow all along a cylindrical runner placed at the outlet of an extruder is studied. A cylindrical runner…”
Get full text
Journal Article -
2
Development of a thermal Reduced Order Model with explicit dependence on viscosity for a generalized Newtonian fluid
Published in Journal of non-Newtonian fluid mechanics (01-10-2018)“…This work falls within the general framework of melted polymers flows characterization. It deals with the development of a thermo-rheological Reduced Order…”
Get full text
Journal Article -
3
Adversarial Robustness by Design Through Analog Computing And Synthetic Gradients
Published in ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) (23-05-2022)“…We propose a new defense mechanism against adversarial at-tacks inspired by an optical co-processor, providing robustness without compromising natural accuracy…”
Get full text
Conference Proceeding -
4
Scaling Laws Beyond Backpropagation
Published 26-10-2022“…Alternatives to backpropagation have long been studied to better understand how biological brains may learn. Recently, they have also garnered interest as a…”
Get full text
Journal Article -
5
Is the Number of Trainable Parameters All That Actually Matters?
Published 24-09-2021“…Recent work has identified simple empirical scaling laws for language models, linking compute budget, dataset size, model size, and autoregressive modeling…”
Get full text
Journal Article -
6
ROPUST: Improving Robustness through Fine-tuning with Photonic Processors and Synthetic Gradients
Published 06-07-2021“…Robustness to adversarial attacks is typically obtained through expensive adversarial training with Projected Gradient Descent. Here we introduce ROPUST, a…”
Get full text
Journal Article -
7
Principled Training of Neural Networks with Direct Feedback Alignment
Published 11-06-2019“…The backpropagation algorithm has long been the canonical training method for neural networks. Modern paradigms are implicitly optimized for it, and numerous…”
Get full text
Journal Article -
8
The RefinedWeb Dataset for Falcon LLM: Outperforming Curated Corpora with Web Data, and Web Data Only
Published 01-06-2023“…Large language models are commonly trained on a mixture of filtered web data and curated high-quality corpora, such as social media conversations, books, or…”
Get full text
Journal Article -
9
Direct Feedback Alignment Scales to Modern Deep Learning Tasks and Architectures
Published 23-06-2020“…Advances in Neural Information Processing Systems, v33, pages 9346--9360, 2020 Despite being the workhorse of deep learning, the backpropagation algorithm is…”
Get full text
Journal Article -
10
The Falcon Series of Open Language Models
Published 28-11-2023“…We introduce the Falcon series: 7B, 40B, and 180B parameters causal decoder-only models trained on a diverse high-quality corpora predominantly assembled from…”
Get full text
Journal Article -
11
Optical training of large-scale Transformers and deep neural networks with direct feedback alignment
Published 01-09-2024“…Modern machine learning relies nearly exclusively on dedicated electronic hardware accelerators. Photonic approaches, with low consumption and high operation…”
Get full text
Journal Article -
12
PAGnol: An Extra-Large French Generative Model
Published 16-10-2021“…Access to large pre-trained models of varied architectures, in many different languages, is central to the democratization of NLP. We introduce PAGnol, a…”
Get full text
Journal Article -
13
Artificial Neural Network Training on an Optical Processor via Direct Feedback Alignment
Published in 2023 Conference on Lasers and Electro-Optics Europe & European Quantum Electronics Conference (CLEO/Europe-EQEC) (26-06-2023)“…Artificial Neural Networks (ANN) are habitually trained via the back-propagation (BP) algorithm. This approach has been extremely successful: Current models…”
Get full text
Conference Proceeding -
14
Photonic Differential Privacy with Direct Feedback Alignment
Published 07-06-2021“…NeurIPS 2021 Optical Processing Units (OPUs) -- low-power photonic chips dedicated to large scale random projections -- have been used in previous work to…”
Get full text
Journal Article -
15
What Language Model Architecture and Pretraining Objective Work Best for Zero-Shot Generalization?
Published 12-04-2022“…Large pretrained Transformer language models have been shown to exhibit zero-shot generalization, i.e. they can perform a wide variety of tasks that they were…”
Get full text
Journal Article -
16
Adversarial Robustness by Design through Analog Computing and Synthetic Gradients
Published 06-01-2021“…ICASSP 2022 - IEEE International Conference on Acoustics, Speech and Signal Processing, We propose a new defense mechanism against adversarial attacks inspired…”
Get full text
Journal Article -
17
Hardware Beyond Backpropagation: a Photonic Co-Processor for Direct Feedback Alignment
Published 11-12-2020“…The scaling hypothesis motivates the expansion of models past trillions of parameters as a path towards better performance. Recent significant developments,…”
Get full text
Journal Article -
18
Light-in-the-loop: using a photonics co-processor for scalable training of neural networks
Published 02-06-2020“…As neural networks grow larger and more complex and data-hungry, training costs are skyrocketing. Especially when lifelong learning is necessary, such as in…”
Get full text
Journal Article -
19
What Language Model to Train if You Have One Million GPU Hours?
Published 27-10-2022“…The crystallization of modeling methods around the Transformer architecture has been a boon for practitioners. Simple, well-motivated architectural variations…”
Get full text
Journal Article -
20
LightOn Optical Processing Unit : Scaling-up AI and HPC with a Non von Neumann co-processor
Published in 2021 IEEE Hot Chips 33 Symposium (HCS) (22-08-2021)“…Beyond pure Von Neumann processing Scalability of AI / HPC models is limited by the Von Neumann bottleneck for accessing massive amounts of memory, driving up…”
Get full text
Conference Proceeding