Search Results - "Kobayashi, Seijin"

  • Showing 1 - 18 results of 18
Refine Results
  1. 1

    Comparison of Hepatic Resection and Radiofrequency Ablation for Small Hepatocellular Carcinoma: A Meta-Analysis of 16,103 Patients by Xu, Qinghua, Kobayashi, Seijin, Ye, Xun, Meng, Xia

    Published in Scientific reports (28-11-2014)
    “…We performed a meta-analysis to evaluate the therapeutic effects of radiofrequency ablation (RFA) and surgical hepatic resection (HR) in the treatment of small…”
    Get full text
    Journal Article
  2. 2

    Wheel Defect Detection With Machine Learning by Krummenacher, Gabriel, Ong, Cheng Soon, Koller, Stefan, Kobayashi, Seijin, Buhmann, Joachim M.

    “…Wheel defects on railway wagons have been identified as an important source of damage to the railway infrastructure and rolling stock. They also cause noise…”
    Get full text
    Journal Article
  3. 3
  4. 4

    Weight decay induces low-rank attention layers by Kobayashi, Seijin, Akram, Yassir, Von Oswald, Johannes

    Published 31-10-2024
    “…The effect of regularizers such as weight decay when training deep neural networks is not well understood. We study the influence of weight decay as well as…”
    Get full text
    Journal Article
  5. 5

    Learning Randomized Algorithms with Transformers by von Oswald, Johannes, Kobayashi, Seijin, Akram, Yassir, Steger, Angelika

    Published 20-08-2024
    “…Randomization is a powerful tool that endows algorithms with remarkable properties. For instance, randomized algorithms excel in adversarial settings, often…”
    Get full text
    Journal Article
  6. 6

    Attention as a Hypernetwork by Schug, Simon, Kobayashi, Seijin, Akram, Yassir, Sacramento, João, Pascanu, Razvan

    Published 09-06-2024
    “…Transformers can under some circumstances generalize to novel problem instances whose constituent parts might have been encountered during training but whose…”
    Get full text
    Journal Article
  7. 7

    Would I have gotten that reward? Long-term credit assignment by counterfactual contribution analysis by Meulemans, Alexander, Schug, Simon, Kobayashi, Seijin, Daw, Nathaniel, Wayne, Gregory

    Published 29-06-2023
    “…To make reinforcement learning more sample efficient, we need better credit assignment methods that measure an action's influence on future rewards. Building…”
    Get full text
    Journal Article
  8. 8

    Disentangling the Predictive Variance of Deep Ensembles through the Neural Tangent Kernel by Kobayashi, Seijin, Aceituno, Pau Vilimelis, von Oswald, Johannes

    Published 18-10-2022
    “…Identifying unfamiliar inputs, also known as out-of-distribution (OOD) detection, is a crucial property of any decision making process. A simple and…”
    Get full text
    Journal Article
  9. 9

    Gated recurrent neural networks discover attention by Zucchet, Nicolas, Kobayashi, Seijin, Akram, Yassir, von Oswald, Johannes, Larcher, Maxime, Steger, Angelika, Sacramento, João

    Published 04-09-2023
    “…Recent architectural developments have enabled recurrent neural networks (RNNs) to reach and even surpass the performance of Transformers on certain sequence…”
    Get full text
    Journal Article
  10. 10

    When can transformers compositionally generalize in-context? by Kobayashi, Seijin, Schug, Simon, Akram, Yassir, Redhardt, Florian, von Oswald, Johannes, Pascanu, Razvan, Lajoie, Guillaume, Sacramento, João

    Published 16-07-2024
    “…Many tasks can be composed from a few independent components. This gives rise to a combinatorial explosion of possible tasks, only some of which might be…”
    Get full text
    Journal Article
  11. 11

    Discovering modular solutions that generalize compositionally by Schug, Simon, Kobayashi, Seijin, Akram, Yassir, Wołczyk, Maciej, Proca, Alexandra, von Oswald, Johannes, Pascanu, Razvan, Sacramento, João, Steger, Angelika

    Published 22-12-2023
    “…Many complex tasks can be decomposed into simpler, independent parts. Discovering such underlying compositional structure has the potential to enable…”
    Get full text
    Journal Article
  12. 12

    Multi-agent cooperation through learning-aware policy gradients by Meulemans, Alexander, Kobayashi, Seijin, von Oswald, Johannes, Scherrer, Nino, Elmoznino, Eric, Richards, Blake, Lajoie, Guillaume, Arcas, Blaise Agüera y, Sacramento, João

    Published 24-10-2024
    “…Self-interested individuals often fail to cooperate, posing a fundamental challenge for multi-agent learning. How can we achieve cooperation among…”
    Get full text
    Journal Article
  13. 13

    Meta-Learning via Classifier(-free) Diffusion Guidance by Nava, Elvis, Kobayashi, Seijin, Yin, Yifei, Katzschmann, Robert K, Grewe, Benjamin F

    Published 17-10-2022
    “…We introduce meta-learning algorithms that perform zero-shot weight-space adaptation of neural network models to unseen tasks. Our methods repurpose the…”
    Get full text
    Journal Article
  14. 14

    The least-control principle for local learning at equilibrium by Meulemans, Alexander, Zucchet, Nicolas, Kobayashi, Seijin, von Oswald, Johannes, Sacramento, João

    Published 04-07-2022
    “…Equilibrium systems are a powerful way to express neural computations. As special cases, they include models of great current interest in both neuroscience and…”
    Get full text
    Journal Article
  15. 15

    Learning where to learn: Gradient sparsity in meta and continual learning by von Oswald, Johannes, Zhao, Dominic, Kobayashi, Seijin, Schug, Simon, Caccia, Massimo, Zucchet, Nicolas, Sacramento, João

    Published 27-10-2021
    “…Finding neural network weights that generalize well from small datasets is difficult. A promising approach is to learn a weight initialization such that a…”
    Get full text
    Journal Article
  16. 16

    Uncovering mesa-optimization algorithms in Transformers by von Oswald, Johannes, Schlegel, Maximilian, Meulemans, Alexander, Kobayashi, Seijin, Niklasson, Eyvind, Zucchet, Nicolas, Scherrer, Nino, Miller, Nolan, Sandler, Mark, Arcas, Blaise Agüera y, Vladymyrov, Max, Pascanu, Razvan, Sacramento, João

    Published 11-09-2023
    “…Some autoregressive models exhibit in-context learning capabilities: being able to learn as an input sequence is processed, without undergoing any parameter…”
    Get full text
    Journal Article
  17. 17

    Neural networks with late-phase weights by von Oswald, Johannes, Kobayashi, Seijin, Meulemans, Alexander, Henning, Christian, Grewe, Benjamin F, Sacramento, João

    Published 25-07-2020
    “…Published as a conference paper at ICLR 2021 The largely successful method of training neural networks is to learn their weights using some variant of…”
    Get full text
    Journal Article
  18. 18

    Posterior Meta-Replay for Continual Learning by Henning, Christian, Cervera, Maria R, D'Angelo, Francesco, von Oswald, Johannes, Traber, Regina, Ehret, Benjamin, Kobayashi, Seijin, Grewe, Benjamin F, Sacramento, João

    Published 01-03-2021
    “…Learning a sequence of tasks without access to i.i.d. observations is a widely studied form of continual learning (CL) that remains challenging. In principle,…”
    Get full text
    Journal Article