Search Results - "Webson, Albert"

  • Showing 1 - 19 results of 19
Refine Results
  1. 1

    Interactive and Visual Prompt Engineering for Ad-hoc Task Adaptation with Large Language Models by Strobelt, Hendrik, Webson, Albert, Sanh, Victor, Hoover, Benjamin, Beyer, Johanna, Pfister, Hanspeter, Rush, Alexander M.

    “…State-of-the-art neural language models can now be used to solve ad-hoc language tasks through zero-shot prompting without the need for supervised training…”
    Get full text
    Journal Article
  2. 2

    Do Prompt-Based Models Really Understand the Meaning of their Prompts? by Webson, Albert, Pavlick, Ellie

    Published 02-09-2021
    “…Recently, a boom of papers has shown extraordinary progress in zero-shot and few-shot learning with various prompt-based models. It is commonly argued that…”
    Get full text
    Journal Article
  3. 3

    In-context Learning Generalizes, But Not Always Robustly: The Case of Syntax by Mueller, Aaron, Webson, Albert, Petty, Jackson, Linzen, Tal

    Published 13-11-2023
    “…In-context learning (ICL) is now a common method for teaching large language models (LLMs) new tasks: given labeled examples in the input context, the LLM…”
    Get full text
    Journal Article
  4. 4

    Are Language Models Worse than Humans at Following Prompts? It's Complicated by Webson, Albert, Loo, Alyssa Marie, Yu, Qinan, Pavlick, Ellie

    Published 17-01-2023
    “…Prompts have been the center of progress in advancing language models' zero-shot and few-shot performance. However, recent work finds that models can perform…”
    Get full text
    Journal Article
  5. 5

    Simfluence: Modeling the Influence of Individual Training Examples by Simulating Training Runs by Guu, Kelvin, Webson, Albert, Pavlick, Ellie, Dixon, Lucas, Tenney, Ian, Bolukbasi, Tolga

    Published 14-03-2023
    “…Training data attribution (TDA) methods offer to trace a model's prediction on any given example back to specific influential training examples. Existing…”
    Get full text
    Journal Article
  6. 6

    Interactive and Visual Prompt Engineering for Ad-hoc Task Adaptation with Large Language Models by Strobelt, Hendrik, Webson, Albert, Sanh, Victor, Hoover, Benjamin, Beyer, Johanna, Pfister, Hanspeter, Rush, Alexander M

    Published 16-08-2022
    “…State-of-the-art neural language models can now be used to solve ad-hoc language tasks through zero-shot prompting without the need for supervised training…”
    Get full text
    Journal Article
  7. 7

    Are "Undocumented Workers" the Same as "Illegal Aliens"? Disentangling Denotation and Connotation in Vector Spaces by Webson, Albert, Chen, Zhizhong, Eickhoff, Carsten, Pavlick, Ellie

    Published 06-10-2020
    “…In politics, neologisms are frequently invented for partisan objectives. For example, "undocumented workers" and "illegal aliens" refer to the same group of…”
    Get full text
    Journal Article
  8. 8

    Larger language models do in-context learning differently by Wei, Jerry, Wei, Jason, Tay, Yi, Tran, Dustin, Webson, Albert, Lu, Yifeng, Chen, Xinyun, Liu, Hanxiao, Huang, Da, Zhou, Denny, Ma, Tengyu

    Published 07-03-2023
    “…We study how in-context learning (ICL) in language models is affected by semantic priors versus input-label mappings. We investigate two setups-ICL with…”
    Get full text
    Journal Article
  9. 9

    The Flan Collection: Designing Data and Methods for Effective Instruction Tuning by Longpre, Shayne, Hou, Le, Vu, Tu, Webson, Albert, Chung, Hyung Won, Tay, Yi, Zhou, Denny, Le, Quoc V, Zoph, Barret, Wei, Jason, Roberts, Adam

    Published 31-01-2023
    “…We study the design decisions of publicly available instruction tuning methods, and break down the development of Flan 2022 (Chung et al., 2022). Through…”
    Get full text
    Journal Article
  10. 10

    Mixture-of-Experts Meets Instruction Tuning:A Winning Combination for Large Language Models by Shen, Sheng, Hou, Le, Zhou, Yanqi, Du, Nan, Longpre, Shayne, Wei, Jason, Chung, Hyung Won, Zoph, Barret, Fedus, William, Chen, Xinyun, Vu, Tu, Wu, Yuexin, Chen, Wuyang, Webson, Albert, Li, Yunxuan, Zhao, Vincent, Yu, Hongkun, Keutzer, Kurt, Darrell, Trevor, Zhou, Denny

    Published 24-05-2023
    “…Sparse Mixture-of-Experts (MoE) is a neural architecture design that can be utilized to add learnable parameters to Large Language Models (LLMs) without…”
    Get full text
    Journal Article
  11. 11
  12. 12
  13. 13
  14. 14

    Crosslingual Generalization through Multitask Finetuning by Muennighoff, Niklas, Wang, Thomas, Sutawika, Lintang, Roberts, Adam, Biderman, Stella, Scao, Teven Le, Bari, M Saiful, Shen, Sheng, Yong, Zheng-Xin, Schoelkopf, Hailey, Tang, Xiangru, Radev, Dragomir, Aji, Alham Fikri, Almubarak, Khalid, Albanie, Samuel, Alyafeai, Zaid, Webson, Albert, Raff, Edward, Raffel, Colin

    Published 03-11-2022
    “…Multitask prompted finetuning (MTF) has been shown to help large language models generalize to new tasks in a zero-shot setting, but so far explorations of MTF…”
    Get full text
    Journal Article
  15. 15

    Gemini 1.5: Unlocking multimodal understanding across millions of tokens of context by Daruki, Samira, Gu, Yiming, Mahdieh, Mahdis, Sundararajan, Mukund, Mentzer, Fabian, He, Antoine, Ho, Lewis, Chowdhery, Aakanksha, Petrushkina, Anastasia, Patil, Piyush, Zhao, Jeffrey, Agarwal, Rishabh, Rajwar, Ravi, Batsaikhan, Bat-Orgil, Kar, Arjun, Sang, Ruoxin, Hand, Steven, Jain, Vihan, Sethi, Nikhil, Bolina, Vijay, Levine, Nir, Addanki, Ravi, Lin, Hanzhao, Rutherford, Eliza, Choe, HyunJeong, Rakićević, Nemanja, Huang, Ronny, Chen, Jeremy, Sulsky, Yury, Greer, Jeremy, Chen, Charlie, Chen, Phil, Tobin, Taylor, Thacker, Phoebe, Davoodi, Elnaz, Proleev, Lev, Baddepudi, Anirudh, Nguyen, Duc Dung, Tagliasacchi, Marco, Du, Cosmo, Martins, Danilo, Shah, Premal, Feinberg, Vladimir, Smith, Charlotte, Shakeri, Siamak, Soparkar, Kedar, Barker, David, Houlsby, Neil, Magni, Alberto, Thakoor, Shantanu, Koop, Anna, Zilka, Lukas, Cai, Honglong, Kallakuri, Praveen, Ghemawat, Sanjay, Tao, David, Green, Tim, Zhou, Denny, Farabet, Clement, Broder, Josef, Repina, Alena, Burns, Andrea, Zhang, Lei, Cobo, Luis C, Simon, Jon, Tenney, Ian, Chen, Minmin, Xu, Yaming, Fernando, Nick, Singhal, Achintya, Jia, Johnson, Crocker, Remi, Liu, Tianqi, White, Anais, McCarthy, Sara, McWilliams, Brian, Zhang, Zhishuai, Felt, Nick, Shenoy, Ashish, Elisseeff, Andre, Zeng, William, Choquette-Choo, Christopher A, Manyika, James, Bolukbasi, Tolga, Salama, Khalid, Askham, Harry, Parisi, Aaron, Vijayakumar, Anitha, Carthy, Sara Mc, Kim, Lucy, Larsen, Rasmus, Ijazi, Joana, Zhao, Shubin, Dhillon, Inderjit, Sedghi, Hanie, Jerome, Sammy, Algymr, Anton, Anklin, Valentin, Baeuml, Martin, Petrov, Slav

    Published 08-03-2024
    “…In this report, we introduce the Gemini 1.5 family of models, representing the next generation of highly compute-efficient multimodal models capable of…”
    Get full text
    Journal Article
  16. 16
  17. 17
  18. 18

    BLOOM: A 176B-Parameter Open-Access Multilingual Language Model by Scao, Teven Le, Fan, Angela, Gallé, Matthias, Webson, Albert, Wang, Thomas, Bekman, Stas, Laurençon, Hugo, Launay, Julien, Raffel, Colin, Simhi, Adi, Alfassy, Amit, Rogers, Anna, Leong, Colin, van Strien, Daniel, Ponferrada, Eduardo González, Levkovizh, Efrat, Benyamina, Hamza, Tran, Hieu, Yu, Ian, Johnson, Isaac, Bhattacharjee, Joydeep, Von Werra, Leandro, Dey, Manan, Jiang, Mike Tian-Jian, Jauhar, Mohammad A, Kassner, Nora, Pyysalo, Sampo, Pai, Suhas, Schick, Timo, Thrush, Tristan, Nikoulina, Vassilina, Laippala, Veronika, Heinzerling, Benjamin, Taşar, Davut Emre, Salesky, Elizabeth, Lee, Wilson Y, Szczechla, Eliza, Chhablani, Gunjan, Wang, Han, Rozen, Jos, Manica, Matteo, Nayak, Nihal, Teehan, Ryan, Albanie, Samuel, Shen, Sheng, Ben-David, Srulik, Kim, Taewoon, Neeraj, Trishala, Roberts, Adam, Tae, Jaesung, Phang, Jason, Press, Ofir, Ryabinin, Max, Peyrounette, Myriam, Patry, Nicolas, Cornette, Pierre, Dettmers, Tim, Ligozat, Anne-Laure, Névéol, Aurélie, Taktasheva, Ekaterina, Kalo, Jan-Christoph, Clive, Jordan, Kim, Najoung, Mirkin, Shachar, Pais, Shani, Pruksachatkun, Yada, Pestana, Amanda, Faranak, Amy, Santos, Ana, HajiHosseini, Azadeh, Ajibade, Benjamin, Saxena, Bharat, Nguyen, Duong A, Rezanejad, Habib, Bhattacharya, Indrani, Nejadgholi, Isar, McKenna, Michael, Burynok, Mykola, Rajani, Nazneen, Samuel, Olanrewaju, Kromann, Rasmus, Shubber, Sarmad, Viguier, Sylvain, Miranda-Escalada, Antonio, Singh, Ayush, Manjavacas, Enrique, Barth, Fabio, Bulchandani, Lokesh, Nezhurina, Marianna, Liu, Minna, Kang, Myungsun, Dahlberg, Nathan, Chandrasekhar, Ramya, Eisenberg, Renata, Canalli, Rodrigo, Schweter, Stefan, Laud, Tanmay, Kainuma, Tomoya, Venkatraman, Yash, Xu, Yingxin

    Published 09-11-2022
    “…Large language models (LLMs) have been shown to be able to perform new tasks based on a few demonstrations or natural language instructions. While these…”
    Get full text
    Journal Article
  19. 19