Search Results - "Phang, Jason"

Refine Results
  1. 1

    An interpretable classifier for high-resolution breast cancer screening images utilizing weakly supervised localization by Shen, Yiqiu, Wu, Nan, Phang, Jason, Park, Jungkyu, Liu, Kangning, Tyagi, Sudarshini, Heacock, Laura, Kim, S. Gene, Moy, Linda, Cho, Kyunghyun, Geras, Krzysztof J.

    Published in Medical image analysis (01-02-2021)
    “…•We propose a novel neural network model for screening mammography interpretation•Our model outperforms popular models such as ResNet-34 and Faster R-CNN•Our…”
    Get full text
    Journal Article
  2. 2
  3. 3

    Investigating the Effectiveness of HyperTuning via Gisting by Phang, Jason

    Published 26-02-2024
    “…Gisting (Mu et al., 2023) is a simple method for training models to compress information into fewer token representations using a modified attention mask, and…”
    Get full text
    Journal Article
  4. 4

    Reducing False-Positive Biopsies using Deep Neural Networks that Utilize both Local and Global Image Context of Screening Mammograms by Wu, Nan, Huang, Zhe, Shen, Yiqiu, Park, Jungkyu, Phang, Jason, Makino, Taro, Gene Kim, S., Cho, Kyunghyun, Heacock, Laura, Moy, Linda, Geras, Krzysztof J.

    Published in Journal of digital imaging (01-12-2021)
    “…Breast cancer is the most common cancer in women, and hundreds of thousands of unnecessary biopsies are done around the world at a tremendous cost. It is…”
    Get full text
    Journal Article
  5. 5

    Investigating Efficiently Extending Transformers for Long Input Summarization by Phang, Jason, Zhao, Yao, Liu, Peter J

    Published 08-08-2022
    “…While large pretrained Transformer models have proven highly capable at tackling natural language tasks, handling long sequence inputs continues to be a…”
    Get full text
    Journal Article
  6. 6

    Unsupervised Sentence Compression using Denoising Auto-Encoders by Févry, Thibault, Phang, Jason

    Published 07-09-2018
    “…In sentence compression, the task of shortening sentences while retaining the original meaning, models tend to be trained on large corpora containing pairs of…”
    Get full text
    Journal Article
  7. 7

    Fine-Tuned Transformers Show Clusters of Similar Representations Across Layers by Phang, Jason, Liu, Haokun, Bowman, Samuel R

    Published 17-09-2021
    “…Despite the success of fine-tuning pretrained language encoders like BERT for downstream natural language understanding (NLU) tasks, it is still poorly…”
    Get full text
    Journal Article
  8. 8

    HyperTuning: Toward Adapting Large Language Models without Back-propagation by Phang, Jason, Mao, Yi, He, Pengcheng, Chen, Weizhu

    Published 22-11-2022
    “…Fine-tuning large language models for different tasks can be costly and inefficient, and even methods that reduce the number of tuned parameters still require…”
    Get full text
    Journal Article
  9. 9

    Large Language Models as Misleading Assistants in Conversation by Hou, Betty Li, Shi, Kejian, Phang, Jason, Aung, James, Adler, Steven, Campbell, Rosie

    Published 16-07-2024
    “…Large Language Models (LLMs) are able to provide assistance on a wide range of information-seeking tasks. However, model outputs may be misleading, whether…”
    Get full text
    Journal Article
  10. 10

    EleutherAI: Going Beyond "Open Science" to "Science in the Open" by Phang, Jason, Bradley, Herbie, Gao, Leo, Castricato, Louis, Biderman, Stella

    Published 12-10-2022
    “…Over the past two years, EleutherAI has established itself as a radically novel initiative aimed at both promoting open-source research and conducting research…”
    Get full text
    Journal Article
  11. 11

    Investigating and Simplifying Masking-based Saliency Methods for Model Interpretability by Phang, Jason, Park, Jungkyu, Geras, Krzysztof J

    Published 19-10-2020
    “…Saliency maps that identify the most informative regions of an image for a classifier are valuable for model interpretability. A common approach to creating…”
    Get full text
    Journal Article
  12. 12

    Struc-Bench: Are Large Language Models Really Good at Generating Complex Structured Data? by Tang, Xiangru, Zong, Yiming, Phang, Jason, Zhao, Yilun, Zhou, Wangchunshu, Cohan, Arman, Gerstein, Mark

    Published 16-09-2023
    “…Despite the remarkable capabilities of Large Language Models (LLMs) like GPT-4, producing complex, structured tabular data remains challenging. Our study…”
    Get full text
    Journal Article
  13. 13

    Adversarially Constructed Evaluation Sets Are More Challenging, but May Not Be Fair by Phang, Jason, Chen, Angelica, Huang, William, Bowman, Samuel R

    Published 15-11-2021
    “…More capable language models increasingly saturate existing task benchmarks, in some cases outperforming humans. This has left little headroom with which to…”
    Get full text
    Journal Article
  14. 14

    Two Failures of Self-Consistency in the Multi-Step Reasoning of LLMs by Chen, Angelica, Phang, Jason, Parrish, Alicia, Padmakumar, Vishakh, Zhao, Chen, Bowman, Samuel R, Cho, Kyunghyun

    Published 23-05-2023
    “…Transactions on Machine Learning Research (2024) Large language models (LLMs) have achieved widespread success on a variety of in-context few-shot tasks, but…”
    Get full text
    Journal Article
  15. 15

    Two-Turn Debate Doesn't Help Humans Answer Hard Reading Comprehension Questions by Parrish, Alicia, Trivedi, Harsh, Nangia, Nikita, Padmakumar, Vishakh, Phang, Jason, Saimbhi, Amanpreet Singh, Bowman, Samuel R

    Published 19-10-2022
    “…The use of language-model-based question-answering systems to aid humans in completing difficult tasks is limited, in part, by the unreliability of the text…”
    Get full text
    Journal Article
  16. 16

    SQuALITY: Building a Long-Document Summarization Dataset the Hard Way by Wang, Alex, Pang, Richard Yuanzhe, Chen, Angelica, Phang, Jason, Bowman, Samuel R

    Published 23-05-2022
    “…Summarization datasets are often assembled either by scraping naturally occurring public-domain summaries -- which are nearly always in difficult-to-work-with…”
    Get full text
    Journal Article
  17. 17

    Single-Turn Debate Does Not Help Humans Answer Hard Reading-Comprehension Questions by Parrish, Alicia, Trivedi, Harsh, Perez, Ethan, Chen, Angelica, Nangia, Nikita, Phang, Jason, Bowman, Samuel R

    Published 11-04-2022
    “…Current QA systems can generate reasonable-sounding yet false answers without explanation or evidence for the generated answer, which is especially problematic…”
    Get full text
    Journal Article
  18. 18

    Pretraining Language Models with Human Preferences by Korbak, Tomasz, Shi, Kejian, Chen, Angelica, Bhalerao, Rasika, Buckley, Christopher L, Phang, Jason, Bowman, Samuel R, Perez, Ethan

    Published 16-02-2023
    “…Language models (LMs) are pretrained to imitate internet text, including content that would violate human preferences if generated by an LM: falsehoods,…”
    Get full text
    Journal Article
  19. 19

    Sentence Encoders on STILTs: Supplementary Training on Intermediate Labeled-data Tasks by Phang, Jason, Févry, Thibault, Bowman, Samuel R

    Published 02-11-2018
    “…Pretraining sentence encoders with language modeling and related unsupervised tasks has recently been shown to be very effective for language understanding…”
    Get full text
    Journal Article
  20. 20

    Do Attention Heads in BERT Track Syntactic Dependencies? by Htut, Phu Mon, Phang, Jason, Bordia, Shikha, Bowman, Samuel R

    Published 27-11-2019
    “…We investigate the extent to which individual attention heads in pretrained transformer language models, such as BERT and RoBERTa, implicitly capture syntactic…”
    Get full text
    Journal Article