Search Results - "Phang, Jason"
-
1
An interpretable classifier for high-resolution breast cancer screening images utilizing weakly supervised localization
Published in Medical image analysis (01-02-2021)“…•We propose a novel neural network model for screening mammography interpretation•Our model outperforms popular models such as ResNet-34 and Faster R-CNN•Our…”
Get full text
Journal Article -
2
Deep Neural Networks Improve Radiologists' Performance in Breast Cancer Screening
Published in IEEE transactions on medical imaging (01-04-2020)“…We present a deep convolutional neural network for breast cancer screening exam classification, trained, and evaluated on over 200000 exams (over 1000000…”
Get full text
Journal Article -
3
Investigating the Effectiveness of HyperTuning via Gisting
Published 26-02-2024“…Gisting (Mu et al., 2023) is a simple method for training models to compress information into fewer token representations using a modified attention mask, and…”
Get full text
Journal Article -
4
Reducing False-Positive Biopsies using Deep Neural Networks that Utilize both Local and Global Image Context of Screening Mammograms
Published in Journal of digital imaging (01-12-2021)“…Breast cancer is the most common cancer in women, and hundreds of thousands of unnecessary biopsies are done around the world at a tremendous cost. It is…”
Get full text
Journal Article -
5
Investigating Efficiently Extending Transformers for Long Input Summarization
Published 08-08-2022“…While large pretrained Transformer models have proven highly capable at tackling natural language tasks, handling long sequence inputs continues to be a…”
Get full text
Journal Article -
6
Unsupervised Sentence Compression using Denoising Auto-Encoders
Published 07-09-2018“…In sentence compression, the task of shortening sentences while retaining the original meaning, models tend to be trained on large corpora containing pairs of…”
Get full text
Journal Article -
7
Fine-Tuned Transformers Show Clusters of Similar Representations Across Layers
Published 17-09-2021“…Despite the success of fine-tuning pretrained language encoders like BERT for downstream natural language understanding (NLU) tasks, it is still poorly…”
Get full text
Journal Article -
8
HyperTuning: Toward Adapting Large Language Models without Back-propagation
Published 22-11-2022“…Fine-tuning large language models for different tasks can be costly and inefficient, and even methods that reduce the number of tuned parameters still require…”
Get full text
Journal Article -
9
Large Language Models as Misleading Assistants in Conversation
Published 16-07-2024“…Large Language Models (LLMs) are able to provide assistance on a wide range of information-seeking tasks. However, model outputs may be misleading, whether…”
Get full text
Journal Article -
10
EleutherAI: Going Beyond "Open Science" to "Science in the Open"
Published 12-10-2022“…Over the past two years, EleutherAI has established itself as a radically novel initiative aimed at both promoting open-source research and conducting research…”
Get full text
Journal Article -
11
Investigating and Simplifying Masking-based Saliency Methods for Model Interpretability
Published 19-10-2020“…Saliency maps that identify the most informative regions of an image for a classifier are valuable for model interpretability. A common approach to creating…”
Get full text
Journal Article -
12
Struc-Bench: Are Large Language Models Really Good at Generating Complex Structured Data?
Published 16-09-2023“…Despite the remarkable capabilities of Large Language Models (LLMs) like GPT-4, producing complex, structured tabular data remains challenging. Our study…”
Get full text
Journal Article -
13
Adversarially Constructed Evaluation Sets Are More Challenging, but May Not Be Fair
Published 15-11-2021“…More capable language models increasingly saturate existing task benchmarks, in some cases outperforming humans. This has left little headroom with which to…”
Get full text
Journal Article -
14
Two Failures of Self-Consistency in the Multi-Step Reasoning of LLMs
Published 23-05-2023“…Transactions on Machine Learning Research (2024) Large language models (LLMs) have achieved widespread success on a variety of in-context few-shot tasks, but…”
Get full text
Journal Article -
15
Two-Turn Debate Doesn't Help Humans Answer Hard Reading Comprehension Questions
Published 19-10-2022“…The use of language-model-based question-answering systems to aid humans in completing difficult tasks is limited, in part, by the unreliability of the text…”
Get full text
Journal Article -
16
SQuALITY: Building a Long-Document Summarization Dataset the Hard Way
Published 23-05-2022“…Summarization datasets are often assembled either by scraping naturally occurring public-domain summaries -- which are nearly always in difficult-to-work-with…”
Get full text
Journal Article -
17
Single-Turn Debate Does Not Help Humans Answer Hard Reading-Comprehension Questions
Published 11-04-2022“…Current QA systems can generate reasonable-sounding yet false answers without explanation or evidence for the generated answer, which is especially problematic…”
Get full text
Journal Article -
18
Pretraining Language Models with Human Preferences
Published 16-02-2023“…Language models (LMs) are pretrained to imitate internet text, including content that would violate human preferences if generated by an LM: falsehoods,…”
Get full text
Journal Article -
19
Sentence Encoders on STILTs: Supplementary Training on Intermediate Labeled-data Tasks
Published 02-11-2018“…Pretraining sentence encoders with language modeling and related unsupervised tasks has recently been shown to be very effective for language understanding…”
Get full text
Journal Article -
20
Do Attention Heads in BERT Track Syntactic Dependencies?
Published 27-11-2019“…We investigate the extent to which individual attention heads in pretrained transformer language models, such as BERT and RoBERTa, implicitly capture syntactic…”
Get full text
Journal Article