Reproducible Speech Research with the Artificial Intelligence--Ready PERCEPT Corpora

Background: Publicly available speech corpora facilitate reproducible research by providing open-access data for participants who have consented/assented to data sharing among different research teams. Such corpora can also support clinical education, including perceptual training and training in th...

Full description

Saved in:
Bibliographic Details
Published in:Journal of speech, language, and hearing research Vol. 66; no. 6; pp. 1986 - 2009
Main Authors: Benway, Nina R, Preston, Jonathan L, Hitchcock, Elaine, Rose, Yvan, Salekin, Asif, Liang, Wendy, McAllister, Tara
Format: Journal Article
Language:English
Published: United States American Speech-Language-Hearing Association 20-06-2023
Subjects:
Online Access:Get more information
Tags: Add Tag
No Tags, Be the first to tag this record!
Abstract Background: Publicly available speech corpora facilitate reproducible research by providing open-access data for participants who have consented/assented to data sharing among different research teams. Such corpora can also support clinical education, including perceptual training and training in the use of speech analysis tools. Purpose: In this research note, we introduce the PERCEPT (Perceptual Error Rating for the Clinical Evaluation of Phonetic Targets) corpora, PERCEPT-R (Rhotics) and PERCEPT-GFTA (Goldman-Fristoe Test of Articulation), which together contain over 36 hr of speech audio (> 125,000 syllable, word, and phrase utterances) from children, adolescents, and young adults aged 6-24 years with speech sound disorder (primarily residual speech sound disorders impacting /[Voiced alveolar and postalveolar approximant]s/) and age-matched peers. We highlight PhonBank as the repository for the corpora and demonstrate use of the associated speech analysis software, Phon, to query PERCEPT-R. A worked example of research with PERCEPT-R, suitable for clinical education and research training, is included as an appendix. Support for end users and information/descriptive statistics for future releases of the PERCEPT corpora can be found in a dedicated Slack channel. Finally, we discuss the potential for PERCEPT corpora to support the training of artificial intelligence clinical speech technology appropriate for use with children with speech sound disorders, the development of which has historically been constrained by the limited representation of either children or individuals with speech impairments in publicly available training corpora. Conclusions: We demonstrate the use of PERCEPT corpora, PhonBank, and Phon for clinical training and research questions appropriate to child citation speech. Increased use of these tools has the potential to enhance reproducibility in the study of speech development and disorders.
AbstractList Publicly available speech corpora facilitate reproducible research by providing open-access data for participants who have consented/assented to data sharing among different research teams. Such corpora can also support clinical education, including perceptual training and training in the use of speech analysis tools. In this research note, we introduce the PERCEPT (Perceptual Error Rating for the Clinical Evaluation of Phonetic Targets) corpora, PERCEPT-R (Rhotics) and PERCEPT-GFTA (Goldman-Fristoe Test of Articulation), which together contain over 36 hr of speech audio (> 125,000 syllable, word, and phrase utterances) from children, adolescents, and young adults aged 6-24 years with speech sound disorder (primarily residual speech sound disorders impacting /ɹ/) and age-matched peers. We highlight PhonBank as the repository for the corpora and demonstrate use of the associated speech analysis software, Phon, to query PERCEPT-R. A worked example of research with PERCEPT-R, suitable for clinical education and research training, is included as an appendix. Support for end users and information/descriptive statistics for future releases of the PERCEPT corpora can be found in a dedicated Slack channel. Finally, we discuss the potential for PERCEPT corpora to support the training of artificial intelligence clinical speech technology appropriate for use with children with speech sound disorders, the development of which has historically been constrained by the limited representation of either children or individuals with speech impairments in publicly available training corpora. We demonstrate the use of PERCEPT corpora, PhonBank, and Phon for clinical training and research questions appropriate to child citation speech. Increased use of these tools has the potential to enhance reproducibility in the study of speech development and disorders.
Background: Publicly available speech corpora facilitate reproducible research by providing open-access data for participants who have consented/assented to data sharing among different research teams. Such corpora can also support clinical education, including perceptual training and training in the use of speech analysis tools. Purpose: In this research note, we introduce the PERCEPT (Perceptual Error Rating for the Clinical Evaluation of Phonetic Targets) corpora, PERCEPT-R (Rhotics) and PERCEPT-GFTA (Goldman-Fristoe Test of Articulation), which together contain over 36 hr of speech audio (> 125,000 syllable, word, and phrase utterances) from children, adolescents, and young adults aged 6-24 years with speech sound disorder (primarily residual speech sound disorders impacting /[Voiced alveolar and postalveolar approximant]s/) and age-matched peers. We highlight PhonBank as the repository for the corpora and demonstrate use of the associated speech analysis software, Phon, to query PERCEPT-R. A worked example of research with PERCEPT-R, suitable for clinical education and research training, is included as an appendix. Support for end users and information/descriptive statistics for future releases of the PERCEPT corpora can be found in a dedicated Slack channel. Finally, we discuss the potential for PERCEPT corpora to support the training of artificial intelligence clinical speech technology appropriate for use with children with speech sound disorders, the development of which has historically been constrained by the limited representation of either children or individuals with speech impairments in publicly available training corpora. Conclusions: We demonstrate the use of PERCEPT corpora, PhonBank, and Phon for clinical training and research questions appropriate to child citation speech. Increased use of these tools has the potential to enhance reproducibility in the study of speech development and disorders.
Author McAllister, Tara
Liang, Wendy
Salekin, Asif
Benway, Nina R
Hitchcock, Elaine
Preston, Jonathan L
Rose, Yvan
Author_xml – sequence: 1
  fullname: Benway, Nina R
– sequence: 2
  fullname: Preston, Jonathan L
– sequence: 3
  fullname: Hitchcock, Elaine
– sequence: 4
  fullname: Rose, Yvan
– sequence: 5
  fullname: Salekin, Asif
– sequence: 6
  fullname: Liang, Wendy
– sequence: 7
  fullname: McAllister, Tara
BackLink http://eric.ed.gov/ERICWebPortal/detail?accno=EJ1396795$$DView record in ERIC
https://www.ncbi.nlm.nih.gov/pubmed/37319018$$D View this record in MEDLINE/PubMed
BookMark eNo9T21LwzAYDDJxL_oLRMkfiOa1bT6O0rmNgaObn0eaPnGRri1ph-zfW5h6X-7gjuNuikZ1UwNCj4y-MCrlK6dcHNa7zTInnBNKhRQ3aMKUSohmlI8GTTUnUiTJGE277osOYDK6Q2MRC6YpSyZon0MbmvJsfVEB3rUA9ohz6MCEQXz7_oj7I-B56L3z1psKr-oeqsp_Qm2BkBxMecHbLE-z7R6nTWibYO7RrTNVBw-_PEMfi2yfLsnm_W2VzjfEipj3hBtbaEoLA0pxMMCdKLWKIqN1DCpyMTNlaQrqQPBhOI8LKalxMnLCUmUUn6Hna297Lk5QHtrgTyZcDn_3hsDTNQDB2387WzOho1gr_gOBiV35
CitedBy_id crossref_primary_10_1121_10_0025536
crossref_primary_10_1121_10_0024632
ContentType Journal Article
DBID 7SW
BJH
BNH
BNI
BNJ
BNO
ERI
PET
REK
WWN
CGR
CUY
CVF
ECM
EIF
NPM
DOI 10.1044/2023_JSLHR-22-00343
DatabaseName ERIC
ERIC (Ovid)
ERIC
ERIC
ERIC (Legacy Platform)
ERIC( SilverPlatter )
ERIC
ERIC PlusText (Legacy Platform)
Education Resources Information Center (ERIC)
ERIC
Medline
MEDLINE
MEDLINE (Ovid)
MEDLINE
MEDLINE
PubMed
DatabaseTitle ERIC
MEDLINE
Medline Complete
MEDLINE with Full Text
PubMed
MEDLINE (Ovid)
DatabaseTitleList MEDLINE
ERIC
Database_xml – sequence: 1
  dbid: BNH
  name: ERIC
  url: http://search.epnet.com/
  sourceTypes: Index Database
DeliveryMethod no_fulltext_linktorsrc
Discipline Medicine
Languages & Literatures
Social Welfare & Social Work
EISSN 1558-9102
ERIC EJ1396795
ExternalDocumentID 37319018
EJ1396795
Genre Research Support, U.S. Gov't, Non-P.H.S
Journal Article
Research Support, N.I.H., Extramural
GrantInformation_xml – fundername: NICHD NIH HHS
  grantid: R01 HD051698
– fundername: NIDCD NIH HHS
  grantid: R15 DC019775
– fundername: NIDCD NIH HHS
  grantid: R01 DC017476
GroupedDBID ---
--Z
-W8
-~X
.GJ
.GO
0-V
04C
0R~
186
18M
1HT
29L
36B
3EH
3V.
4.4
53G
5GY
6NX
6PF
7RV
7SW
7X7
85S
88E
88I
8A4
8AF
8FI
8FJ
8G5
8R4
8R5
AAHSB
AAWTL
AAYRB
ABDBF
ABIVO
ABOPQ
ABPPZ
ABTAH
ABUWG
ABWJO
ABZEH
ACGFO
ACGOD
ACHQT
ACNCT
ACUXI
ADBBV
ADOJX
AENEX
AERSA
AFKRA
AGHSJ
AHMBA
AI.
AIKWM
ALIPV
ALMA_UNASSIGNED_HOLDINGS
ALSLI
ARALO
AZQEC
B-7
BCR
BENPR
BJH
BKEYQ
BLC
BMSDO
BNH
BNI
BNJ
BNO
BPHCQ
BVXVI
CCPQU
CJNVE
CPGLG
CRLPW
CS3
DU5
DWQXO
EAD
EAP
EAS
EBD
EBO
EBS
ECE
ECF
ECT
EDJ
EIHBH
EJD
EMB
EMK
EMOBN
ERI
ESX
EX3
F5P
F9R
FJW
FYUFA
G8K
GNUQQ
GUQSH
H13
HCIFZ
HMCUK
HZ~
H~9
I-F
IAO
ICO
IEA
IER
IHR
IHW
IN-
INH
INIJC
INR
IOF
IPO
IPY
ITC
M0P
M1P
M2M
M2O
M2P
M2Q
M2R
MLAFT
MVM
NAPCQ
O9-
OHT
P-O
P2P
PADUT
PCD
PEA
PET
PQEDU
PQQKQ
PROAC
PSQYO
PSYQQ
Q2X
QF4
QM7
QN7
QO4
QO5
REK
RWL
S0X
S10
SJA
SV3
TAE
TH9
TN5
TUS
TWZ
UHB
UKHRP
UPT
VH1
VJK
VQA
WH7
WOW
WQ9
WWN
YCJ
YQT
ZCA
ZCG
ZHY
ZXP
ZY4
CGR
CUY
CVF
ECM
EIF
NPM
ID FETCH-LOGICAL-c372t-2acb900bae552eae2f3d9566a997e56f71addab0fe3201427b440af46f3c05a52
ISSN 1092-4388
IngestDate Sat Nov 02 12:31:35 EDT 2024
Fri Sep 06 12:14:24 EDT 2024
IsDoiOpenAccess false
IsOpenAccess true
IsPeerReviewed true
IsScholarly true
Issue 6
Language English
LinkModel OpenURL
MergedId FETCHMERGED-LOGICAL-c372t-2acb900bae552eae2f3d9566a997e56f71addab0fe3201427b440af46f3c05a52
ORCID 0000-0003-0955-9495
0000-0003-1835-0976
0000-0001-9971-6321
0000-0003-2230-2897
OpenAccessLink https://osf.io/hzwvf/download
PMID 37319018
PageCount 24
ParticipantIDs pubmed_primary_37319018
eric_primary_EJ1396795
PublicationCentury 2000
PublicationDate 2023-06-20
PublicationDateYYYYMMDD 2023-06-20
PublicationDate_xml – month: 06
  year: 2023
  text: 2023-06-20
  day: 20
PublicationDecade 2020
PublicationPlace United States
PublicationPlace_xml – name: United States
PublicationTitle Journal of speech, language, and hearing research
PublicationTitleAlternate J Speech Lang Hear Res
PublicationYear 2023
Publisher American Speech-Language-Hearing Association
Publisher_xml – name: American Speech-Language-Hearing Association
SSID ssj0000146
Score 2.467747
Snippet Background: Publicly available speech corpora facilitate reproducible research by providing open-access data for participants who have consented/assented to...
Publicly available speech corpora facilitate reproducible research by providing open-access data for participants who have consented/assented to data sharing...
SourceID pubmed
eric
SourceType Index Database
StartPage 1986
SubjectTerms Adolescent
Allied Health Personnel
Articulation (Speech)
Artificial Intelligence
Audio Equipment
Child
Clinical Diagnosis
Humans
Phonetics
Reproducibility of Results
Speech
Speech Disorders
Speech Language Pathology
Speech Sound Disorder
Technology Uses in Education
Training
Title Reproducible Speech Research with the Artificial Intelligence--Ready PERCEPT Corpora
URI http://eric.ed.gov/ERICWebPortal/detail?accno=EJ1396795
https://www.ncbi.nlm.nih.gov/pubmed/37319018
Volume 66
hasFullText
inHoldings 1
isFullTextHit
isPrint
link http://sdu.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwtV3db9MwELdakBgvExQ2NgbyA-KlskhtJ2ketxIUTd00rUXAU2UnjjYxZVPLh_jvuYs_EmCT4IEXK3Iby7r76Xy-3P2OkFd4i6gSWTKTiAqjVYplXHOmSqPhNDaxaimFikV6-nH6Npf5YOC7ZHVz_1XTMAe6xsrZf9B2WBQm4Bl0DiNoHca_0jt41C2J6yWWRC1ujCkvQnqdi7piMcm6zRGyTBsdKSdjmFT_Y3yWn8_ys-XY8Rzf4cFu2tVRST7q6XNBsU02xiAclVAIOR-Z5ruyVv2yUV2yIiaCuDx-H84f96LigKxgt69ULxPA94X89M2B3MUvOPaSYDzqEOc_TFmJsLnbMCvcTn-HqbXTUYbFXrYjoDfktn2LA2zfKk8yS7f9x3ERSYkVMbCr1fFiXpwzuJkjY4_oTseQs5gfg7OcpFk8JENwq_D-f1r0mMpsIZvflye4kvLNLctvka2wXEi0_-Va07o3y0dk22mVHlpAPSYD04zIrhfThr6m80DAvRmRBycuJ2NEDmyFN_1grmq1NvBPP3G9_vyELPuApFb81AOSIiApAJJ2gKS3AJI6QFIHyKfk_bt8OSuYa-XBSpHyL4yrUmdRpJWJY26U4bWo4GaeqCxLTZzU6QTOWaWj2gjwSCVPtZSRqmVSizKKVcx3yL3mujHPCNWCZ2qq5bSqY1lNhE5Lro0uq6SSoOhoj-ygPFc3lq1lFQS9R3athMNPIhXoD0_373rlOXnYQfaA3K_BSJgXZLipvr5s9Q9jPjv5CSDQh1o
link.rule.ids 782
linkProvider EBSCOhost
openUrl ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=Reproducible+Speech+Research+with+the+Artificial+Intelligence--Ready+PERCEPT+Corpora&rft.jtitle=Journal+of+speech%2C+language%2C+and+hearing+research&rft.au=Benway%2C+Nina+R&rft.au=Preston%2C+Jonathan+L&rft.au=Hitchcock%2C+Elaine&rft.au=Rose%2C+Yvan&rft.date=2023-06-20&rft.pub=American+Speech-Language-Hearing+Association&rft.issn=1092-4388&rft.volume=66&rft.issue=6&rft.spage=1986&rft_id=info:doi/10.1044%2F2023_JSLHR-22-00343&rft.externalDocID=EJ1396795
thumbnail_l http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/lc.gif&issn=1092-4388&client=summon
thumbnail_m http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/mc.gif&issn=1092-4388&client=summon
thumbnail_s http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/sc.gif&issn=1092-4388&client=summon