Speech Enhancement Based on Multi-Task Adaptive Knowledge Distillation
Purposes In order to solve the computational cost problem of complex model in time and hardware, and improve the performance of speech enhancement algorithm, a speech enhancement algorithm using multi-task adaptive knowledge distillation is proposed. Methods First, the idea of knowledge distillation...
Saved in:
Published in: | Taiyuan li gong da xue xue bao = Journal of Taiyuan University of Technology Vol. 55; no. 4; pp. 720 - 726 |
---|---|
Main Authors: | , , , , |
Format: | Journal Article |
Language: | English |
Published: |
Editorial Office of Journal of Taiyuan University of Technology
01-07-2024
|
Subjects: | |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Abstract | Purposes In order to solve the computational cost problem of complex model in time and hardware, and improve the performance of speech enhancement algorithm, a speech enhancement algorithm using multi-task adaptive knowledge distillation is proposed. Methods First, the idea of knowledge distillation is adopted to solve the problems that the existing speech enhancement model is too large, has many parameters, and has high calculation cost. Second, the differences between different time-frequency units are fully considered, and the weighting factor is introduced to optimize the traditional loss function to improve the network performance of students. In order to avoid the uncertainty of teacher network prediction affecting the performance of student network, the knowledge distillation network of multi-task adaptive learning is built to better utilize the correlation between different tasks to optimize the model. Findings The simulation results show that the proposed algorithm can effectively improve the performance of speech enhancement model while reducing the number of parameters and shortening the calculation time. |
---|---|
AbstractList | Purposes In order to solve the computational cost problem of complex model in time and hardware, and improve the performance of speech enhancement algorithm, a speech enhancement algorithm using multi-task adaptive knowledge distillation is proposed. Methods First, the idea of knowledge distillation is adopted to solve the problems that the existing speech enhancement model is too large, has many parameters, and has high calculation cost. Second, the differences between different time-frequency units are fully considered, and the weighting factor is introduced to optimize the traditional loss function to improve the network performance of students. In order to avoid the uncertainty of teacher network prediction affecting the performance of student network, the knowledge distillation network of multi-task adaptive learning is built to better utilize the correlation between different tasks to optimize the model. Findings The simulation results show that the proposed algorithm can effectively improve the performance of speech enhancement model while reducing the number of parameters and shortening the calculation time. |
Author | ZHANG Gangmin LI Yarong WANG Xianxia DUAN Shufei JIA Hairong |
Author_xml | – sequence: 1 fullname: ZHANG Gangmin organization: College of Electronic Information and Optical Engineering, Taiyuan University of Technology, Jinzhong 030600, China – sequence: 2 fullname: LI Yarong organization: College of Electronic Information and Optical Engineering, Taiyuan University of Technology, Jinzhong 030600, China – sequence: 3 fullname: JIA Hairong organization: College of Electronic Information and Optical Engineering, Taiyuan University of Technology, Jinzhong 030600, China – sequence: 4 fullname: WANG Xianxia organization: College of Mathematics, Taiyuan University of Technology, Jinzhong 030600, China – sequence: 5 fullname: DUAN Shufei organization: College of Electronic Information and Optical Engineering, Taiyuan University of Technology, Jinzhong 030600, China |
BookMark | eNqtzM1OAjEYheEuIBGQe2jcz9hO20GW8heJcSX75pv2AzqWdjItGu5eQohX4Nm8ybM4YzIIMSAhT5yVvBZKPbdlvpxzyRmbFXMpqrJilWCVmg_I6A8fyDSlll2nXq6mRmTz2SGaI12HIwSDJwyZLiChpTHQj7PPrthB-qKvFrrsvpG-h_jj0R6QrlzKznvILoZHMtyDTzi9d0K2m_Vu-VbYCK3ueneC_qIjOH2D2B809NkZj9oC23MzaxphuEQmwVhprZGGN3VTq1r859cvNepgkg |
ContentType | Journal Article |
DBID | DOA |
DOI | 10.16355/j.tyut.1007-9432.20230259 |
DatabaseName | Directory of Open Access Journals |
DatabaseTitleList | |
Database_xml | – sequence: 1 dbid: DOA name: Directory of Open Access Journals url: http://www.doaj.org/ sourceTypes: Open Website |
DeliveryMethod | fulltext_linktorsrc |
Discipline | Engineering |
EndPage | 726 |
ExternalDocumentID | oai_doaj_org_article_da0f1c7bb3c14e04acd4ddc4c1b6b656 |
GroupedDBID | -03 5XA 5XD 92E 92I ACGFS ALMA_UNASSIGNED_HOLDINGS CCEZO CEKLB CW9 GROUPED_DOAJ TCJ TGP U1G U5M |
ID | FETCH-doaj_primary_oai_doaj_org_article_da0f1c7bb3c14e04acd4ddc4c1b6b6563 |
IEDL.DBID | DOA |
ISSN | 1007-9432 |
IngestDate | Tue Oct 22 15:14:57 EDT 2024 |
IsDoiOpenAccess | true |
IsOpenAccess | true |
IsPeerReviewed | false |
IsScholarly | true |
Issue | 4 |
Language | English |
LinkModel | DirectLink |
MergedId | FETCHMERGED-doaj_primary_oai_doaj_org_article_da0f1c7bb3c14e04acd4ddc4c1b6b6563 |
OpenAccessLink | https://doaj.org/article/da0f1c7bb3c14e04acd4ddc4c1b6b656 |
ParticipantIDs | doaj_primary_oai_doaj_org_article_da0f1c7bb3c14e04acd4ddc4c1b6b656 |
PublicationCentury | 2000 |
PublicationDate | 2024-07-01 |
PublicationDateYYYYMMDD | 2024-07-01 |
PublicationDate_xml | – month: 07 year: 2024 text: 2024-07-01 day: 01 |
PublicationDecade | 2020 |
PublicationTitle | Taiyuan li gong da xue xue bao = Journal of Taiyuan University of Technology |
PublicationYear | 2024 |
Publisher | Editorial Office of Journal of Taiyuan University of Technology |
Publisher_xml | – name: Editorial Office of Journal of Taiyuan University of Technology |
SSID | ssj0000580075 |
Score | 4.624763 |
Snippet | Purposes In order to solve the computational cost problem of complex model in time and hardware, and improve the performance of speech enhancement algorithm, a... |
SourceID | doaj |
SourceType | Open Website |
StartPage | 720 |
SubjectTerms | knowledge distillation multi-task adaptive learning speech enhancement weighted loss function |
Title | Speech Enhancement Based on Multi-Task Adaptive Knowledge Distillation |
URI | https://doaj.org/article/da0f1c7bb3c14e04acd4ddc4c1b6b656 |
Volume | 55 |
hasFullText | 1 |
inHoldings | 1 |
isFullTextHit | |
isPrint | |
link | http://sdu.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwrZ07T8MwEMdP0AkGxFO85YE1adI4D48tTVSExNIObJHtc1QBSiOaDP32-BGqMjHAmkiJfSff_2ydfwfwkAqeZokKdeYWM49WppF7FqOnlNQCV3FUljM7m6cvr9k0N5icbasvUxPm8MDOcEPkQRXKVIhIhlQFlEukiJLKUCRCJyM2-gbZzmbKUb2zwFF2TRWAx2g06omjRmCHb3676Vp_-843TcS1-rMf5H4rMcUxHPW5IRm7MZ3AnqpP4XCHGHgGxbzR81uSvF4ad5mjPTLRQoRkVRN7mdZb8PU7GSNvTCAjz99nZmRqVvOHK307h6ciXzzOPDOMsnHEidIwoO0DbZmyt0z5m2WiCxjUq1pdAuHxKGVVwFFWSBPJGIo00rmH0GLNZBReweTv_7v-j4_cwIH2BXUFsbcwaD87dQf7a-zurYe_AFohsM8 |
link.rule.ids | 315,782,786,866,2108,27935,27936 |
linkProvider | Directory of Open Access Journals |
openUrl | ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=Speech+Enhancement+Based+on+Multi-Task+Adaptive+Knowledge+Distillation&rft.jtitle=Taiyuan+li+gong+da+xue+xue+bao+%3D+Journal+of+Taiyuan+University+of+Technology&rft.au=ZHANG+Gangmin&rft.au=LI+Yarong&rft.au=JIA+Hairong&rft.au=WANG+Xianxia&rft.date=2024-07-01&rft.pub=Editorial+Office+of+Journal+of+Taiyuan+University+of+Technology&rft.issn=1007-9432&rft.volume=55&rft.issue=4&rft.spage=720&rft.epage=726&rft_id=info:doi/10.16355%2Fj.tyut.1007-9432.20230259&rft.externalDBID=DOA&rft.externalDocID=oai_doaj_org_article_da0f1c7bb3c14e04acd4ddc4c1b6b656 |
thumbnail_l | http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/lc.gif&issn=1007-9432&client=summon |
thumbnail_m | http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/mc.gif&issn=1007-9432&client=summon |
thumbnail_s | http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/sc.gif&issn=1007-9432&client=summon |