Speech Enhancement Based on Multi-Task Adaptive Knowledge Distillation

Purposes In order to solve the computational cost problem of complex model in time and hardware, and improve the performance of speech enhancement algorithm, a speech enhancement algorithm using multi-task adaptive knowledge distillation is proposed. Methods First, the idea of knowledge distillation...

Full description

Saved in:
Bibliographic Details
Published in:Taiyuan li gong da xue xue bao = Journal of Taiyuan University of Technology Vol. 55; no. 4; pp. 720 - 726
Main Authors: ZHANG Gangmin, LI Yarong, JIA Hairong, WANG Xianxia, DUAN Shufei
Format: Journal Article
Language:English
Published: Editorial Office of Journal of Taiyuan University of Technology 01-07-2024
Subjects:
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Abstract Purposes In order to solve the computational cost problem of complex model in time and hardware, and improve the performance of speech enhancement algorithm, a speech enhancement algorithm using multi-task adaptive knowledge distillation is proposed. Methods First, the idea of knowledge distillation is adopted to solve the problems that the existing speech enhancement model is too large, has many parameters, and has high calculation cost. Second, the differences between different time-frequency units are fully considered, and the weighting factor is introduced to optimize the traditional loss function to improve the network performance of students. In order to avoid the uncertainty of teacher network prediction affecting the performance of student network, the knowledge distillation network of multi-task adaptive learning is built to better utilize the correlation between different tasks to optimize the model. Findings The simulation results show that the proposed algorithm can effectively improve the performance of speech enhancement model while reducing the number of parameters and shortening the calculation time.
AbstractList Purposes In order to solve the computational cost problem of complex model in time and hardware, and improve the performance of speech enhancement algorithm, a speech enhancement algorithm using multi-task adaptive knowledge distillation is proposed. Methods First, the idea of knowledge distillation is adopted to solve the problems that the existing speech enhancement model is too large, has many parameters, and has high calculation cost. Second, the differences between different time-frequency units are fully considered, and the weighting factor is introduced to optimize the traditional loss function to improve the network performance of students. In order to avoid the uncertainty of teacher network prediction affecting the performance of student network, the knowledge distillation network of multi-task adaptive learning is built to better utilize the correlation between different tasks to optimize the model. Findings The simulation results show that the proposed algorithm can effectively improve the performance of speech enhancement model while reducing the number of parameters and shortening the calculation time.
Author ZHANG Gangmin
LI Yarong
WANG Xianxia
DUAN Shufei
JIA Hairong
Author_xml – sequence: 1
  fullname: ZHANG Gangmin
  organization: College of Electronic Information and Optical Engineering, Taiyuan University of Technology, Jinzhong 030600, China
– sequence: 2
  fullname: LI Yarong
  organization: College of Electronic Information and Optical Engineering, Taiyuan University of Technology, Jinzhong 030600, China
– sequence: 3
  fullname: JIA Hairong
  organization: College of Electronic Information and Optical Engineering, Taiyuan University of Technology, Jinzhong 030600, China
– sequence: 4
  fullname: WANG Xianxia
  organization: College of Mathematics, Taiyuan University of Technology, Jinzhong 030600, China
– sequence: 5
  fullname: DUAN Shufei
  organization: College of Electronic Information and Optical Engineering, Taiyuan University of Technology, Jinzhong 030600, China
BookMark eNqtzM1OAjEYheEuIBGQe2jcz9hO20GW8heJcSX75pv2AzqWdjItGu5eQohX4Nm8ybM4YzIIMSAhT5yVvBZKPbdlvpxzyRmbFXMpqrJilWCVmg_I6A8fyDSlll2nXq6mRmTz2SGaI12HIwSDJwyZLiChpTHQj7PPrthB-qKvFrrsvpG-h_jj0R6QrlzKznvILoZHMtyDTzi9d0K2m_Vu-VbYCK3ueneC_qIjOH2D2B809NkZj9oC23MzaxphuEQmwVhprZGGN3VTq1r859cvNepgkg
ContentType Journal Article
DBID DOA
DOI 10.16355/j.tyut.1007-9432.20230259
DatabaseName Directory of Open Access Journals
DatabaseTitleList
Database_xml – sequence: 1
  dbid: DOA
  name: Directory of Open Access Journals
  url: http://www.doaj.org/
  sourceTypes: Open Website
DeliveryMethod fulltext_linktorsrc
Discipline Engineering
EndPage 726
ExternalDocumentID oai_doaj_org_article_da0f1c7bb3c14e04acd4ddc4c1b6b656
GroupedDBID -03
5XA
5XD
92E
92I
ACGFS
ALMA_UNASSIGNED_HOLDINGS
CCEZO
CEKLB
CW9
GROUPED_DOAJ
TCJ
TGP
U1G
U5M
ID FETCH-doaj_primary_oai_doaj_org_article_da0f1c7bb3c14e04acd4ddc4c1b6b6563
IEDL.DBID DOA
ISSN 1007-9432
IngestDate Tue Oct 22 15:14:57 EDT 2024
IsDoiOpenAccess true
IsOpenAccess true
IsPeerReviewed false
IsScholarly true
Issue 4
Language English
LinkModel DirectLink
MergedId FETCHMERGED-doaj_primary_oai_doaj_org_article_da0f1c7bb3c14e04acd4ddc4c1b6b6563
OpenAccessLink https://doaj.org/article/da0f1c7bb3c14e04acd4ddc4c1b6b656
ParticipantIDs doaj_primary_oai_doaj_org_article_da0f1c7bb3c14e04acd4ddc4c1b6b656
PublicationCentury 2000
PublicationDate 2024-07-01
PublicationDateYYYYMMDD 2024-07-01
PublicationDate_xml – month: 07
  year: 2024
  text: 2024-07-01
  day: 01
PublicationDecade 2020
PublicationTitle Taiyuan li gong da xue xue bao = Journal of Taiyuan University of Technology
PublicationYear 2024
Publisher Editorial Office of Journal of Taiyuan University of Technology
Publisher_xml – name: Editorial Office of Journal of Taiyuan University of Technology
SSID ssj0000580075
Score 4.624763
Snippet Purposes In order to solve the computational cost problem of complex model in time and hardware, and improve the performance of speech enhancement algorithm, a...
SourceID doaj
SourceType Open Website
StartPage 720
SubjectTerms knowledge distillation
multi-task adaptive learning
speech enhancement
weighted loss function
Title Speech Enhancement Based on Multi-Task Adaptive Knowledge Distillation
URI https://doaj.org/article/da0f1c7bb3c14e04acd4ddc4c1b6b656
Volume 55
hasFullText 1
inHoldings 1
isFullTextHit
isPrint
link http://sdu.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwrZ07T8MwEMdP0AkGxFO85YE1adI4D48tTVSExNIObJHtc1QBSiOaDP32-BGqMjHAmkiJfSff_2ydfwfwkAqeZokKdeYWM49WppF7FqOnlNQCV3FUljM7m6cvr9k0N5icbasvUxPm8MDOcEPkQRXKVIhIhlQFlEukiJLKUCRCJyM2-gbZzmbKUb2zwFF2TRWAx2g06omjRmCHb3676Vp_-843TcS1-rMf5H4rMcUxHPW5IRm7MZ3AnqpP4XCHGHgGxbzR81uSvF4ad5mjPTLRQoRkVRN7mdZb8PU7GSNvTCAjz99nZmRqVvOHK307h6ciXzzOPDOMsnHEidIwoO0DbZmyt0z5m2WiCxjUq1pdAuHxKGVVwFFWSBPJGIo00rmH0GLNZBReweTv_7v-j4_cwIH2BXUFsbcwaD87dQf7a-zurYe_AFohsM8
link.rule.ids 315,782,786,866,2108,27935,27936
linkProvider Directory of Open Access Journals
openUrl ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=Speech+Enhancement+Based+on+Multi-Task+Adaptive+Knowledge+Distillation&rft.jtitle=Taiyuan+li+gong+da+xue+xue+bao+%3D+Journal+of+Taiyuan+University+of+Technology&rft.au=ZHANG+Gangmin&rft.au=LI+Yarong&rft.au=JIA+Hairong&rft.au=WANG+Xianxia&rft.date=2024-07-01&rft.pub=Editorial+Office+of+Journal+of+Taiyuan+University+of+Technology&rft.issn=1007-9432&rft.volume=55&rft.issue=4&rft.spage=720&rft.epage=726&rft_id=info:doi/10.16355%2Fj.tyut.1007-9432.20230259&rft.externalDBID=DOA&rft.externalDocID=oai_doaj_org_article_da0f1c7bb3c14e04acd4ddc4c1b6b656
thumbnail_l http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/lc.gif&issn=1007-9432&client=summon
thumbnail_m http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/mc.gif&issn=1007-9432&client=summon
thumbnail_s http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/sc.gif&issn=1007-9432&client=summon