Speaker Adaptation on Articulation and Acoustics for Articulation-to-Speech Synthesis

Silent speech interfaces (SSIs) convert non-audio bio-signals, such as articulatory movement, to speech. This technology has the potential to recover the speech ability of individuals who have lost their voice but can still articulate (e.g., laryngectomees). Articulation-to-speech (ATS) synthesis is...

Celý popis

Uložené v:
Podrobná bibliografia
Vydané v:Sensors (Basel, Switzerland) Ročník 22; číslo 16; s. 6056
Hlavní autori: Cao, Beiming, Wisler, Alan, Wang, Jun
Médium: Journal Article
Jazyk:English
Vydavateľské údaje: Switzerland MDPI AG 13.08.2022
MDPI
Predmet:
ISSN:1424-8220, 1424-8220
On-line prístup:Získať plný text
Tagy: Pridať tag
Žiadne tagy, Buďte prvý, kto otaguje tento záznam!
Abstract Silent speech interfaces (SSIs) convert non-audio bio-signals, such as articulatory movement, to speech. This technology has the potential to recover the speech ability of individuals who have lost their voice but can still articulate (e.g., laryngectomees). Articulation-to-speech (ATS) synthesis is an algorithm design of SSI that has the advantages of easy-implementation and low-latency, and therefore is becoming more popular. Current ATS studies focus on speaker-dependent (SD) models to avoid large variations of articulatory patterns and acoustic features across speakers. However, these designs are limited by the small data size from individual speakers. Speaker adaptation designs that include multiple speakers’ data have the potential to address the issue of limited data size from single speakers; however, few prior studies have investigated their performance in ATS. In this paper, we investigated speaker adaptation on both the input articulation and the output acoustic signals (with or without direct inclusion of data from test speakers) using the publicly available electromagnetic articulatory (EMA) dataset. We used Procrustes matching and voice conversion for articulation and voice adaptation, respectively. The performance of the ATS models was measured objectively by the mel-cepstral distortions (MCDs). The synthetic speech samples were generated and are provided in the supplementary material. The results demonstrated the improvement brought by both Procrustes matching and voice conversion on speaker-independent ATS. With the direct inclusion of target speaker data in the training process, the speaker-adaptive ATS achieved a comparable performance to speaker-dependent ATS. To our knowledge, this is the first study that has demonstrated that speaker-adaptive ATS can achieve a non-statistically different performance to speaker-dependent ATS.
AbstractList Silent speech interfaces (SSIs) convert non-audio bio-signals, such as articulatory movement, to speech. This technology has the potential to recover the speech ability of individuals who have lost their voice but can still articulate (e.g., laryngectomees). Articulation-to-speech (ATS) synthesis is an algorithm design of SSI that has the advantages of easy-implementation and low-latency, and therefore is becoming more popular. Current ATS studies focus on speaker-dependent (SD) models to avoid large variations of articulatory patterns and acoustic features across speakers. However, these designs are limited by the small data size from individual speakers. Speaker adaptation designs that include multiple speakers’ data have the potential to address the issue of limited data size from single speakers; however, few prior studies have investigated their performance in ATS. In this paper, we investigated speaker adaptation on both the input articulation and the output acoustic signals (with or without direct inclusion of data from test speakers) using the publicly available electromagnetic articulatory (EMA) dataset. We used Procrustes matching and voice conversion for articulation and voice adaptation, respectively. The performance of the ATS models was measured objectively by the mel-cepstral distortions (MCDs). The synthetic speech samples were generated and are provided in the supplementary material. The results demonstrated the improvement brought by both Procrustes matching and voice conversion on speaker-independent ATS. With the direct inclusion of target speaker data in the training process, the speaker-adaptive ATS achieved a comparable performance to speaker-dependent ATS. To our knowledge, this is the first study that has demonstrated that speaker-adaptive ATS can achieve a non-statistically different performance to speaker-dependent ATS.
Silent speech interfaces (SSIs) convert non-audio bio-signals, such as articulatory movement, to speech. This technology has the potential to recover the speech ability of individuals who have lost their voice but can still articulate (e.g., laryngectomees). Articulation-to-speech (ATS) synthesis is an algorithm design of SSI that has the advantages of easy-implementation and low-latency, and therefore is becoming more popular. Current ATS studies focus on speaker-dependent (SD) models to avoid large variations of articulatory patterns and acoustic features across speakers. However, these designs are limited by the small data size from individual speakers. Speaker adaptation designs that include multiple speakers' data have the potential to address the issue of limited data size from single speakers; however, few prior studies have investigated their performance in ATS. In this paper, we investigated speaker adaptation on both the input articulation and the output acoustic signals (with or without direct inclusion of data from test speakers) using the publicly available electromagnetic articulatory (EMA) dataset. We used Procrustes matching and voice conversion for articulation and voice adaptation, respectively. The performance of the ATS models was measured objectively by the mel-cepstral distortions (MCDs). The synthetic speech samples were generated and are provided in the supplementary material. The results demonstrated the improvement brought by both Procrustes matching and voice conversion on speaker-independent ATS. With the direct inclusion of target speaker data in the training process, the speaker-adaptive ATS achieved a comparable performance to speaker-dependent ATS. To our knowledge, this is the first study that has demonstrated that speaker-adaptive ATS can achieve a non-statistically different performance to speaker-dependent ATS.Silent speech interfaces (SSIs) convert non-audio bio-signals, such as articulatory movement, to speech. This technology has the potential to recover the speech ability of individuals who have lost their voice but can still articulate (e.g., laryngectomees). Articulation-to-speech (ATS) synthesis is an algorithm design of SSI that has the advantages of easy-implementation and low-latency, and therefore is becoming more popular. Current ATS studies focus on speaker-dependent (SD) models to avoid large variations of articulatory patterns and acoustic features across speakers. However, these designs are limited by the small data size from individual speakers. Speaker adaptation designs that include multiple speakers' data have the potential to address the issue of limited data size from single speakers; however, few prior studies have investigated their performance in ATS. In this paper, we investigated speaker adaptation on both the input articulation and the output acoustic signals (with or without direct inclusion of data from test speakers) using the publicly available electromagnetic articulatory (EMA) dataset. We used Procrustes matching and voice conversion for articulation and voice adaptation, respectively. The performance of the ATS models was measured objectively by the mel-cepstral distortions (MCDs). The synthetic speech samples were generated and are provided in the supplementary material. The results demonstrated the improvement brought by both Procrustes matching and voice conversion on speaker-independent ATS. With the direct inclusion of target speaker data in the training process, the speaker-adaptive ATS achieved a comparable performance to speaker-dependent ATS. To our knowledge, this is the first study that has demonstrated that speaker-adaptive ATS can achieve a non-statistically different performance to speaker-dependent ATS.
Audience Academic
Author Cao, Beiming
Wang, Jun
Wisler, Alan
AuthorAffiliation 3 Department of Mathematics and Statistics, Utah State University, Logan, UT 84322, USA
1 Department of Electrical and Computer Engineering, University of Texas at Austin, Austin, TX 78712, USA
2 Department of Speech, Language, and Hearing Sciences, University of Texas at Austin, Austin, TX 78712, USA
4 Department of Neurology, Dell Medical School, University of Texas at Austin, Austin, TX 78712, USA
AuthorAffiliation_xml – name: 3 Department of Mathematics and Statistics, Utah State University, Logan, UT 84322, USA
– name: 2 Department of Speech, Language, and Hearing Sciences, University of Texas at Austin, Austin, TX 78712, USA
– name: 1 Department of Electrical and Computer Engineering, University of Texas at Austin, Austin, TX 78712, USA
– name: 4 Department of Neurology, Dell Medical School, University of Texas at Austin, Austin, TX 78712, USA
Author_xml – sequence: 1
  givenname: Beiming
  orcidid: 0000-0003-1223-3146
  surname: Cao
  fullname: Cao, Beiming
– sequence: 2
  givenname: Alan
  orcidid: 0000-0003-2601-2846
  surname: Wisler
  fullname: Wisler, Alan
– sequence: 3
  givenname: Jun
  orcidid: 0000-0001-7265-217X
  surname: Wang
  fullname: Wang, Jun
BackLink https://www.ncbi.nlm.nih.gov/pubmed/36015817$$D View this record in MEDLINE/PubMed
BookMark eNplkktr3DAQx01JaR7toV-gLPTSHpyMHpbkS8GEPgKBHtKchazHrrZeaSvZgXz7ynUSsikSSBr956eZ0ZxWRyEGW1XvEZwT0sJFxhgxBg17VZ0gimktMIajZ_vj6jTnLQAmhIg31TFhgBqB-El1e7O36rdNq86o_ahGH8OqzC6NXk_DclbBrDodp1xseeViOriux1gXiNWb1c19GDc2-_y2eu3UkO27h_Wsuv329dflj_r65_ery-661g2IsW5a3Gjsem2EIrzHWAMSyLbOMXBEOWWd4MJhrQkz4AQQJoTRjjGLqWI9OauuFq6Jaiv3ye9UupdRefnPENNaqjnUwUrgyhAAy5DtKfRNz0XPrQFmkG65aQrry8LaT_3OGm3DmNRwAD28CX4j1_FOthQxSmkBfHoApPhnsnmUO5-1HQYVbCmexBw4Q0DFLP34QrqNUwqlVLOKUcKxmCM6X1RrVRLwwcXyri7D2J3XpQWcL_aOU9YQBM2M_fA8hafYH7-7CC4WgU4x52Sd1H759EL2g0Qg54aSTw1VPD6_8HiE_q_9C4hIymw
CitedBy_id crossref_primary_10_3390_biomimetics10050275
crossref_primary_10_3390_biomimetics9020080
crossref_primary_10_3390_s22228601
crossref_primary_10_1109_TIM_2024_3449948
crossref_primary_10_1044_2023_JSLHR_22_00319
crossref_primary_10_1109_TNSRE_2023_3342068
crossref_primary_10_3390_su142013631
Cites_doi 10.1109/TASLP.2017.2758999
10.1109/89.222875
10.1177/000348948008900608
10.21437/Interspeech.2018-2484
10.21437/Interspeech.2017-939
10.21437/Interspeech.2021-1466
10.1109/ICASSP.2010.5495552
10.21437/Interspeech.2014-266
10.3390/s22020649
10.1109/TASL.2007.907344
10.1109/ICASSP.2019.8683143
10.1109/TASLP.2017.2752365
10.1109/TASLP.2017.2757263
10.1145/3172944.3172977
10.1587/transinf.2015EDP7457
10.1007/BF00453713
10.1109/TASL.2009.2016394
10.2147/MDER.S133225
10.21437/Interspeech.2015-522
10.1109/ICASSP.2007.367298
10.1016/j.specom.2017.01.008
10.1250/ast.27.349
10.21437/Interspeech.2011-316
10.1002/hed.24353
10.1109/SLT48900.2021.9383619
10.1016/j.specom.2009.08.002
10.18653/v1/W15-5109
10.21437/Interspeech.2020-1031
10.21437/Interspeech.2020-1630
10.1109/ICASSP.2014.6855102
10.1109/JSEN.2020.3046469
10.1590/S1807-59322005000200010
10.1587/transinf.2019EDP7211
10.1007/978-3-540-74048-3
10.1109/ICASSP.2013.6639215
10.1002/ecja.4400660203
10.1109/TBME.2017.2654361
10.1016/j.anl.2006.11.010
10.21437/Interspeech.2021-1842
10.1109/ACCESS.2020.3026579
10.3115/1075671.1075688
10.1109/BIOCAS.2018.8584786
10.1109/TAU.1969.1162058
10.1016/0093-934X(87)90058-7
10.1016/j.specom.2009.11.004
10.1121/1.4987629
10.1007/BF02291478
ContentType Journal Article
Copyright COPYRIGHT 2022 MDPI AG
2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
2022 by the authors. 2022
Copyright_xml – notice: COPYRIGHT 2022 MDPI AG
– notice: 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
– notice: 2022 by the authors. 2022
DBID AAYXX
CITATION
CGR
CUY
CVF
ECM
EIF
NPM
3V.
7X7
7XB
88E
8FI
8FJ
8FK
ABUWG
AFKRA
AZQEC
BENPR
CCPQU
DWQXO
FYUFA
GHDGH
K9.
M0S
M1P
PHGZM
PHGZT
PIMPY
PJZUB
PKEHL
PPXIY
PQEST
PQQKQ
PQUKI
PRINS
7X8
5PM
DOA
DOI 10.3390/s22166056
DatabaseName CrossRef
Medline
MEDLINE
MEDLINE (Ovid)
MEDLINE
MEDLINE
PubMed
ProQuest Central (Corporate)
Health & Medical Collection
ProQuest Central (purchase pre-March 2016)
Medical Database (Alumni Edition)
Hospital Premium Collection
Hospital Premium Collection (Alumni Edition)
ProQuest Central (Alumni) (purchase pre-March 2016)
ProQuest Central (Alumni)
ProQuest Central UK/Ireland
ProQuest Central Essentials
ProQuest Central
ProQuest One Community College
ProQuest Central
Health Research Premium Collection
Health Research Premium Collection (Alumni)
ProQuest Health & Medical Complete (Alumni)
Health & Medical Collection (Alumni Edition)
PML(ProQuest Medical Library)
ProQuest Central Premium
ProQuest One Academic
Publicly Available Content Database
ProQuest Health & Medical Research Collection
ProQuest One Academic Middle East (New)
One Health & Nursing
ProQuest One Academic Eastern Edition (DO NOT USE)
ProQuest One Academic (retired)
ProQuest One Academic UKI Edition
ProQuest Central China
MEDLINE - Academic
PubMed Central (Full Participant titles)
DOAJ Directory of Open Access Journals
DatabaseTitle CrossRef
MEDLINE
Medline Complete
MEDLINE with Full Text
PubMed
MEDLINE (Ovid)
Publicly Available Content Database
ProQuest One Academic Middle East (New)
ProQuest Central Essentials
ProQuest Health & Medical Complete (Alumni)
ProQuest Central (Alumni Edition)
ProQuest One Community College
ProQuest One Health & Nursing
ProQuest Central China
ProQuest Central
ProQuest Health & Medical Research Collection
Health Research Premium Collection
Health and Medicine Complete (Alumni Edition)
ProQuest Central Korea
Health & Medical Research Collection
ProQuest Central (New)
ProQuest Medical Library (Alumni)
ProQuest One Academic Eastern Edition
ProQuest Hospital Collection
Health Research Premium Collection (Alumni)
ProQuest Hospital Collection (Alumni)
ProQuest Health & Medical Complete
ProQuest Medical Library
ProQuest One Academic UKI Edition
ProQuest One Academic
ProQuest One Academic (New)
ProQuest Central (Alumni)
MEDLINE - Academic
DatabaseTitleList Publicly Available Content Database


MEDLINE - Academic
MEDLINE

CrossRef
Database_xml – sequence: 1
  dbid: DOA
  name: DOAJ Directory of Open Access Journals
  url: https://www.doaj.org/
  sourceTypes: Open Website
– sequence: 2
  dbid: NPM
  name: PubMed
  url: http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?db=PubMed
  sourceTypes: Index Database
– sequence: 3
  dbid: PIMPY
  name: Publicly Available Content Database
  url: http://search.proquest.com/publiccontent
  sourceTypes: Aggregation Database
DeliveryMethod fulltext_linktorsrc
Discipline Engineering
EISSN 1424-8220
ExternalDocumentID oai_doaj_org_article_07ad300e61eb40b5b78b7ed06d1c97d5
PMC9416444
A746531054
36015817
10_3390_s22166056
Genre Journal Article
GeographicLocations United States
GeographicLocations_xml – name: United States
GrantInformation_xml – fundername: NIH HHS
  grantid: R01DC016621
– fundername: NIDCD NIH HHS
  grantid: R01 DC016621
– fundername: Other Communication Disorders (NIDCD) of the National Institutes of Health (NIH)
  grantid: R01DC016621
– fundername: National Institute on Deafness
GroupedDBID ---
123
2WC
53G
5VS
7X7
88E
8FE
8FG
8FI
8FJ
AADQD
AAHBH
AAYXX
ABDBF
ABUWG
ACUHS
ADBBV
ADMLS
AENEX
AFFHD
AFKRA
AFZYC
ALMA_UNASSIGNED_HOLDINGS
BENPR
BPHCQ
BVXVI
CCPQU
CITATION
CS3
D1I
DU5
E3Z
EBD
ESX
F5P
FYUFA
GROUPED_DOAJ
GX1
HH5
HMCUK
HYE
IAO
ITC
KQ8
L6V
M1P
M48
MODMG
M~E
OK1
OVT
P2P
P62
PHGZM
PHGZT
PIMPY
PJZUB
PPXIY
PQQKQ
PROAC
PSQYO
RNS
RPM
TUS
UKHRP
XSB
~8M
3V.
ABJCF
ALIPV
ARAPS
CGR
CUY
CVF
ECM
EIF
HCIFZ
KB.
M7S
NPM
PDBOC
7XB
8FK
AZQEC
DWQXO
K9.
PKEHL
PQEST
PQUKI
PRINS
7X8
PUEGO
5PM
ID FETCH-LOGICAL-c508t-5925c2fbcd8a37b22c0181e9ff60f3afaef878f2cc36d0f803688dcf66e24a6b3
IEDL.DBID DOA
ISICitedReferencesCount 8
ISICitedReferencesURI http://www.webofscience.com/api/gateway?GWVersion=2&SrcApp=Summon&SrcAuth=ProQuest&DestLinkType=CitingArticles&DestApp=WOS_CPL&KeyUT=000845322000001&url=https%3A%2F%2Fcvtisr.summon.serialssolutions.com%2F%23%21%2Fsearch%3Fho%3Df%26include.ft.matches%3Dt%26l%3Dnull%26q%3D
ISSN 1424-8220
IngestDate Tue Oct 14 18:44:26 EDT 2025
Tue Nov 04 01:52:30 EST 2025
Thu Oct 02 07:16:41 EDT 2025
Tue Oct 07 07:12:16 EDT 2025
Tue Nov 04 18:17:38 EST 2025
Wed Feb 19 02:18:40 EST 2025
Sat Nov 29 07:13:46 EST 2025
Tue Nov 18 22:44:28 EST 2025
IsDoiOpenAccess true
IsOpenAccess true
IsPeerReviewed true
IsScholarly true
Issue 16
Keywords speaker adaption
articulation-to-speech synthesis
silent speech interface
voice conversion
Language English
License Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
LinkModel DirectLink
MergedId FETCHMERGED-LOGICAL-c508t-5925c2fbcd8a37b22c0181e9ff60f3afaef878f2cc36d0f803688dcf66e24a6b3
Notes ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 14
content type line 23
ORCID 0000-0003-1223-3146
0000-0003-2601-2846
0000-0001-7265-217X
OpenAccessLink https://doaj.org/article/07ad300e61eb40b5b78b7ed06d1c97d5
PMID 36015817
PQID 2706437285
PQPubID 2032333
ParticipantIDs doaj_primary_oai_doaj_org_article_07ad300e61eb40b5b78b7ed06d1c97d5
pubmedcentral_primary_oai_pubmedcentral_nih_gov_9416444
proquest_miscellaneous_2707610484
proquest_journals_2706437285
gale_infotracacademiconefile_A746531054
pubmed_primary_36015817
crossref_citationtrail_10_3390_s22166056
crossref_primary_10_3390_s22166056
PublicationCentury 2000
PublicationDate 20220813
PublicationDateYYYYMMDD 2022-08-13
PublicationDate_xml – month: 8
  year: 2022
  text: 20220813
  day: 13
PublicationDecade 2020
PublicationPlace Switzerland
PublicationPlace_xml – name: Switzerland
– name: Basel
PublicationTitle Sensors (Basel, Switzerland)
PublicationTitleAlternate Sensors (Basel)
PublicationYear 2022
Publisher MDPI AG
MDPI
Publisher_xml – name: MDPI AG
– name: MDPI
References Kim (ref_11) 2017; 25
Gonzalez (ref_30) 2017; 25
ref_58
ref_57
ref_12
ref_56
ref_10
ref_54
ref_53
Nijdam (ref_2) 1982; 237
Yamagishi (ref_19) 2009; 17
ref_18
Schultz (ref_8) 2017; 25
ref_17
ref_16
ref_15
Hueber (ref_33) 2010; 52
ref_59
Huang (ref_13) 1993; 1
Liu (ref_4) 2007; 34
ref_23
ref_22
Gower (ref_26) 1975; 40
ref_21
Kawahara (ref_51) 2006; 27
ref_20
Mohammadi (ref_55) 2017; 88
Tiede (ref_25) 2017; 141
ref_29
ref_28
ref_27
Arfib (ref_48) 2002; 9
Kaye (ref_5) 2017; 10
Denby (ref_7) 2010; 52
Singer (ref_3) 1980; 89
Toda (ref_24) 2007; 15
Electrical (ref_43) 1969; 17
Lee (ref_38) 2020; 103
Sebkhi (ref_41) 2020; 21
ref_36
Eadie (ref_6) 2016; 38
ref_35
ref_34
ref_31
Braz (ref_1) 2005; 60
ref_39
ref_37
Gomez (ref_9) 2020; 8
ref_47
ref_46
ref_45
ref_44
Imai (ref_50) 1983; 66
ref_42
Sebkhi (ref_32) 2017; 64
ref_40
Morise (ref_52) 2016; 99
Wenig (ref_14) 1987; 31
ref_49
References_xml – volume: 25
  start-page: 2323
  year: 2017
  ident: ref_11
  article-title: Speaker-Independent Silent Speech Recognition from Flesh-Point Articulatory Movements Using an LSTM Neural Network
  publication-title: IEEE/ACM Trans. Audio Speech Lang. Process. (TASLP)
  doi: 10.1109/TASLP.2017.2758999
– volume: 1
  start-page: 150
  year: 1993
  ident: ref_13
  article-title: On speaker-independent, speaker-dependent, and speaker-adaptive speech recognition
  publication-title: IEEE Trans. Speech Audio Process.
  doi: 10.1109/89.222875
– volume: 89
  start-page: 529
  year: 1980
  ident: ref_3
  article-title: An Endoscopic Technique for Restoration of Voice after Laryngectomy
  publication-title: Ann. Otol. Rhinol. Laryngol.
  doi: 10.1177/000348948008900608
– ident: ref_15
  doi: 10.21437/Interspeech.2018-2484
– ident: ref_18
  doi: 10.21437/Interspeech.2017-939
– ident: ref_20
  doi: 10.21437/Interspeech.2021-1466
– ident: ref_37
  doi: 10.1109/ICASSP.2010.5495552
– ident: ref_35
– ident: ref_23
– ident: ref_16
  doi: 10.21437/Interspeech.2014-266
– ident: ref_40
  doi: 10.3390/s22020649
– ident: ref_58
– volume: 15
  start-page: 2222
  year: 2007
  ident: ref_24
  article-title: Voice Conversion Based on Maximum-Likelihood Estimation of Spectral Parameter Trajectory
  publication-title: IEEE Trans. Audio Speech Lang. Process.
  doi: 10.1109/TASL.2007.907344
– ident: ref_27
– ident: ref_28
  doi: 10.1109/ICASSP.2019.8683143
– volume: 25
  start-page: 2257
  year: 2017
  ident: ref_8
  article-title: Biosignal-based Spoken Communication: A Survey
  publication-title: IEEE/ACM Trans. Audio Speech Lang. Process.
  doi: 10.1109/TASLP.2017.2752365
– volume: 25
  start-page: 2362
  year: 2017
  ident: ref_30
  article-title: Direct Speech Reconstruction from Articulatory Sensor Data by Machine Learning
  publication-title: IEEE/ACM Trans. Audio Speech Lang. Process.
  doi: 10.1109/TASLP.2017.2757263
– ident: ref_39
  doi: 10.1145/3172944.3172977
– volume: 99
  start-page: 1877
  year: 2016
  ident: ref_52
  article-title: WORLD: A vocoder-based high-quality speech synthesis system for real-time applications
  publication-title: IEICE Trans. Inf. Syst.
  doi: 10.1587/transinf.2015EDP7457
– volume: 237
  start-page: 27
  year: 1982
  ident: ref_2
  article-title: A New Prosthesis for Voice Rehabilitation after Laryngectomy
  publication-title: Arch. Oto-Rhino-Laryngol.
  doi: 10.1007/BF00453713
– volume: 17
  start-page: 1208
  year: 2009
  ident: ref_19
  article-title: Robust speaker-adaptive HMM-based text-to-speech synthesis
  publication-title: IEEE Trans. Audio Speech Lang. Process.
  doi: 10.1109/TASL.2009.2016394
– ident: ref_17
– volume: 10
  start-page: 133
  year: 2017
  ident: ref_5
  article-title: The Electrolarynx: Voice Restoration after Total Laryngectomy
  publication-title: Med. Devices
  doi: 10.2147/MDER.S133225
– ident: ref_59
– ident: ref_29
  doi: 10.21437/Interspeech.2015-522
– ident: ref_49
  doi: 10.1109/ICASSP.2007.367298
– volume: 88
  start-page: 65
  year: 2017
  ident: ref_55
  article-title: An Overview of Voice Conversion Systems
  publication-title: Speech Commun.
  doi: 10.1016/j.specom.2017.01.008
– ident: ref_53
– volume: 27
  start-page: 349
  year: 2006
  ident: ref_51
  article-title: STRAIGHT, exploitation of the other aspect of VOCODER: Perceptually isomorphic decomposition of speech sounds
  publication-title: Acoust. Sci. Technol.
  doi: 10.1250/ast.27.349
– ident: ref_44
  doi: 10.21437/Interspeech.2011-316
– volume: 38
  start-page: E1955
  year: 2016
  ident: ref_6
  article-title: The Relationship between Communicative Participation and Postlaryngectomy Speech Outcomes
  publication-title: Head Neck
  doi: 10.1002/hed.24353
– ident: ref_21
  doi: 10.1109/SLT48900.2021.9383619
– volume: 52
  start-page: 270
  year: 2010
  ident: ref_7
  article-title: Silent Speech Interfaces
  publication-title: Speech Commun.
  doi: 10.1016/j.specom.2009.08.002
– ident: ref_57
  doi: 10.18653/v1/W15-5109
– ident: ref_47
– ident: ref_34
  doi: 10.21437/Interspeech.2020-1031
– ident: ref_42
  doi: 10.21437/Interspeech.2020-1630
– ident: ref_45
  doi: 10.1109/ICASSP.2014.6855102
– volume: 21
  start-page: 7964
  year: 2020
  ident: ref_41
  article-title: Inertial Measurements for Tongue Motion Tracking Based on Magnetic Localization With Orientation Compensation
  publication-title: IEEE Sens. J.
  doi: 10.1109/JSEN.2020.3046469
– volume: 60
  start-page: 135
  year: 2005
  ident: ref_1
  article-title: Quality of life and depression in patients undergoing total and partial laryngectomy
  publication-title: Clinics
  doi: 10.1590/S1807-59322005000200010
– volume: 103
  start-page: 1875
  year: 2020
  ident: ref_38
  article-title: Silent speech interface using ultrasonic Doppler sonar
  publication-title: IEICE Trans. Inf. Syst.
  doi: 10.1587/transinf.2019EDP7211
– ident: ref_56
  doi: 10.1007/978-3-540-74048-3
– ident: ref_12
  doi: 10.1109/ICASSP.2013.6639215
– volume: 66
  start-page: 10
  year: 1983
  ident: ref_50
  article-title: Mel log spectrum approximation (MLSA) filter for speech synthesis
  publication-title: Electron. Commun. Jpn. (Part I Commun.)
  doi: 10.1002/ecja.4400660203
– volume: 64
  start-page: 2639
  year: 2017
  ident: ref_32
  article-title: Multimodal Speech Capture System for Speech Rehabilitation and Learning
  publication-title: IEEE Trans. Biomed. Eng.
  doi: 10.1109/TBME.2017.2654361
– volume: 34
  start-page: 327
  year: 2007
  ident: ref_4
  article-title: Electrolarynx in Voice Rehabilitation
  publication-title: Auris Nasus Larynx
  doi: 10.1016/j.anl.2006.11.010
– ident: ref_10
  doi: 10.21437/Interspeech.2021-1842
– volume: 8
  start-page: 177995
  year: 2020
  ident: ref_9
  article-title: Silent Speech Interfaces for Speech Restoration: A Review
  publication-title: IEEE Access
  doi: 10.1109/ACCESS.2020.3026579
– ident: ref_54
– ident: ref_46
– ident: ref_22
  doi: 10.3115/1075671.1075688
– ident: ref_31
  doi: 10.1109/BIOCAS.2018.8584786
– volume: 17
  start-page: 225
  year: 1969
  ident: ref_43
  article-title: IEEE recommended practice for speech quality measurements
  publication-title: IEEE Trans. Audio Electroacoust.
  doi: 10.1109/TAU.1969.1162058
– volume: 31
  start-page: 26
  year: 1987
  ident: ref_14
  article-title: Electromagnetic articulography: Use of alternating magnetic fields for tracking movements of multiple points inside and outside the vocal tract
  publication-title: Brain Lang.
  doi: 10.1016/0093-934X(87)90058-7
– volume: 52
  start-page: 288
  year: 2010
  ident: ref_33
  article-title: Development of a silent speech interface driven by ultrasound and optical images of the tongue and lips
  publication-title: Speech Commun.
  doi: 10.1016/j.specom.2009.11.004
– volume: 141
  start-page: 3580
  year: 2017
  ident: ref_25
  article-title: Quantifying kinematic aspects of reduction in a contrasting rate production task
  publication-title: J. Acoust. Soc. Am.
  doi: 10.1121/1.4987629
– ident: ref_36
– volume: 9
  start-page: 299
  year: 2002
  ident: ref_48
  article-title: Source-filter processing
  publication-title: DAFX–Digital Audio Eff.
– volume: 40
  start-page: 33
  year: 1975
  ident: ref_26
  article-title: Generalized Procrustes Analysis
  publication-title: Psychometrika
  doi: 10.1007/BF02291478
SSID ssj0023338
Score 2.4314144
Snippet Silent speech interfaces (SSIs) convert non-audio bio-signals, such as articulatory movement, to speech. This technology has the potential to recover the...
SourceID doaj
pubmedcentral
proquest
gale
pubmed
crossref
SourceType Open Website
Open Access Repository
Aggregation Database
Index Database
Enrichment Source
StartPage 6056
SubjectTerms Acoustics
Adaptation
articulation-to-speech synthesis
Computer interfaces
Datasets
Esophagus
Humans
Larynx
Sensors
silent speech interface
speaker adaption
Speaking
Speech
Speech Acoustics
Speech Perception
Speech processing systems
Statistical methods
Testing
Tongue
Ultrasonic imaging
Voice
voice conversion
Voice recognition
SummonAdditionalLinks – databaseName: ProQuest Central
  dbid: BENPR
  link: http://cvtisr.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwpV1Lb9QwEB7BlgMceD8CBQWEBJeoiRM_ckJb1IrTqqJU6i3yk65A2WWzIPHvmUm8YSMQF6ScbMcaazwvj_0NwGtP9ZC0NBmpuowAujJjuM3qUGpHFksb3RebkIuFurysz-KBWxevVe50Yq-o3crSGfkRk0OOSfF3628ZVY2i7GosoXEdDgiprJrBwfHJ4uzjGHKVGIENeEIlBvdHHWOFQAdeTKxQD9b_p0res0nT-5J7Buj0zv-SfhduR9cznQ975R5c8-19uLUHSPgALs7XXn_xm3Tu9HpI0qf49b_EOl-pbl06t6u-CliXos876c62qwwn8fYqPf_ZonfZLbuHcHF68un9hywWXsgs-mvbjNeMWxaMdUqX0jBmCdbL1yGIHHkYtA9KqsCsLYXLg0IrqJSzQQjPKi1M-Qhm7ar1TyAthNec25pAYSpuvaqZ0BwbgzIEzJPA2x0jGhtRyak4xtcGoxPiWTPyLIFX49D1AMXxt0HHxM1xAKFn9w2rzecmCmOTS-3KPPei8KbKDTdSGeldLlxha-l4Am9oLzQk40iM1fGpAi6J0LKauSRUOvRMqwQOdyxvovB3zW9-J_By7EaxpVyMbj1yiMZI9FwrhVM8HnbXSHOJQTJXhUxATvbdZFHTnnZ51UOD1-hfV1X19N9kPYObjF5xELJveQiz7ea7fw437I_tstu8iDL0C38eKJw
  priority: 102
  providerName: ProQuest
Title Speaker Adaptation on Articulation and Acoustics for Articulation-to-Speech Synthesis
URI https://www.ncbi.nlm.nih.gov/pubmed/36015817
https://www.proquest.com/docview/2706437285
https://www.proquest.com/docview/2707610484
https://pubmed.ncbi.nlm.nih.gov/PMC9416444
https://doaj.org/article/07ad300e61eb40b5b78b7ed06d1c97d5
Volume 22
WOSCitedRecordID wos000845322000001&url=https%3A%2F%2Fcvtisr.summon.serialssolutions.com%2F%23%21%2Fsearch%3Fho%3Df%26include.ft.matches%3Dt%26l%3Dnull%26q%3D
hasFullText 1
inHoldings 1
isFullTextHit
isPrint
journalDatabaseRights – providerCode: PRVAON
  databaseName: DOAJ Directory of Open Access Journals
  customDbUrl:
  eissn: 1424-8220
  dateEnd: 99991231
  omitProxy: false
  ssIdentifier: ssj0023338
  issn: 1424-8220
  databaseCode: DOA
  dateStart: 20010101
  isFulltext: true
  titleUrlDefault: https://www.doaj.org/
  providerName: Directory of Open Access Journals
– providerCode: PRVHPJ
  databaseName: ROAD: Directory of Open Access Scholarly Resources
  customDbUrl:
  eissn: 1424-8220
  dateEnd: 99991231
  omitProxy: false
  ssIdentifier: ssj0023338
  issn: 1424-8220
  databaseCode: M~E
  dateStart: 20010101
  isFulltext: true
  titleUrlDefault: https://road.issn.org
  providerName: ISSN International Centre
– providerCode: PRVPQU
  databaseName: AUTh Library subscriptions: ProQuest Central
  customDbUrl:
  eissn: 1424-8220
  dateEnd: 99991231
  omitProxy: false
  ssIdentifier: ssj0023338
  issn: 1424-8220
  databaseCode: BENPR
  dateStart: 20010101
  isFulltext: true
  titleUrlDefault: https://www.proquest.com/central
  providerName: ProQuest
– providerCode: PRVPQU
  databaseName: Health & Medical Collection
  customDbUrl:
  eissn: 1424-8220
  dateEnd: 99991231
  omitProxy: false
  ssIdentifier: ssj0023338
  issn: 1424-8220
  databaseCode: 7X7
  dateStart: 20010101
  isFulltext: true
  titleUrlDefault: https://search.proquest.com/healthcomplete
  providerName: ProQuest
– providerCode: PRVPQU
  databaseName: Publicly Available Content Database
  customDbUrl:
  eissn: 1424-8220
  dateEnd: 99991231
  omitProxy: false
  ssIdentifier: ssj0023338
  issn: 1424-8220
  databaseCode: PIMPY
  dateStart: 20010101
  isFulltext: true
  titleUrlDefault: http://search.proquest.com/publiccontent
  providerName: ProQuest
link http://cvtisr.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwrV3Ni9QwFH_o6sE9iN9W16GKoJeybdIm6XFWZtHDDoPrwngq-WQHpTNMZwUv_u2-13bKFAUvQskhSUPy8pL3eyT5PYC3nuIhaWkS2uoSIuhKjClsUgauHVksbXQbbELO52q5LBcHob7oTlhHD9wJ7jSV2vE09SLzJk9NYaQy0rtUuMyW0rXspYh69s5U72px9Lw6HiGOTv1pw1gmELiLkfVpSfr_3IoPbNH4nuSB4Tl_APd7xBhPu54-hFu-fgTHBzyCj-HqcuP1N7-Np05vurP1GL_2lz48V6xrF0_tug3e1cQIVUfFyW6dYCPeXseXP2sEhc2qeQJX57MvHz4mfbyExCLM2iVFyQrLgrFOaS4NY5bYuHwZgkhR9EH7oKQKzFouXBoUGi-lnA1CeJZrYfhTOKrXtX8OcSa8LgpbEpdLXlivSiZ0gZlBGeLTieD9Xo6V7cnEKabF9wqdChJ5NYg8gjdD1U3HoPG3Smc0GUMFIr1uM1AVql4Vqn-pQgTvaCorWprYGav7FwY4JCK5qqaSyOQQUOYRnOxnu-rXbFMx2Z1iKmzo9VCMq42OUHTtcYaojkTAmSts4lmnHEOfOfq2hcpkBHKkNqNBjUvq1XXL6F0iLM7z_MX_kMJLuMfoiQbR9vITONptb_wruGt_7FbNdgK35VK2qZrAnbPZfPF50i4dTC9-zTBv8eli8fU3ttYf3A
linkProvider Directory of Open Access Journals
linkToHtml http://cvtisr.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMw1V1bb9MwFD4aHRLwwP0SGBAQCF6ipU5iOw8Ilcu0altVaZs0noJvYRUoLU0B7U_xGzknNxqBeNsDUp5ix7Lrz-dS298H8MyRHpISOiBTFxBBV6B1YoI0j5Qlj6W0qsQmxGQiT07S6Qb8bO_C0LHK1iZWhtrODf1Hvs1Evcckk9eLrwGpRtHuaiuhUcNiz539wJStfDV-h_P7nLGd90dvd4NGVSAwGIysgiRliWG5NlaqSGjGDHFWuTTPeYgdzJXLpZA5MybiNswlmngprck5dyxWXEfY7gXYjBHscgCb0_HB9EOX4kWY8dX8RVGUhtslY0OOCQPveb1KHOBPF7DmA_vnM9cc3s61_-2nug5Xm9DaH9Vr4QZsuOImXFkjXLwFx4cLpz67pT-yalEfQvDxqT5pdMx8VVh_ZOaVylnpY0zfKw5W8wAbcebUPzwrMHouZ-VtOD6Xcd2BQTEv3D3wh9ypJDEpkd7EiXEyZVwl-DKXmoiHPHjZTnxmGtZ1Ev_4kmH2RRjJOox48LSruqipRv5W6Q2hp6tA7ODVi_nyU9YYmywUykZh6PjQ6TjUiRZSC2dDbocmFTbx4AVhLyMbhp0xqrmKgUMiNrBsJIh1DyPv2IOtFmJZY9zK7De-PHjSFaNZor0mVTicIaojMDKPJTZxt0Zz1-eIYwwqh8ID0cN5b1D9kmJ2WlGfp5g_xHF8_9_degyXdo8O9rP98WTvAVxmdGOFWIyjLRislt_cQ7hovq9m5fJRs359-Hje6-AX8k6H0g
linkToPdf http://cvtisr.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMw1V1bb9MwFD4aG0LwwP0SGBAQCF6iJk5iOw8IFUZFNagqjUnbU_AtrAKlpSmg_TV-HeckaWgE4m0PSHmKHctOPp9LbH8fwBNHekhK6IBMXUAEXYHWqQmyIlaWPJbSqhabEJOJPDrKplvwc30WhrZVrm1ibajt3NA_8gETzRqTTAdFuy1iujd6ufgakIIUrbSu5TQaiOy70x-YvlUvxnv4rZ8yNnrz4fXboFUYCAwGJqsgzVhqWKGNlSoWmjFD_FUuKwoeYmcL5QopZMGMibkNC4nmXkprCs4dSxTXMbZ7DnYwJE9wju1Mx--nx126F2P213AZxXEWDirGIo7JA-95wFoo4E93sOEP-3s1N5zf6Mr__NquwuU25PaHzRy5BluuvA6XNogYb8DhwcKpz27pD61aNJsTfLzqR1p9M1-V1h-aea1-VvkY6_eKg9U8wEacOfEPTkuMqqtZdRMOz2Rct2C7nJfuDvgRdypNTUZkOElqnMwYVyneLKQmQiIPnq9BkJuWjZ1EQb7kmJURXvIOLx487qouGgqSv1V6RUjqKhBreH1jvvyUt0YoD4WycRg6HjmdhDrVQmrhbMhtZDJhUw-eEQ5zsm3YGaPaIxo4JGIJy4eC2PgwIk882F3DLW-NXpX_xpoHj7piNFe0BqVKh1-I6giM2BOJTdxukN31OeYYm8pIeCB6mO8Nql9Szk5qSvQM84okSe7-u1sP4QKCP383nuzfg4uMDrIQuXG8C9ur5Td3H86b76tZtXzQTmUfPp71NPgF7qmQkg
openUrl ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=Speaker+Adaptation+on+Articulation+and+Acoustics+for+Articulation-to-Speech+Synthesis&rft.jtitle=Sensors+%28Basel%2C+Switzerland%29&rft.au=Cao%2C+Beiming&rft.au=Wisler%2C+Alan&rft.au=Wang%2C+Jun&rft.date=2022-08-13&rft.issn=1424-8220&rft.eissn=1424-8220&rft.volume=22&rft.issue=16&rft_id=info:doi/10.3390%2Fs22166056&rft.externalDBID=NO_FULL_TEXT
thumbnail_l http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/lc.gif&issn=1424-8220&client=summon
thumbnail_m http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/mc.gif&issn=1424-8220&client=summon
thumbnail_s http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/sc.gif&issn=1424-8220&client=summon