{"id":1639,"date":"2026-03-20T12:13:40","date_gmt":"2026-03-20T12:13:40","guid":{"rendered":"https:\/\/iberspeech.tech\/2026\/?p=1639"},"modified":"2026-03-20T12:19:36","modified_gmt":"2026-03-20T12:19:36","slug":"%f0%9f%93%a2-call-for-papers-iberspeech-2026","status":"publish","type":"post","link":"https:\/\/iberspeech.tech\/2026\/%f0%9f%93%a2-call-for-papers-iberspeech-2026\/","title":{"rendered":"\ud83d\udce2 Call for Papers \u2014 IberSPEECH 2026"},"content":{"rendered":"\n[et_pb_section fb_built=&#8221;1&#8243; theme_builder_area=&#8221;post_content&#8221; _builder_version=&#8221;4.27.4&#8243; _module_preset=&#8221;default&#8221;][et_pb_row _builder_version=&#8221;4.27.4&#8243; _module_preset=&#8221;default&#8221; theme_builder_area=&#8221;post_content&#8221;][et_pb_column _builder_version=&#8221;4.27.4&#8243; _module_preset=&#8221;default&#8221; type=&#8221;4_4&#8243; theme_builder_area=&#8221;post_content&#8221;][et_pb_text _builder_version=&#8221;4.27.4&#8243; _module_preset=&#8221;default&#8221; theme_builder_area=&#8221;post_content&#8221; hover_enabled=&#8221;0&#8243; sticky_enabled=&#8221;0&#8243;]<h2 data-section-id=\"1crp2o0\" data-start=\"86\" data-end=\"125\">\ud83d\udce2 Call for Papers \u2014 IberSPEECH 2026<\/h2>\n<p data-start=\"127\" data-end=\"312\"><strong data-start=\"127\" data-end=\"146\">IberSPEECH 2026<\/strong> will take place in Madrid, Spain, from <strong data-start=\"186\" data-end=\"213\">November 18 to 20, 2026<\/strong>, bringing together researchers and practitioners in the field of speech and language technologies.<\/p>\n<p data-start=\"314\" data-end=\"553\">This edition combines the <strong data-start=\"340\" data-end=\"381\">XIV Jornadas en Tecnolog\u00edas del Habla<\/strong> and the <strong data-start=\"390\" data-end=\"419\">X Iberian SLTech Workshop<\/strong>, aiming to foster collaboration between academia and industry, with a particular \u2014 though not exclusive \u2014 focus on Iberian languages.<\/p>\n<p data-start=\"555\" data-end=\"605\">\ud83d\udd17 More information: <a href=\"https:\/\/iberspeech.tech\/2026\/call-for-papers\/\">https:\/\/iberspeech.tech\/2026\/call-for-papers\/<\/a><\/p>\n<hr data-start=\"607\" data-end=\"610\" \/>\n<h3 data-section-id=\"1dw83ba\" data-start=\"612\" data-end=\"635\">\ud83d\uddd3\ufe0f Important Dates<\/h3>\n<ul data-start=\"637\" data-end=\"858\">\n<li data-section-id=\"jql23b\" data-start=\"637\" data-end=\"706\">\n<p data-start=\"639\" data-end=\"706\"><strong data-start=\"639\" data-end=\"656\">June 15, 2026<\/strong> \u2014 Initial submission (title, authors, abstract)<\/p>\n<\/li>\n<li data-section-id=\"1wmqivj\" data-start=\"707\" data-end=\"752\">\n<p data-start=\"709\" data-end=\"752\"><strong data-start=\"709\" data-end=\"726\">June 22, 2026<\/strong> \u2014 Full paper submission<\/p>\n<\/li>\n<li data-section-id=\"an2z19\" data-start=\"753\" data-end=\"805\">\n<p data-start=\"755\" data-end=\"805\"><strong data-start=\"755\" data-end=\"777\">September 11, 2026<\/strong> \u2014 Acceptance notification<\/p>\n<\/li>\n<li data-section-id=\"dmn09h\" data-start=\"806\" data-end=\"858\">\n<p data-start=\"808\" data-end=\"858\"><strong data-start=\"808\" data-end=\"830\">September 20, 2026<\/strong> \u2014 Camera-ready papers due<\/p>\n<\/li>\n<\/ul>\n<hr data-start=\"860\" data-end=\"863\" \/>\n<h3 data-section-id=\"1rit9pk\" data-start=\"865\" data-end=\"890\">\ud83c\udfaf Topics of Interest<\/h3>\n<p data-start=\"892\" data-end=\"956\">We invite contributions in areas including (but not limited to):<\/p>\n<ul data-start=\"958\" data-end=\"1423\">\n<li data-section-id=\"1cun99u\" data-start=\"958\" data-end=\"1012\">\n<p data-start=\"960\" data-end=\"1012\">Speech Processing and Automatic Speech Recognition<\/p>\n<\/li>\n<li data-section-id=\"12ey21s\" data-start=\"1013\" data-end=\"1052\">\n<p data-start=\"1015\" data-end=\"1052\">Speaker and Language Identification<\/p>\n<\/li>\n<li data-section-id=\"2jitwn\" data-start=\"1053\" data-end=\"1104\">\n<p data-start=\"1055\" data-end=\"1104\">Speech Synthesis and Spoken Language Generation<\/p>\n<\/li>\n<li data-section-id=\"7zdcn\" data-start=\"1105\" data-end=\"1159\">\n<p data-start=\"1107\" data-end=\"1159\">Spoken Language Understanding and Dialogue Systems<\/p>\n<\/li>\n<li data-section-id=\"c62puz\" data-start=\"1160\" data-end=\"1182\">\n<p data-start=\"1162\" data-end=\"1182\">Speech Translation<\/p>\n<\/li>\n<li data-section-id=\"1gcmdpx\" data-start=\"1183\" data-end=\"1238\">\n<p data-start=\"1185\" data-end=\"1238\">Emotion Recognition and Affective Speech Processing<\/p>\n<\/li>\n<li data-section-id=\"1rb5asl\" data-start=\"1239\" data-end=\"1290\">\n<p data-start=\"1241\" data-end=\"1290\">Speech Enhancement and Acoustic Event Detection<\/p>\n<\/li>\n<li data-section-id=\"nt326g\" data-start=\"1291\" data-end=\"1355\">\n<p data-start=\"1293\" data-end=\"1355\">Speech Technologies for Health, Education, and Accessibility<\/p>\n<\/li>\n<li data-section-id=\"1l423d7\" data-start=\"1356\" data-end=\"1423\">\n<p data-start=\"1358\" data-end=\"1423\">Applications and Systems using Speech and Language Technologies<\/p>\n<\/li>\n<\/ul>\n<hr data-start=\"1425\" data-end=\"1428\" \/>\n<h3 data-section-id=\"1o37f19\" data-start=\"1430\" data-end=\"1457\">\ud83e\uddea Albayzin Evaluations<\/h3>\n<p data-start=\"1459\" data-end=\"1583\">The <strong data-start=\"1463\" data-end=\"1487\">Albayzin Evaluations<\/strong> will be held in conjunction with IberSPEECH 2026, with results presented during the conference.<\/p>\n<hr data-start=\"1585\" data-end=\"1588\" \/>\n<h3 data-section-id=\"1rc7h6l\" data-start=\"1590\" data-end=\"1617\">\ud83d\udc65 Organizing Committee<\/h3>\n<ul data-start=\"1619\" data-end=\"1947\">\n<li data-section-id=\"5fc1tj\" data-start=\"1619\" data-end=\"1708\">\n<p data-start=\"1621\" data-end=\"1708\"><strong data-start=\"1621\" data-end=\"1639\">General Chair:<\/strong> Jos\u00e9 Manuel Pardo Mu\u00f1oz (Universidad Polit\u00e9cnica de Madrid, Spain)<\/p>\n<\/li>\n<li data-section-id=\"akws7p\" data-start=\"1709\" data-end=\"1796\">\n<p data-start=\"1711\" data-end=\"1796\"><strong data-start=\"1711\" data-end=\"1739\">Technical Program Chair:<\/strong> Inma Hern\u00e1ez Rioja (Universidad del Pa\u00eds Vasco, Spain)<\/p>\n<\/li>\n<li data-section-id=\"150ggd8\" data-start=\"1797\" data-end=\"1877\">\n<p data-start=\"1799\" data-end=\"1877\"><strong data-start=\"1799\" data-end=\"1820\">Challenges Chair:<\/strong> Eduardo Lleida Solano (Universidad de Zaragoza, Spain)<\/p>\n<\/li>\n<li data-section-id=\"6ueglx\" data-start=\"1878\" data-end=\"1947\">\n<p data-start=\"1880\" data-end=\"1947\"><strong data-start=\"1880\" data-end=\"1907\">Special Sessions Chair:<\/strong> Dayana Ribas (SONOC \/ BTS Group, Spain)<\/p>\n<\/li>\n<\/ul>\n<p data-start=\"2100\" data-end=\"2184\" data-is-last-node=\"\"><\/p>[\/et_pb_text][\/et_pb_column][\/et_pb_row][\/et_pb_section]\n","protected":false},"excerpt":{"rendered":"<p>Join IberSPEECH 2026 in Madrid this November and share your latest advances in speech and language technologies. Paper submissions open until June 2026.<\/p>\n","protected":false},"author":4,"featured_media":0,"comment_status":"closed","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"_et_pb_use_builder":"on","_et_pb_old_content":"","_et_gb_content_width":"2880","footnotes":""},"categories":[19],"tags":[],"class_list":["post-1639","post","type-post","status-publish","format-standard","hentry","category-news"],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v27.0 - https:\/\/yoast.com\/product\/yoast-seo-wordpress\/ -->\n<title>\ud83d\udce2 Call for Papers \u2014 IberSPEECH 2026 - Iberspeech<\/title>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/iberspeech.tech\/2026\/\ud83d\udce2-call-for-papers-iberspeech-2026\/\" \/>\n<meta property=\"og:locale\" content=\"en_US\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"\ud83d\udce2 Call for Papers \u2014 IberSPEECH 2026 - Iberspeech\" \/>\n<meta property=\"og:description\" content=\"Join IberSPEECH 2026 in Madrid this November and share your latest advances in speech and language technologies. Paper submissions open until June 2026.\" \/>\n<meta property=\"og:url\" content=\"https:\/\/iberspeech.tech\/2026\/\ud83d\udce2-call-for-papers-iberspeech-2026\/\" \/>\n<meta property=\"og:site_name\" content=\"Iberspeech\" \/>\n<meta property=\"article:published_time\" content=\"2026-03-20T12:13:40+00:00\" \/>\n<meta property=\"article:modified_time\" content=\"2026-03-20T12:19:36+00:00\" \/>\n<meta name=\"author\" content=\"Fernando Fern\u00e1ndez-Mart\u00ednez\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:label1\" content=\"Written by\" \/>\n\t<meta name=\"twitter:data1\" content=\"Fernando Fern\u00e1ndez-Mart\u00ednez\" \/>\n\t<meta name=\"twitter:label2\" content=\"Est. reading time\" \/>\n\t<meta name=\"twitter:data2\" content=\"1 minute\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\/\/schema.org\",\"@graph\":[{\"@type\":\"Article\",\"@id\":\"https:\/\/iberspeech.tech\/2026\/%f0%9f%93%a2-call-for-papers-iberspeech-2026\/#article\",\"isPartOf\":{\"@id\":\"https:\/\/iberspeech.tech\/2026\/%f0%9f%93%a2-call-for-papers-iberspeech-2026\/\"},\"author\":{\"name\":\"Fernando Fern\u00e1ndez-Mart\u00ednez\",\"@id\":\"https:\/\/iberspeech.tech\/2026\/#\/schema\/person\/fbc06d727153139c99b3febcc11f5977\"},\"headline\":\"\ud83d\udce2 Call for Papers \u2014 IberSPEECH 2026\",\"datePublished\":\"2026-03-20T12:13:40+00:00\",\"dateModified\":\"2026-03-20T12:19:36+00:00\",\"mainEntityOfPage\":{\"@id\":\"https:\/\/iberspeech.tech\/2026\/%f0%9f%93%a2-call-for-papers-iberspeech-2026\/\"},\"wordCount\":290,\"articleSection\":[\"NEWS\"],\"inLanguage\":\"en-US\"},{\"@type\":\"WebPage\",\"@id\":\"https:\/\/iberspeech.tech\/2026\/%f0%9f%93%a2-call-for-papers-iberspeech-2026\/\",\"url\":\"https:\/\/iberspeech.tech\/2026\/%f0%9f%93%a2-call-for-papers-iberspeech-2026\/\",\"name\":\"\ud83d\udce2 Call for Papers \u2014 IberSPEECH 2026 - Iberspeech\",\"isPartOf\":{\"@id\":\"https:\/\/iberspeech.tech\/2026\/#website\"},\"datePublished\":\"2026-03-20T12:13:40+00:00\",\"dateModified\":\"2026-03-20T12:19:36+00:00\",\"author\":{\"@id\":\"https:\/\/iberspeech.tech\/2026\/#\/schema\/person\/fbc06d727153139c99b3febcc11f5977\"},\"breadcrumb\":{\"@id\":\"https:\/\/iberspeech.tech\/2026\/%f0%9f%93%a2-call-for-papers-iberspeech-2026\/#breadcrumb\"},\"inLanguage\":\"en-US\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\/\/iberspeech.tech\/2026\/%f0%9f%93%a2-call-for-papers-iberspeech-2026\/\"]}]},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\/\/iberspeech.tech\/2026\/%f0%9f%93%a2-call-for-papers-iberspeech-2026\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"Home\",\"item\":\"https:\/\/iberspeech.tech\/2026\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"\ud83d\udce2 Call for Papers \u2014 IberSPEECH 2026\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\/\/iberspeech.tech\/2026\/#website\",\"url\":\"https:\/\/iberspeech.tech\/2026\/\",\"name\":\"Iberspeech\",\"description\":\"\",\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\/\/iberspeech.tech\/2026\/?s={search_term_string}\"},\"query-input\":{\"@type\":\"PropertyValueSpecification\",\"valueRequired\":true,\"valueName\":\"search_term_string\"}}],\"inLanguage\":\"en-US\"},{\"@type\":\"Person\",\"@id\":\"https:\/\/iberspeech.tech\/2026\/#\/schema\/person\/fbc06d727153139c99b3febcc11f5977\",\"name\":\"Fernando Fern\u00e1ndez-Mart\u00ednez\",\"image\":{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\/\/iberspeech.tech\/2026\/#\/schema\/person\/image\/\",\"url\":\"https:\/\/secure.gravatar.com\/avatar\/bc0c875fa52c5a36a2878e47042a282ef0574dc119487ad6c4b6792d9a5ee881?s=96&d=mm&r=g\",\"contentUrl\":\"https:\/\/secure.gravatar.com\/avatar\/bc0c875fa52c5a36a2878e47042a282ef0574dc119487ad6c4b6792d9a5ee881?s=96&d=mm&r=g\",\"caption\":\"Fernando Fern\u00e1ndez-Mart\u00ednez\"},\"url\":\"https:\/\/iberspeech.tech\/2026\/author\/ffm\/\"}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"\ud83d\udce2 Call for Papers \u2014 IberSPEECH 2026 - Iberspeech","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/iberspeech.tech\/2026\/\ud83d\udce2-call-for-papers-iberspeech-2026\/","og_locale":"en_US","og_type":"article","og_title":"\ud83d\udce2 Call for Papers \u2014 IberSPEECH 2026 - Iberspeech","og_description":"Join IberSPEECH 2026 in Madrid this November and share your latest advances in speech and language technologies. Paper submissions open until June 2026.","og_url":"https:\/\/iberspeech.tech\/2026\/\ud83d\udce2-call-for-papers-iberspeech-2026\/","og_site_name":"Iberspeech","article_published_time":"2026-03-20T12:13:40+00:00","article_modified_time":"2026-03-20T12:19:36+00:00","author":"Fernando Fern\u00e1ndez-Mart\u00ednez","twitter_card":"summary_large_image","twitter_misc":{"Written by":"Fernando Fern\u00e1ndez-Mart\u00ednez","Est. reading time":"1 minute"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"Article","@id":"https:\/\/iberspeech.tech\/2026\/%f0%9f%93%a2-call-for-papers-iberspeech-2026\/#article","isPartOf":{"@id":"https:\/\/iberspeech.tech\/2026\/%f0%9f%93%a2-call-for-papers-iberspeech-2026\/"},"author":{"name":"Fernando Fern\u00e1ndez-Mart\u00ednez","@id":"https:\/\/iberspeech.tech\/2026\/#\/schema\/person\/fbc06d727153139c99b3febcc11f5977"},"headline":"\ud83d\udce2 Call for Papers \u2014 IberSPEECH 2026","datePublished":"2026-03-20T12:13:40+00:00","dateModified":"2026-03-20T12:19:36+00:00","mainEntityOfPage":{"@id":"https:\/\/iberspeech.tech\/2026\/%f0%9f%93%a2-call-for-papers-iberspeech-2026\/"},"wordCount":290,"articleSection":["NEWS"],"inLanguage":"en-US"},{"@type":"WebPage","@id":"https:\/\/iberspeech.tech\/2026\/%f0%9f%93%a2-call-for-papers-iberspeech-2026\/","url":"https:\/\/iberspeech.tech\/2026\/%f0%9f%93%a2-call-for-papers-iberspeech-2026\/","name":"\ud83d\udce2 Call for Papers \u2014 IberSPEECH 2026 - Iberspeech","isPartOf":{"@id":"https:\/\/iberspeech.tech\/2026\/#website"},"datePublished":"2026-03-20T12:13:40+00:00","dateModified":"2026-03-20T12:19:36+00:00","author":{"@id":"https:\/\/iberspeech.tech\/2026\/#\/schema\/person\/fbc06d727153139c99b3febcc11f5977"},"breadcrumb":{"@id":"https:\/\/iberspeech.tech\/2026\/%f0%9f%93%a2-call-for-papers-iberspeech-2026\/#breadcrumb"},"inLanguage":"en-US","potentialAction":[{"@type":"ReadAction","target":["https:\/\/iberspeech.tech\/2026\/%f0%9f%93%a2-call-for-papers-iberspeech-2026\/"]}]},{"@type":"BreadcrumbList","@id":"https:\/\/iberspeech.tech\/2026\/%f0%9f%93%a2-call-for-papers-iberspeech-2026\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Home","item":"https:\/\/iberspeech.tech\/2026\/"},{"@type":"ListItem","position":2,"name":"\ud83d\udce2 Call for Papers \u2014 IberSPEECH 2026"}]},{"@type":"WebSite","@id":"https:\/\/iberspeech.tech\/2026\/#website","url":"https:\/\/iberspeech.tech\/2026\/","name":"Iberspeech","description":"","potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/iberspeech.tech\/2026\/?s={search_term_string}"},"query-input":{"@type":"PropertyValueSpecification","valueRequired":true,"valueName":"search_term_string"}}],"inLanguage":"en-US"},{"@type":"Person","@id":"https:\/\/iberspeech.tech\/2026\/#\/schema\/person\/fbc06d727153139c99b3febcc11f5977","name":"Fernando Fern\u00e1ndez-Mart\u00ednez","image":{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/iberspeech.tech\/2026\/#\/schema\/person\/image\/","url":"https:\/\/secure.gravatar.com\/avatar\/bc0c875fa52c5a36a2878e47042a282ef0574dc119487ad6c4b6792d9a5ee881?s=96&d=mm&r=g","contentUrl":"https:\/\/secure.gravatar.com\/avatar\/bc0c875fa52c5a36a2878e47042a282ef0574dc119487ad6c4b6792d9a5ee881?s=96&d=mm&r=g","caption":"Fernando Fern\u00e1ndez-Mart\u00ednez"},"url":"https:\/\/iberspeech.tech\/2026\/author\/ffm\/"}]}},"_links":{"self":[{"href":"https:\/\/iberspeech.tech\/2026\/wp-json\/wp\/v2\/posts\/1639","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/iberspeech.tech\/2026\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/iberspeech.tech\/2026\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/iberspeech.tech\/2026\/wp-json\/wp\/v2\/users\/4"}],"replies":[{"embeddable":true,"href":"https:\/\/iberspeech.tech\/2026\/wp-json\/wp\/v2\/comments?post=1639"}],"version-history":[{"count":6,"href":"https:\/\/iberspeech.tech\/2026\/wp-json\/wp\/v2\/posts\/1639\/revisions"}],"predecessor-version":[{"id":1647,"href":"https:\/\/iberspeech.tech\/2026\/wp-json\/wp\/v2\/posts\/1639\/revisions\/1647"}],"wp:attachment":[{"href":"https:\/\/iberspeech.tech\/2026\/wp-json\/wp\/v2\/media?parent=1639"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/iberspeech.tech\/2026\/wp-json\/wp\/v2\/categories?post=1639"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/iberspeech.tech\/2026\/wp-json\/wp\/v2\/tags?post=1639"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}