2016 |
Sanchez-Cortina, Isaias; Andrés-Ferrer, Jesús; Sanchis, Alberto; Juan, Alfons Speaker-adapted confidence measures for speech recognition of video lectures Journal Article Computer Speech & Language, 37 , pp. 11–23, 2016, ISBN: 0885-2308. Abstract | Links | BibTeX | Tags: Confidence measures, Log-linear models, Online video lectures, Speaker adaptation, Speech Recognition @article{SanchezCortina2016, title = {Speaker-adapted confidence measures for speech recognition of video lectures}, author = {Isaias Sanchez-Cortina and Jesús Andrés-Ferrer and Alberto Sanchis and Alfons Juan}, url = {http://www.sciencedirect.com/science/article/pii/S0885230815000960 http://authors.elsevier.com/a/1SAsB39HpSHRc0}, isbn = {0885-2308}, year = {2016}, date = {2016-01-01}, journal = {Computer Speech & Language}, volume = {37}, pages = {11--23}, abstract = {Abstract Automatic Speech Recognition applications can benefit from a confidence measure (CM) to predict the reliability of the output. Previous works showed that a word-dependent naïve Bayes (NB) classifier outperforms the conventional word posterior probability as a CM. However, a discriminative formulation usually renders improved performance due to the available training techniques. Taking this into account, we propose a logistic regression (LR) classifier defined with simple input functions to approximate to the \\{NB\\} behaviour. Additionally, as a main contribution, we propose to adapt the \\{CM\\} to the speaker in cases in which it is possible to identify the speakers, such as online lecture repositories. The experiments have shown that speaker-adapted models outperform their non-adapted counterparts on two difficult tasks from English (videoLectures.net) and Spanish (poliMedia) educational lectures. They have also shown that the \\{NB\\} model is clearly superseded by the proposed \\{LR\\} classifier.}, keywords = {Confidence measures, Log-linear models, Online video lectures, Speaker adaptation, Speech Recognition}, pubstate = {published}, tppubtype = {article} } Abstract Automatic Speech Recognition applications can benefit from a confidence measure (CM) to predict the reliability of the output. Previous works showed that a word-dependent naïve Bayes (NB) classifier outperforms the conventional word posterior probability as a CM. However, a discriminative formulation usually renders improved performance due to the available training techniques. Taking this into account, we propose a logistic regression (LR) classifier defined with simple input functions to approximate to the \{NB\} behaviour. Additionally, as a main contribution, we propose to adapt the \{CM\} to the speaker in cases in which it is possible to identify the speakers, such as online lecture repositories. The experiments have shown that speaker-adapted models outperform their non-adapted counterparts on two difficult tasks from English (videoLectures.net) and Spanish (poliMedia) educational lectures. They have also shown that the \{NB\} model is clearly superseded by the proposed \{LR\} classifier. |
2015 |
del-Agua, Miguel Ángel; Martínez-Villaronga, Adrià; Piqueras, Santiago; Giménez, Adrià; Sanchis, Alberto; Civera, Jorge; Juan, Alfons The MLLP ASR Systems for IWSLT 2015 Inproceedings Proc. of 12th Intl. Workshop on Spoken Language Translation (IWSLT 2015), pp. 39–44, Da Nang (Vietnam), 2015. Abstract | Links | BibTeX | Tags: @inproceedings{delAgua15, title = {The MLLP ASR Systems for IWSLT 2015}, author = {Miguel Ángel del-Agua and Adrià Martínez-Villaronga and Santiago Piqueras and Adrià Giménez and Alberto Sanchis and Jorge Civera and Alfons Juan}, url = {https://aclanthology.org/2015.iwslt-evaluation.5/}, year = {2015}, date = {2015-12-03}, booktitle = {Proc. of 12th Intl. Workshop on Spoken Language Translation (IWSLT 2015)}, pages = {39--44}, address = {Da Nang (Vietnam)}, abstract = {This paper describes the Machine Learning and Language Processing (MLLP) ASR systems for the 2015 IWSLT evaluation campaing. The English system is based on the combination of five different subsystems which consist of two types of Neural Networks architectures (Deep feed-forward and Convolutional), two types of activation functions (sigmoid and rectified linear) and two types of input features (fMLLR and FBANK). All subsystems perform a speaker adaptation step based on confidence measures, the output of which is then combined with ROVER. This system achieves a Word Error Rate (WER) of 13.3% on the official IWSLT 2015 English test set.}, keywords = {}, pubstate = {published}, tppubtype = {inproceedings} } This paper describes the Machine Learning and Language Processing (MLLP) ASR systems for the 2015 IWSLT evaluation campaing. The English system is based on the combination of five different subsystems which consist of two types of Neural Networks architectures (Deep feed-forward and Convolutional), two types of activation functions (sigmoid and rectified linear) and two types of input features (fMLLR and FBANK). All subsystems perform a speaker adaptation step based on confidence measures, the output of which is then combined with ROVER. This system achieves a Word Error Rate (WER) of 13.3% on the official IWSLT 2015 English test set. |
Valor Miró, Juan Daniel ; Silvestre-Cerdà, Joan Albert ; Civera, Jorge ; Turró, Carlos ; Juan, Alfons Efficient Generation of High-Quality Multilingual Subtitles for Video Lecture Repositories Inproceedings Proc. of 10th European Conf. on Technology Enhanced Learning (EC-TEL 2015), pp. 485–490, Toledo (Spain), 2015, ISBN: 978-3-319-24258-3. Abstract | Links | BibTeX | Tags: Automatic Speech Recognition, Docencia en Red, Efficient video subtitling, Polimedia, Statistical machine translation, video lecture repositories @inproceedings{valor2015efficient, title = {Efficient Generation of High-Quality Multilingual Subtitles for Video Lecture Repositories}, author = {Valor Miró, Juan Daniel and Silvestre-Cerdà, Joan Albert and Civera, Jorge and Turró, Carlos and Juan, Alfons}, url = {http://link.springer.com/chapter/10.1007/978-3-319-24258-3_44 http://www.mllp.upv.es/wp-content/uploads/2016/03/paper.pdf }, isbn = {978-3-319-24258-3}, year = {2015}, date = {2015-09-17}, booktitle = {Proc. of 10th European Conf. on Technology Enhanced Learning (EC-TEL 2015)}, pages = {485--490}, address = {Toledo (Spain)}, abstract = {Video lectures are a valuable educational tool in higher education to support or replace face-to-face lectures in active learning strategies. In 2007 the Universitat Polit‘ecnica de Val‘encia (UPV) implemented its video lecture capture system, resulting in a high quality educational video repository, called poliMedia, with more than 10.000 mini lectures created by 1.373 lecturers. Also, in the framework of the European project transLectures, UPV has automatically generated transcriptions and translations in Spanish, Catalan and English for all videos included in the poliMedia video repository. transLectures’s objective responds to the widely-recognised need for subtitles to be provided with video lectures, as an essential service for non-native speakers and hearing impaired persons, and to allow advanced repository functionalities. Although high-quality automatic transcriptions and translations were generated in transLectures, they were not error-free. For this reason, lecturers need to manually review video subtitles to guarantee the absence of errors. The aim of this study is to evaluate the efficiency of the manual review process from automatic subtitles in comparison with the conventional generation of video subtitles from scratch. The reported results clearly indicate the convenience of providing automatic subtitles as a first step in the generation of video subtitles and the significant savings in time of up to almost 75% involved in reviewing subtitles.}, keywords = {Automatic Speech Recognition, Docencia en Red, Efficient video subtitling, Polimedia, Statistical machine translation, video lecture repositories}, pubstate = {published}, tppubtype = {inproceedings} } Video lectures are a valuable educational tool in higher education to support or replace face-to-face lectures in active learning strategies. In 2007 the Universitat Polit‘ecnica de Val‘encia (UPV) implemented its video lecture capture system, resulting in a high quality educational video repository, called poliMedia, with more than 10.000 mini lectures created by 1.373 lecturers. Also, in the framework of the European project transLectures, UPV has automatically generated transcriptions and translations in Spanish, Catalan and English for all videos included in the poliMedia video repository. transLectures’s objective responds to the widely-recognised need for subtitles to be provided with video lectures, as an essential service for non-native speakers and hearing impaired persons, and to allow advanced repository functionalities. Although high-quality automatic transcriptions and translations were generated in transLectures, they were not error-free. For this reason, lecturers need to manually review video subtitles to guarantee the absence of errors. The aim of this study is to evaluate the efficiency of the manual review process from automatic subtitles in comparison with the conventional generation of video subtitles from scratch. The reported results clearly indicate the convenience of providing automatic subtitles as a first step in the generation of video subtitles and the significant savings in time of up to almost 75% involved in reviewing subtitles. |
Pérez González de Martos, Alejandro ; Silvestre-Cerdà, Joan Albert ; Valor Miró, Juan Daniel ; Civera, Jorge ; Juan, Alfons MLLP Transcription and Translation Platform Miscellaneous 2015, (Short paper for demo presentation accepted at 10th European Conf. on Technology Enhanced Learning (EC-TEL 2015), Toledo (Spain), 2015.). Abstract | Links | BibTeX | Tags: Automatic Speech Recognition, Docencia en Red, Document translation, Efficient video subtitling, Machine Translation, MLLP, Post-editing, Video Lectures @misc{mllpttp, title = {MLLP Transcription and Translation Platform}, author = {Pérez González de Martos, Alejandro and Silvestre-Cerdà, Joan Albert and Valor Miró, Juan Daniel and Civera, Jorge and Juan, Alfons}, url = {http://hdl.handle.net/10251/65747 http://www.mllp.upv.es/wp-content/uploads/2015/09/ttp_platform_demo_ectel2015.pdf http://ectel2015.httc.de/index.php?id=722}, year = {2015}, date = {2015-09-16}, booktitle = {Tenth European Conference On Technology Enhanced Learning (EC-TEL 2015)}, abstract = {This paper briefly presents the main features of MLLP’s Transcription and Translation Platform, which uses state-of-the-art automatic speech recognition and machine translation systems to generate multilingual subtitles of educational audiovisual and textual content. It has proven to reduce user effort up to 1/3 of the time needed to generate transcriptions and translations from scratch.}, note = {Short paper for demo presentation accepted at 10th European Conf. on Technology Enhanced Learning (EC-TEL 2015), Toledo (Spain), 2015.}, keywords = {Automatic Speech Recognition, Docencia en Red, Document translation, Efficient video subtitling, Machine Translation, MLLP, Post-editing, Video Lectures}, pubstate = {published}, tppubtype = {misc} } This paper briefly presents the main features of MLLP’s Transcription and Translation Platform, which uses state-of-the-art automatic speech recognition and machine translation systems to generate multilingual subtitles of educational audiovisual and textual content. It has proven to reduce user effort up to 1/3 of the time needed to generate transcriptions and translations from scratch. |
Valor Miró, Juan Daniel ; Turró, C; Civera, J; Juan, A Evaluación de la revisión de transcripciones y traducciones automáticas de vídeos poliMedia Inproceedings Proc. of I Congreso Nacional de Innovación Educativa y Docencia en Red (IN-RED 2015), pp. 464–468, València (Spain), 2015. Links | BibTeX | Tags: Docencia en Red, evaluaciones con usuario, Polimedia, traducciones, transcripciones @inproceedings{Valor-InRed2015, title = {Evaluación de la revisión de transcripciones y traducciones automáticas de vídeos poliMedia}, author = {Valor Miró, Juan Daniel and Turró, C. and Civera, J. and Juan, A.}, url = {http://hdl.handle.net/10251/52755 http://www.mllp.upv.es/wp-content/uploads/2015/06/1574-3087-1-PB.pdf}, year = {2015}, date = {2015-06-30}, booktitle = {Proc. of I Congreso Nacional de Innovación Educativa y Docencia en Red (IN-RED 2015)}, pages = {464--468}, address = {València (Spain)}, keywords = {Docencia en Red, evaluaciones con usuario, Polimedia, traducciones, transcripciones}, pubstate = {published}, tppubtype = {inproceedings} } |
Valor Miró, Juan Daniel ; Silvestre-Cerdà, Joan Albert; Civera, Jorge; Turró, Carlos; Juan, Alfons Efficiency and usability study of innovative computer-aided transcription strategies for video lecture repositories Journal Article Speech Communication, 74 , pp. 65–75, 2015, ISSN: 0167-6393. Abstract | Links | BibTeX | Tags: Automatic Speech Recognition, Computer-assisted transcription, Interface design strategies, Usability study, video lecture repositories @article{Valor201565, title = {Efficiency and usability study of innovative computer-aided transcription strategies for video lecture repositories}, author = {Valor Miró, Juan Daniel and Joan Albert Silvestre-Cerdà and Jorge Civera and Carlos Turró and Alfons Juan}, url = {http://www.sciencedirect.com/science/article/pii/S0167639315001016 http://www.mllp.upv.es/wp-content/uploads/2016/03/paper1.pdf}, issn = {0167-6393}, year = {2015}, date = {2015-01-01}, journal = {Speech Communication}, volume = {74}, pages = {65--75}, abstract = {Abstract Video lectures are widely used in education to support and complement face-to-face lectures. However, the utility of these audiovisual assets could be further improved by adding subtitles that can be exploited to incorporate added-value functionalities such as searchability, accessibility, translatability, note-taking, and discovery of content-related videos, among others. Today, automatic subtitles are prone to error, and need to be reviewed and post-edited in order to ensure that what students see on-screen are of an acceptable quality. This work investigates different user interface design strategies for this post-editing task to discover the best way to incorporate automatic transcription technologies into large educational video repositories. Our three-phase study involved lecturers from the Universitat Politècnica de València (UPV) with videos available on the poliMedia video lecture repository, which is currently over 10,000 video objects. Simply by conventional post-editing automatic transcriptions users almost reduced to half the time that would require to generate the transcription from scratch. As expected, this study revealed that the time spent by lecturers reviewing automatic transcriptions correlated directly with the accuracy of said transcriptions. However, it is also shown that the average time required to perform each individual editing operation could be precisely derived and could be applied in the definition of a user model. In addition, the second phase of this study presents a transcription review strategy based on confidence measures (CM) and compares it to the conventional post-editing strategy. Finally, a third strategy resulting from the combination of that based on \\{CM\\} with massive adaptation techniques for automatic speech recognition (ASR), achieved to improve the transcription review efficiency in comparison with the two aforementioned strategies.}, keywords = {Automatic Speech Recognition, Computer-assisted transcription, Interface design strategies, Usability study, video lecture repositories}, pubstate = {published}, tppubtype = {article} } Abstract Video lectures are widely used in education to support and complement face-to-face lectures. However, the utility of these audiovisual assets could be further improved by adding subtitles that can be exploited to incorporate added-value functionalities such as searchability, accessibility, translatability, note-taking, and discovery of content-related videos, among others. Today, automatic subtitles are prone to error, and need to be reviewed and post-edited in order to ensure that what students see on-screen are of an acceptable quality. This work investigates different user interface design strategies for this post-editing task to discover the best way to incorporate automatic transcription technologies into large educational video repositories. Our three-phase study involved lecturers from the Universitat Politècnica de València (UPV) with videos available on the poliMedia video lecture repository, which is currently over 10,000 video objects. Simply by conventional post-editing automatic transcriptions users almost reduced to half the time that would require to generate the transcription from scratch. As expected, this study revealed that the time spent by lecturers reviewing automatic transcriptions correlated directly with the accuracy of said transcriptions. However, it is also shown that the average time required to perform each individual editing operation could be precisely derived and could be applied in the definition of a user model. In addition, the second phase of this study presents a transcription review strategy based on confidence measures (CM) and compares it to the conventional post-editing strategy. Finally, a third strategy resulting from the combination of that based on \{CM\} with massive adaptation techniques for automatic speech recognition (ASR), achieved to improve the transcription review efficiency in comparison with the two aforementioned strategies. |
Brouns, Francis; Serrano Martínez-Santos, Nicolás ; Civera, Jorge; Kalz, Marco; Juan, Alfons Supporting language diversity of European MOOCs with the EMMA platform Inproceedings Proc. of the European MOOC Stakeholder Summit EMOOCs 2015, pp. 157–165, Mons (Belgium), 2015. Abstract | Links | BibTeX | Tags: Automatic Speech Recognition, EMMA, Statistical machine translation @inproceedings{Brouns2015, title = {Supporting language diversity of European MOOCs with the EMMA platform}, author = {Francis Brouns and Serrano Martínez-Santos, Nicolás and Jorge Civera and Marco Kalz and Alfons Juan}, url = {http://www.emoocs2015.eu/node/55}, year = {2015}, date = {2015-01-01}, booktitle = {Proc. of the European MOOC Stakeholder Summit EMOOCs 2015}, pages = {157--165}, address = {Mons (Belgium)}, abstract = {This paper introduces the cross-language support of the EMMA MOOC platform. Based on a discussion of language diversity in Europe, we introduce the development and evaluation of automated translation of texts and subtitling of videos from Dutch into English. The development of an Automatic Speech Recognition (ASR) system and a Statistical Machine Translation (SMT) system is described. The resources employed and evaluation approach is introduced. Initial evaluation results are presented. Finally, we provide an outlook into future research and development.}, keywords = {Automatic Speech Recognition, EMMA, Statistical machine translation}, pubstate = {published}, tppubtype = {inproceedings} } This paper introduces the cross-language support of the EMMA MOOC platform. Based on a discussion of language diversity in Europe, we introduce the development and evaluation of automated translation of texts and subtitling of videos from Dutch into English. The development of an Automatic Speech Recognition (ASR) system and a Statistical Machine Translation (SMT) system is described. The resources employed and evaluation approach is introduced. Initial evaluation results are presented. Finally, we provide an outlook into future research and development. |
Khoury, Ihab; Giménez, Adrià; Juan, Alfons; Andrés-Ferrer, Jesús Window Repositioning for Printed Arabic Recognition Journal Article Pattern Recognition Letters, 51 , pp. 86–93, 2015, ISSN: 0167-8655. Abstract | Links | BibTeX | Tags: Bernoulli HMMs, Printed Arabic Recognition, Repositioning, Sliding window @article{Kho14, title = {Window Repositioning for Printed Arabic Recognition}, author = {Ihab Khoury and Adrià Giménez and Alfons Juan and Jesús Andrés-Ferrer}, url = {http://dx.doi.org/10.1016/j.patrec.2014.08.009}, issn = {0167-8655}, year = {2015}, date = {2015-01-01}, journal = {Pattern Recognition Letters}, volume = {51}, pages = {86--93}, abstract = {Bernoulli HMMs are conventional HMMs in which the emission probabilities are modeled with Bernoulli mixtures. They have recently been applied, with good results, in off-line text recognition in many languages, in particular, Arabic. A key idea that has proven to be very effective in this application of Bernoulli HMMs is the use of a sliding window of adequate width for feature extraction. This idea has allowed us to obtain very competitive results in the recognition of both Arabic handwriting and printed text. Indeed, a system based on it ranked first at the ICDAR 2011 Arabic recognition competition on the Arabic Printed Text Image (APTI) database. More recently, this idea has been refined by using repositioning techniques for extracted windows, leading to further improvements in Arabic handwriting recognition. In the case of printed text, this refinement led to an improved system which ranked second at the ICDAR 2013 second competition on APTI, only at a marginal distance from the best system. In this work, we describe the development of this improved system. Following evaluation protocols similar to those of the competitions on APTI, exhaustive experiments are detailed from which state-of-the-art results are obtained.}, keywords = {Bernoulli HMMs, Printed Arabic Recognition, Repositioning, Sliding window}, pubstate = {published}, tppubtype = {article} } Bernoulli HMMs are conventional HMMs in which the emission probabilities are modeled with Bernoulli mixtures. They have recently been applied, with good results, in off-line text recognition in many languages, in particular, Arabic. A key idea that has proven to be very effective in this application of Bernoulli HMMs is the use of a sliding window of adequate width for feature extraction. This idea has allowed us to obtain very competitive results in the recognition of both Arabic handwriting and printed text. Indeed, a system based on it ranked first at the ICDAR 2011 Arabic recognition competition on the Arabic Printed Text Image (APTI) database. More recently, this idea has been refined by using repositioning techniques for extracted windows, leading to further improvements in Arabic handwriting recognition. In the case of printed text, this refinement led to an improved system which ranked second at the ICDAR 2013 second competition on APTI, only at a marginal distance from the best system. In this work, we describe the development of this improved system. Following evaluation protocols similar to those of the competitions on APTI, exhaustive experiments are detailed from which state-of-the-art results are obtained. |
2014 |
Valor Miró, Juan Daniel ; Spencer, R N; Pérez González de Martos, A; Garcés Díaz-Munío, G; Turró, C; Civera, J; Juan, A Evaluación del proceso de revisión de transcripciones automáticas para vídeos Polimedia Inproceedings Proc. of I Jornadas de Innovación Educativa y Docencia en Red (IN-RED 2014), pp. 272–278, Valencia (Spain), 2014. Abstract | Links | BibTeX | Tags: ASR, Docencia en Red, evaluaciones, Polimedia, transcripciones @inproceedings{Valor14-InRed, title = {Evaluación del proceso de revisión de transcripciones automáticas para vídeos Polimedia}, author = {Valor Miró, Juan Daniel and Spencer, R.N. and Pérez González de Martos, A. and Garcés Díaz-Munío, G. and Turró, C. and Civera, J. and Juan, A.}, url = {http://hdl.handle.net/10251/40404 http://dx.doi.org/10.4995/INRED.2014 http://www.mllp.upv.es/wp-content/uploads/2015/04/paper1.pdf https://www.mllp.upv.es/wp-content/uploads/2019/09/poster.pdf}, year = {2014}, date = {2014-07-01}, booktitle = {Proc. of I Jornadas de Innovación Educativa y Docencia en Red (IN-RED 2014)}, pages = {272--278}, address = {Valencia (Spain)}, abstract = {[EN] Video lectures are a tool of proven value and wide acceptance in universities which is leading to video lecture platforms like poliMèdia (Universitat Politècnica de València). transLectures is an EU project that generates high-quality automatic transcriptions and translations for the poliMèdia platform, and improves them by using massive adaptation and intelligent interaction techniques. In this paper we present the evaluation with lecturers carried out under the Docència en Xarxa 2012-2013 action plan with the aim of studying the process of transcription post-editing, in contrast with transcribing from scratch. [ES] Los vídeos docentes son una herramienta de demostrada utilidad y gran aceptación en el mundo universitario que está dando lugar a plataformas de vídeos docentes como poliMèdia (Universitat Politècnica de València). transLectures es un proyecto europeo que genera transcripciones y traducciones automáticas de alta calidad para la plataforma poliMèdia, mediante técnicas de adaptación masiva e interacción inteligente. En este artículo presentamos la evaluación con profesores que se realizó en el marco del plan de acción Docència en Xarxa 2012-2013, con el objetivo de estudiar el proceso de supervisión de transcripciones, comparándolo con la obtención de la transcripción sin disponer de una transcripción automática previa. [CA] Els vídeos docents són una eina d'utilitat demostrada i amb gran acceptació en el món universitari que està donant lloc a plataformes de vídeos docents com poliMèdia (Universitat Politècnica de València). transLectures és un projecte europeu que genera transcripcions i traduccions automàtiques d'alta qualitat per a la plataforma poliMèdia, mitjançant tècniques d'adaptació massiva i interacció intel·ligent. En aquest article presentem l'avaluació amb professors que es realitzà en el marc del pla d'acció Docència en Xarxa 2012-2013, amb l'objectiu d'estudiar el procés de supervisió de transcripcions, comparant-lo amb l'obtenció de la transcripció sense disposar d'una transcripció automàtica prèvia.}, keywords = {ASR, Docencia en Red, evaluaciones, Polimedia, transcripciones}, pubstate = {published}, tppubtype = {inproceedings} } [EN] Video lectures are a tool of proven value and wide acceptance in universities which is leading to video lecture platforms like poliMèdia (Universitat Politècnica de València). transLectures is an EU project that generates high-quality automatic transcriptions and translations for the poliMèdia platform, and improves them by using massive adaptation and intelligent interaction techniques. In this paper we present the evaluation with lecturers carried out under the Docència en Xarxa 2012-2013 action plan with the aim of studying the process of transcription post-editing, in contrast with transcribing from scratch. [ES] Los vídeos docentes son una herramienta de demostrada utilidad y gran aceptación en el mundo universitario que está dando lugar a plataformas de vídeos docentes como poliMèdia (Universitat Politècnica de València). transLectures es un proyecto europeo que genera transcripciones y traducciones automáticas de alta calidad para la plataforma poliMèdia, mediante técnicas de adaptación masiva e interacción inteligente. En este artículo presentamos la evaluación con profesores que se realizó en el marco del plan de acción Docència en Xarxa 2012-2013, con el objetivo de estudiar el proceso de supervisión de transcripciones, comparándolo con la obtención de la transcripción sin disponer de una transcripción automática previa. [CA] Els vídeos docents són una eina d'utilitat demostrada i amb gran acceptació en el món universitari que està donant lloc a plataformes de vídeos docents com poliMèdia (Universitat Politècnica de València). transLectures és un projecte europeu que genera transcripcions i traduccions automàtiques d'alta qualitat per a la plataforma poliMèdia, mitjançant tècniques d'adaptació massiva i interacció intel·ligent. En aquest article presentem l'avaluació amb professors que es realitzà en el marc del pla d'acció Docència en Xarxa 2012-2013, amb l'objectiu d'estudiar el procés de supervisió de transcripcions, comparant-lo amb l'obtenció de la transcripció sense disposar d'una transcripció automàtica prèvia. |
Martínez-Villaronga, A; del-Agua, M A; Silvestre-Cerdà, J A; Andrés-Ferrer, J; Juan, A Language model adaptation for lecture transcription by document retrieval Inproceedings Proc. of VIII Jornadas en Tecnología del Habla and IV Iberian SLTech Workshop (IberSpeech 2014), Las Palmas de Gran Canaria (Spain), 2014. @inproceedings{MarAgu14, title = {Language model adaptation for lecture transcription by document retrieval}, author = {A. Martínez-Villaronga and M. A. del-Agua and J.A. Silvestre-Cerdà and J. Andrés-Ferrer and A. Juan}, url = {http://www.mllp.upv.es/wp-content/uploads/2015/04/ibsp14-cameraReady.pdf}, year = {2014}, date = {2014-01-01}, booktitle = {Proc. of VIII Jornadas en Tecnología del Habla and IV Iberian SLTech Workshop (IberSpeech 2014)}, address = {Las Palmas de Gran Canaria (Spain)}, keywords = {}, pubstate = {published}, tppubtype = {inproceedings} } |
Piqueras, S; del-Agua, M A; Giménez, A; Civera, J; Juan, A Statistical text-to-speech synthesis of Spanish subtitles Inproceedings Proc. of VIII Jornadas en Tecnología del Habla and IV Iberian SLTech Workshop (IberSpeech 2014), Las Palmas de Gran Canaria (Spain), 2014. @inproceedings{PiqAgu14, title = {Statistical text-to-speech synthesis of Spanish subtitles}, author = {S. Piqueras and M. A. del-Agua and A. Giménez and J. Civera and A. Juan}, url = {http://www.mllp.upv.es/wp-content/uploads/2015/04/paper3.pdf http://link.springer.com/chapter/10.1007%2F978-3-319-13623-3_5}, year = {2014}, date = {2014-01-01}, booktitle = {Proc. of VIII Jornadas en Tecnología del Habla and IV Iberian SLTech Workshop (IberSpeech 2014)}, address = {Las Palmas de Gran Canaria (Spain)}, keywords = {}, pubstate = {published}, tppubtype = {inproceedings} } |
Pérez-González-de-Martos, A; Silvestre-Cerdá, J A; Rihtar, M; Juan, A; Civera, J Using Automatic Speech Transcriptions in Lecture Recommendation Systems Inproceedings Proc. of VIII Jornadas en Tecnología del Habla and IV Iberian SLTech Workshop (IberSpeech 2014), Las Palmas de Gran Canaria (Spain), 2014. @inproceedings{PerSil14, title = {Using Automatic Speech Transcriptions in Lecture Recommendation Systems}, author = {A. Pérez-González-de-Martos and J. A. Silvestre-Cerdá and M. Rihtar and A. Juan and J. Civera}, url = {http://www.mllp.upv.es/wp-content/uploads/2015/04/lavie_is2014_camready1.pdf}, year = {2014}, date = {2014-01-01}, booktitle = {Proc. of VIII Jornadas en Tecnología del Habla and IV Iberian SLTech Workshop (IberSpeech 2014)}, address = {Las Palmas de Gran Canaria (Spain)}, keywords = {}, pubstate = {published}, tppubtype = {inproceedings} } |
del-Agua, M A; Giménez, A; Serrano, N; Andrés-Ferrer, J; Civera, J; Sanchis, A; Juan, A The transLectures-UPV toolkit Inproceedings Proc. of VIII Jornadas en Tecnología del Habla and IV Iberian SLTech Workshop (IberSpeech 2014), Las Palmas de Gran Canaria (Spain), 2014. @inproceedings{AguGim14, title = {The transLectures-UPV toolkit}, author = {M. A. del-Agua and A. Giménez and N. Serrano and J. Andrés-Ferrer and J. Civera and A. Sanchis and A. Juan}, url = {http://www.mllp.upv.es/wp-content/uploads/2015/04/IberSpeech2014-TLK-camready1.pdf}, year = {2014}, date = {2014-01-01}, booktitle = {Proc. of VIII Jornadas en Tecnología del Habla and IV Iberian SLTech Workshop (IberSpeech 2014)}, address = {Las Palmas de Gran Canaria (Spain)}, keywords = {}, pubstate = {published}, tppubtype = {inproceedings} } |
Wuebker, Joern; Ney, Hermann; Martínez-Villaronga, Adrià; Giménez, Adrià; Juan, Alfons; Servan, Christophe; Dymetman, Marc; Mirkin, Shachar Comparison of Data Selection Techniques for the Translation of Video Lectures Inproceedings Proc. of the Eleventh Biennial Conf. of the Association for Machine Translation in the Americas (AMTA-2014), pp. 193–207, Vancouver (Canada), 2014. @inproceedings{WueMarSer14, title = {Comparison of Data Selection Techniques for the Translation of Video Lectures}, author = {Joern Wuebker and Hermann Ney and Adrià Martínez-Villaronga and Adrià Giménez and Alfons Juan and Christophe Servan and Marc Dymetman and Shachar Mirkin}, url = {https://aclanthology.org/2014.amta-researchers.15/}, year = {2014}, date = {2014-01-01}, booktitle = {Proc. of the Eleventh Biennial Conf. of the Association for Machine Translation in the Americas (AMTA-2014)}, pages = {193--207}, address = {Vancouver (Canada)}, keywords = {}, pubstate = {published}, tppubtype = {inproceedings} } |
Serrano, Nicolás; Giménez, Adrià; Civera, Jorge; Sanchis, Alberto; Juan, Alfons Interactive Handwriting Recognition with Limited User effort Journal Article Intl. Journal on Document Analysis and Recognition (IJDAR), 17 , pp. 47–59, 2014. @article{Serrano14a, title = {Interactive Handwriting Recognition with Limited User effort}, author = {Nicolás Serrano and Adrià Giménez and Jorge Civera and Alberto Sanchis and Alfons Juan}, url = {http://dx.doi.org/10.1007/s10032-013-0204-5}, year = {2014}, date = {2014-01-01}, journal = {Intl. Journal on Document Analysis and Recognition (IJDAR)}, volume = {17}, pages = {47--59}, keywords = {}, pubstate = {published}, tppubtype = {article} } |
Giménez, Adrià; Khoury, Ihab; Andrés-Ferrer, Jesús; Juan, Alfons Handwriting word recognition using windowed Bernoulli HMMs Journal Article Pattern Recognition Letters, 35 (0), pp. 149–156, 2014, ISSN: 0167-8655, (Frontiers in Handwriting Processing). Links | BibTeX | Tags: Sliding window @article{Giménez2014149, title = {Handwriting word recognition using windowed Bernoulli HMMs}, author = {Adrià Giménez and Ihab Khoury and Jesús Andrés-Ferrer and Alfons Juan}, url = {http://dx.doi.org/10.1016/j.patrec.2012.09.002 http://hdl.handle.net/10251/37326}, issn = {0167-8655}, year = {2014}, date = {2014-01-01}, journal = {Pattern Recognition Letters}, volume = {35}, number = {0}, pages = {149--156}, note = {Frontiers in Handwriting Processing}, keywords = {Sliding window}, pubstate = {published}, tppubtype = {article} } |
Giménez, Adrià; Andrés-Ferrer, Jesús; Juan, Alfons Discriminative Bernoulli HMMs for isolated handwritten word recognition Journal Article Pattern Recognition Letters, 35 (0), pp. 157–168, 2014, ISSN: 0167-8655, (Frontiers in Handwriting Processing). @article{Giménez2014157, title = {Discriminative Bernoulli HMMs for isolated handwritten word recognition}, author = {Adrià Giménez and Jesús Andrés-Ferrer and Alfons Juan}, url = {http://dx.doi.org/10.1016/j.patrec.2013.05.016}, issn = {0167-8655}, year = {2014}, date = {2014-01-01}, journal = {Pattern Recognition Letters}, volume = {35}, number = {0}, pages = {157--168}, note = {Frontiers in Handwriting Processing}, keywords = {RIMES}, pubstate = {published}, tppubtype = {article} } |
Serrano, Nicolás; Civera, Jorge; Sanchis, Alberto; Juan, A Effective balancing error and user effort in interactive handwriting recognition Journal Article Pattern Recognition Letters, 37 , pp. 135–142, 2014. @article{Serrano14b, title = {Effective balancing error and user effort in interactive handwriting recognition}, author = {Nicolás Serrano and Jorge Civera and Alberto Sanchis and A. Juan}, url = {http://dx.doi.org/10.1016/j.patrec.2013.03.010}, year = {2014}, date = {2014-01-01}, journal = {Pattern Recognition Letters}, volume = {37}, pages = {135--142}, keywords = {}, pubstate = {published}, tppubtype = {article} } |
Valor Miró, Juan Daniel ; Spencer, R N; Pérez González de Martos, A; Garcés Díaz-Munío, G; Turró, C; Civera, J; Juan, A Evaluating intelligent interfaces for post-editing automatic transcriptions of online video lectures Journal Article Open Learning: The Journal of Open, Distance and e-Learning, 29 (1), pp. 72–85, 2014. Abstract | Links | BibTeX | Tags: @article{doi:10.1080/02680513.2014.909722, title = {Evaluating intelligent interfaces for post-editing automatic transcriptions of online video lectures}, author = {Valor Miró, Juan Daniel and Spencer, R.N. and Pérez González de Martos, A. and Garcés Díaz-Munío, G. and Turró, C. and Civera, J. and Juan, A.}, url = {http://hdl.handle.net/10251/55925 http://dx.doi.org/10.1080/02680513.2014.909722 http://www.mllp.upv.es/wp-content/uploads/2015/04/author_version.pdf}, year = {2014}, date = {2014-01-01}, journal = {Open Learning: The Journal of Open, Distance and e-Learning}, volume = {29}, number = {1}, pages = {72--85}, abstract = {[EN] Video lectures are fast becoming an everyday educational resource in higher education. They are being incorporated into existing university curricula around the world, while also emerging as a key component of the open education movement. In 2007 the Universitat Politècnica de València (UPV) implemented its poliMèdia lecture capture system for the creation and publication of quality educational video content and now has a collection of over 10,000 video objects. In 2011 it embarked on the EU-subsidised transLectures project to add automatic subtitles to these videos in both Spanish and other languages. By doing so, it allows access to their educational content by non-native speakers and the deaf and hard-of-hearing, as well as enabling advanced repository management functions. In this paper, following a short introduction to poliMèdia, transLectures and Docència en Xarxa, the UPV's action plan to boost the use of digital resources at the university, we will discuss the three-stage evaluation process carried out with the collaboration of UPV lecturers to find the best interaction protocol for the task of post-editing automatic subtitles. [CA] "Avaluació d'interfícies intel·ligents per a la postedició de transcripcions automàtiques de vídeos docents en línia": Els vídeos docents s'estan convertint en un recurs d'ús quotidià en l'educació superior. Estan entrant en els plans d'estudis universitaris de tot el món, al mateix temps que es defineixen com un component clau del moviment de l'educació lliure. L'any 2007, la Universitat Politècnica de València (UPV) implementà el seu sistema d'enregistrament de classes poliMèdia, que permet la producció i publicació de continguts audiovisuals d'alta qualitat, i que ja acumula més de 10.000 vídeos. L'any 2011, la UPV va entrar en el projecte europeu transLectures per a afegir subtítols automàtics a aquests vídeos en castellà, català i altres llengües. Així, es facilita l'accés a aquests continguts educatius per part de parlants d'altres llengües i de persones sordes o amb dificultats auditives, i també es proporcionen funcions avançades de gestió del repositori. En aquest article, després de presentar poliMèdia, transLectures i Docència en Xarxa (el pla d'acció de la UPV per a impulsar l'ús de recursos digitals en la universitat), explicarem el procés d'avaluació en tres fases que s'ha realitzat amb la col·laboració de professors de la UPV per a trobar el millor protocol d'interacció per a la postedició de subtítols automàtics.}, keywords = {}, pubstate = {published}, tppubtype = {article} } [EN] Video lectures are fast becoming an everyday educational resource in higher education. They are being incorporated into existing university curricula around the world, while also emerging as a key component of the open education movement. In 2007 the Universitat Politècnica de València (UPV) implemented its poliMèdia lecture capture system for the creation and publication of quality educational video content and now has a collection of over 10,000 video objects. In 2011 it embarked on the EU-subsidised transLectures project to add automatic subtitles to these videos in both Spanish and other languages. By doing so, it allows access to their educational content by non-native speakers and the deaf and hard-of-hearing, as well as enabling advanced repository management functions. In this paper, following a short introduction to poliMèdia, transLectures and Docència en Xarxa, the UPV's action plan to boost the use of digital resources at the university, we will discuss the three-stage evaluation process carried out with the collaboration of UPV lecturers to find the best interaction protocol for the task of post-editing automatic subtitles. [CA] "Avaluació d'interfícies intel·ligents per a la postedició de transcripcions automàtiques de vídeos docents en línia": Els vídeos docents s'estan convertint en un recurs d'ús quotidià en l'educació superior. Estan entrant en els plans d'estudis universitaris de tot el món, al mateix temps que es defineixen com un component clau del moviment de l'educació lliure. L'any 2007, la Universitat Politècnica de València (UPV) implementà el seu sistema d'enregistrament de classes poliMèdia, que permet la producció i publicació de continguts audiovisuals d'alta qualitat, i que ja acumula més de 10.000 vídeos. L'any 2011, la UPV va entrar en el projecte europeu transLectures per a afegir subtítols automàtics a aquests vídeos en castellà, català i altres llengües. Així, es facilita l'accés a aquests continguts educatius per part de parlants d'altres llengües i de persones sordes o amb dificultats auditives, i també es proporcionen funcions avançades de gestió del repositori. En aquest article, després de presentar poliMèdia, transLectures i Docència en Xarxa (el pla d'acció de la UPV per a impulsar l'ús de recursos digitals en la universitat), explicarem el procés d'avaluació en tres fases que s'ha realitzat amb la col·laboració de professors de la UPV per a trobar el millor protocol d'interacció per a la postedició de subtítols automàtics. |
2013 |
Khoury, Ihab ; Giménez, Adrià ; Andrés-Ferrer, Jesús ; Juan, Alfons ; Sánchez, Joan Andreu The UPV Handwriting Recognition and Translation System for OpenHaRT 2013 Inproceedings Proc. of the NIST Open Handwriting Recognition and Translation Evaluation Workshop (OpenHaRT 2013), Washington DC (USA), 2013. Links | BibTeX | Tags: Arabic HTR, Bernoulli HMM, NIST OpenHaRT, Repositioning, Sliding window @inproceedings{Khoury2013, title = {The UPV Handwriting Recognition and Translation System for OpenHaRT 2013}, author = {Khoury, Ihab and Giménez, Adrià and Andrés-Ferrer, Jesús and Juan, Alfons and Sánchez, Joan Andreu}, url = {http://www.nist.gov/itl/iad/mig/upload/OpenHaRT2013_SysDesc_UPV.pdf}, year = {2013}, date = {2013-01-01}, booktitle = {Proc. of the NIST Open Handwriting Recognition and Translation Evaluation Workshop (OpenHaRT 2013)}, address = {Washington DC (USA)}, keywords = {Arabic HTR, Bernoulli HMM, NIST OpenHaRT, Repositioning, Sliding window}, pubstate = {published}, tppubtype = {inproceedings} } |
Silvestre-Cerdà, Joan Albert; Pérez, Alejandro; Jiménez, Manuel; Turró, Carlos; Juan, Alfons; Civera, Jorge A System Architecture to Support Cost-Effective Transcription and Translation of Large Video Lecture Repositories Inproceedings Proc. of the IEEE Intl. Conf. on Systems, Man, and Cybernetics SMC 2013 , pp. 3994-3999, Manchester (UK), 2013. Abstract | Links | BibTeX | Tags: Accessibility, Automatic Speech Recognition, Education, Intelligent Interaction, Language Technologies, Machine Translation, Massive Adaptation, Multilingualism, Opencast Matterhorn, Video Lectures @inproceedings{Silvestre-Cerdà2013, title = {A System Architecture to Support Cost-Effective Transcription and Translation of Large Video Lecture Repositories}, author = {Joan Albert Silvestre-Cerdà and Alejandro Pérez and Manuel Jiménez and Carlos Turró and Alfons Juan and Jorge Civera}, url = {http://dx.doi.org/10.1109/SMC.2013.682}, year = {2013}, date = {2013-01-01}, booktitle = {Proc. of the IEEE Intl. Conf. on Systems, Man, and Cybernetics SMC 2013 }, pages = {3994-3999}, address = {Manchester (UK)}, abstract = {Online video lecture repositories are rapidly growing and becoming established as fundamental knowledge assets. However, most lectures are neither transcribed nor translated because of the lack of cost-effective solutions that can give accurate enough results. In this paper, we describe a system architecture that supports the cost-effective transcription and translation of large video lecture repositories. This architecture has been adopted in the EU project transLectures and is now being tested on a repository of more than 9000 video lectures at the Universitat Politecnica de Valencia. Following a brief description of this repository and of the transLectures project, we describe the proposed system architecture in detail. We also report empirical results on the quality of the transcriptions and translations currently being maintained and steadily improved.}, keywords = {Accessibility, Automatic Speech Recognition, Education, Intelligent Interaction, Language Technologies, Machine Translation, Massive Adaptation, Multilingualism, Opencast Matterhorn, Video Lectures}, pubstate = {published}, tppubtype = {inproceedings} } Online video lecture repositories are rapidly growing and becoming established as fundamental knowledge assets. However, most lectures are neither transcribed nor translated because of the lack of cost-effective solutions that can give accurate enough results. In this paper, we describe a system architecture that supports the cost-effective transcription and translation of large video lecture repositories. This architecture has been adopted in the EU project transLectures and is now being tested on a repository of more than 9000 video lectures at the Universitat Politecnica de Valencia. Following a brief description of this repository and of the transLectures project, we describe the proposed system architecture in detail. We also report empirical results on the quality of the transcriptions and translations currently being maintained and steadily improved. |
Alkhoury, Ihab; Giménez, Adrià; Juan, Alfons; Andrés-Ferrer, Jesús Arabic Printed Word Recognition Using Windowed Bernoulli HMMs Inproceedings Proc. of the 17th Intl. Conf. on Image, Analysis and Processings (ICIAP 2013), pp. 330 – 339, Naples (Italy), 2013. @inproceedings{khoury13a, title = {Arabic Printed Word Recognition Using Windowed Bernoulli HMMs}, author = {Ihab Alkhoury and Adrià Giménez and Alfons Juan and Jesús Andrés-Ferrer}, url = {http://dx.doi.org/10.1007/978-3-642-41181-6_34}, year = {2013}, date = {2013-01-01}, booktitle = {Proc. of the 17th Intl. Conf. on Image, Analysis and Processings (ICIAP 2013)}, pages = {330 -- 339}, address = {Naples (Italy)}, keywords = {}, pubstate = {published}, tppubtype = {inproceedings} } |
Martínez-Villaronga, A; del Agua, M A; Andrés-Ferrer, J; Juan, A Language model adaptation for video lectures transcription Inproceedings Proc. of the IEEE Intl. Conf. on Acoustics, Speech and Signal Processing ICASSP 2013, pp. 8450-8454, Vancouver (Canada), 2013. Links | BibTeX | Tags: language model adaptation, Video Lectures @inproceedings{Martinez-Villaronga2013, title = {Language model adaptation for video lectures transcription}, author = {A. Martínez-Villaronga and M.A. del Agua and J. Andrés-Ferrer and A. Juan}, url = {http://dx.doi.org/10.1109/ICASSP.2013.6639314}, year = {2013}, date = {2013-01-01}, booktitle = {Proc. of the IEEE Intl. Conf. on Acoustics, Speech and Signal Processing ICASSP 2013}, pages = {8450-8454}, address = {Vancouver (Canada)}, keywords = {language model adaptation, Video Lectures}, pubstate = {published}, tppubtype = {inproceedings} } |
2012 |
Silvestre-Cerdà, Joan Albert ; Del Agua, Miguel ; Garcés, Gonçal; Gascó, Guillem; Giménez-Pastor, Adrià; Martínez, Adrià; Pérez González de Martos, Alejandro ; Sánchez, Isaías; Serrano Martínez-Santos, Nicolás ; Spencer, Rachel; Valor Miró, Juan Daniel ; Andrés-Ferrer, Jesús; Civera, Jorge; Sanchís, Alberto; Juan, Alfons transLectures Inproceedings Proceedings (Online) of IberSPEECH 2012, pp. 345–351, Madrid (Spain), 2012. Abstract | Links | BibTeX | Tags: Accessibility, Automatic Speech Recognition, Education, Intelligent Interaction, Language Technologies, Machine Translation, Massive Adaptation, Multilingualism, Opencast Matterhorn, Video Lectures @inproceedings{Silvestre-Cerdà2012b, title = {transLectures}, author = {Silvestre-Cerdà, Joan Albert and Del Agua, Miguel and Gonçal Garcés and Guillem Gascó and Adrià Giménez-Pastor and Adrià Martínez and Pérez González de Martos, Alejandro and Isaías Sánchez and Serrano Martínez-Santos, Nicolás and Rachel Spencer and Valor Miró, Juan Daniel and Jesús Andrés-Ferrer and Jorge Civera and Alberto Sanchís and Alfons Juan}, url = {http://hdl.handle.net/10251/37290 http://lorien.die.upm.es/~lapiz/rtth/JORNADAS/VII/IberSPEECH2012_OnlineProceedings.pdf https://web.archive.org/web/20130609073144/http://iberspeech2012.ii.uam.es/IberSPEECH2012_OnlineProceedings.pdf http://www.mllp.upv.es/wp-content/uploads/2015/04/1209IberSpeech.pdf}, year = {2012}, date = {2012-11-22}, booktitle = {Proceedings (Online) of IberSPEECH 2012}, pages = {345--351}, address = {Madrid (Spain)}, abstract = {[EN] transLectures (Transcription and Translation of Video Lectures) is an EU STREP project in which advanced automatic speech recognition and machine translation techniques are being tested on large video lecture repositories. The project began in November 2011 and will run for three years. This paper will outline the project's main motivation and objectives, and give a brief description of the two main repositories being considered: VideoLectures.NET and poliMèdia. The first results obtained by the UPV group for the poliMedia repository will also be provided. [CA] transLectures (Transcription and Translation of Video Lectures) és un projecte del 7PM de la Unió Europea en el qual s'estan posant a prova tècniques avançades de reconeixement automàtic de la parla i de traducció automàtica sobre grans repositoris digitals de vídeos docents. El projecte començà al novembre de 2011 i tindrà una duració de tres anys. En aquest article exposem la motivació i els objectius del projecte, i descrivim breument els dos repositoris principals sobre els quals es treballa: VideoLectures.NET i poliMèdia. També oferim els primers resultats obtinguts per l'equip de la UPV al repositori poliMèdia.}, keywords = {Accessibility, Automatic Speech Recognition, Education, Intelligent Interaction, Language Technologies, Machine Translation, Massive Adaptation, Multilingualism, Opencast Matterhorn, Video Lectures}, pubstate = {published}, tppubtype = {inproceedings} } [EN] transLectures (Transcription and Translation of Video Lectures) is an EU STREP project in which advanced automatic speech recognition and machine translation techniques are being tested on large video lecture repositories. The project began in November 2011 and will run for three years. This paper will outline the project's main motivation and objectives, and give a brief description of the two main repositories being considered: VideoLectures.NET and poliMèdia. The first results obtained by the UPV group for the poliMedia repository will also be provided. [CA] transLectures (Transcription and Translation of Video Lectures) és un projecte del 7PM de la Unió Europea en el qual s'estan posant a prova tècniques avançades de reconeixement automàtic de la parla i de traducció automàtica sobre grans repositoris digitals de vídeos docents. El projecte començà al novembre de 2011 i tindrà una duració de tres anys. En aquest article exposem la motivació i els objectius del projecte, i descrivim breument els dos repositoris principals sobre els quals es treballa: VideoLectures.NET i poliMèdia. També oferim els primers resultats obtinguts per l'equip de la UPV al repositori poliMèdia. |
Silvestre-Cerdà, Joan Albert; Giménez, Adrià; Andrés-Ferrer, Jesús; Civera, Jorge; Juan, Alfons Albayzin Evaluation: The PRHLT-UPV Audio Segmentation System Inproceedings Proceedings (Online) of IberSPEECH 2012, pp. 596-600, Madrid (Spain), 2012. Abstract | Links | BibTeX | Tags: @inproceedings{Silvestre-Cerdà2012c, title = {Albayzin Evaluation: The PRHLT-UPV Audio Segmentation System}, author = {Joan Albert Silvestre-Cerdà and Adrià Giménez and Jesús Andrés-Ferrer and Jorge Civera and Alfons Juan}, url = {http://hdl.handle.net/10251/53699 http://iberspeech2012.ii.uam.es/IberSPEECH2012_OnlineProceedings.pdf}, year = {2012}, date = {2012-11-22}, booktitle = {Proceedings (Online) of IberSPEECH 2012}, pages = {596-600}, address = {Madrid (Spain)}, abstract = {This paper describes the audio segmentation system developed by the PRHLT research group at the UPV for the Albayzin Audio Segmentation Evaluation 2012. The PRHLT-UPV audio segmentation system is based on a conventional GMM-HMM speech recognition approach in which the vocabulary set is defined by the power set of segment classes. MFCC features were extracted to represent the acoustic signal and the AK toolkit was used for both, training acoustic models and performing audio segmentation. Experimental results reveals that our system provides an excellent performance on speech detection, so it could be successfully employed to provide speech segments to a diarization or speech recognition system.}, keywords = {}, pubstate = {published}, tppubtype = {inproceedings} } This paper describes the audio segmentation system developed by the PRHLT research group at the UPV for the Albayzin Audio Segmentation Evaluation 2012. The PRHLT-UPV audio segmentation system is based on a conventional GMM-HMM speech recognition approach in which the vocabulary set is defined by the power set of segment classes. MFCC features were extracted to represent the acoustic signal and the AK toolkit was used for both, training acoustic models and performing audio segmentation. Experimental results reveals that our system provides an excellent performance on speech detection, so it could be successfully employed to provide speech segments to a diarization or speech recognition system. |
Toselli, Alejandro H; Serrano, Nicolás ; Giménez-Pastor, Adrià ; Khoury, Ihab ; Juan, Alfons ; Vidal, Enrique Language Technology for Handwritten Text Recognition Incollection Advances in Speech and Language Technologies for Iberian Languages (IberSpeech 2012), 328 , pp. 178-186, Springer Berlin Heidelberg, 2012. Links | BibTeX | Tags: Hidden Markov Model Emission Probability, Mixture of Bernoulli Distributions, Mixture of Gaussian Densities, Off-Line Continuous Handwritten Text Recognition @incollection{Toselli2012, title = {Language Technology for Handwritten Text Recognition}, author = {Toselli, Alejandro H. and Serrano, Nicolás and Giménez-Pastor, Adrià and Khoury, Ihab and Juan, Alfons and Vidal, Enrique}, url = {http://hdl.handle.net/10251/36452}, year = {2012}, date = {2012-01-01}, booktitle = {Advances in Speech and Language Technologies for Iberian Languages (IberSpeech 2012)}, volume = {328}, pages = {178-186}, publisher = {Springer Berlin Heidelberg}, keywords = {Hidden Markov Model Emission Probability, Mixture of Bernoulli Distributions, Mixture of Gaussian Densities, Off-Line Continuous Handwritten Text Recognition}, pubstate = {published}, tppubtype = {incollection} } |
Doetsch, Patrick; Hamdani, Mahdi; Giménez-Pastor, Adrià; Andrés-Ferrer, Jesús; Juan, Alfons; Ney, Hermann Comparison of Bernoulli and Gaussian HMMs using a vertical repositioning technique for off-line handwriting recognition Inproceedings Proc. of the 2012 Intl. Conf. on Frontiers in Handwriting Recognition (ICFHR 2012), pp. 3 – 7, 2012. BibTeX | Tags: @inproceedings{doetsch12, title = {Comparison of Bernoulli and Gaussian HMMs using a vertical repositioning technique for off-line handwriting recognition}, author = {Patrick Doetsch and Mahdi Hamdani and Adrià Giménez-Pastor and Jesús Andrés-Ferrer and Alfons Juan and Hermann Ney}, year = {2012}, date = {2012-01-01}, booktitle = {Proc. of the 2012 Intl. Conf. on Frontiers in Handwriting Recognition (ICFHR 2012)}, pages = {3 -- 7}, keywords = {}, pubstate = {published}, tppubtype = {inproceedings} } |
Khoury, Ihab; Giménez-Pastor, Adrià; Juan, Alfons Guide to OCR for Arabic Scripts Book Chapter Märgner Volkerand El Abed, Haikal (Ed.): Chapter Arabic Handwriting Recognition Using Ber, pp. 255-272, Springer, 2012. BibTeX | Tags: @inbook{Khoury12-AHRBook-chp10, title = {Guide to OCR for Arabic Scripts}, author = {Ihab Khoury and Adrià Giménez-Pastor and Alfons Juan}, editor = {Märgner, Volkerand El Abed, Haikal }, year = {2012}, date = {2012-01-01}, pages = {255-272}, publisher = {Springer}, chapter = {Arabic Handwriting Recognition Using Ber}, keywords = {}, pubstate = {published}, tppubtype = {inbook} } |
Valor Miró, Juan Daniel ; Pérez González de Martos, Alejandro ; Civera, Jorge ; Juan, Alfons Integrating a State-of-the-Art ASR System into the Opencast Matterhorn Platform Incollection Advances in Speech and Language Technologies for Iberian Languages (IberSpeech 2012), 328 , pp. 237-246, Springer Berlin Heidelberg, 2012, ISBN: 978-3-642-35291-1, (doi: 10.1007/978-3-642-35292-8_20). Links | BibTeX | Tags: Google N-Gram, Language Modeling, Linear Combination, Opencast Matterhorn, Speech Recognition @incollection{Valor2012, title = {Integrating a State-of-the-Art ASR System into the Opencast Matterhorn Platform}, author = {Valor Miró, Juan Daniel and Pérez González de Martos, Alejandro and Civera, Jorge and Juan, Alfons}, url = {http://hdl.handle.net/10251/35190 http://www.mllp.upv.es/wp-content/uploads/2015/04/paper2.pdf}, isbn = {978-3-642-35291-1}, year = {2012}, date = {2012-01-01}, booktitle = {Advances in Speech and Language Technologies for Iberian Languages (IberSpeech 2012)}, volume = {328}, pages = {237-246}, publisher = {Springer Berlin Heidelberg}, series = {Communications in Computer and Information Science}, note = {doi: 10.1007/978-3-642-35292-8_20}, keywords = {Google N-Gram, Language Modeling, Linear Combination, Opencast Matterhorn, Speech Recognition}, pubstate = {published}, tppubtype = {incollection} } |
Sánchez-Cortina, Isaías; Serrano, Nicolás; Sanchis, Alberto; Juan, Alfons A Prototype for Interactive Speech Transcription Balancing Error and Supervision Effort Inproceedings Proc. of the 2012 ACM Intl. Conf. on Intelligent User Interfaces, pp. 325–326, 2012. BibTeX | Tags: @inproceedings{b, title = {A Prototype for Interactive Speech Transcription Balancing Error and Supervision Effort}, author = {Isaías Sánchez-Cortina and Nicolás Serrano and Alberto Sanchis and Alfons Juan}, year = {2012}, date = {2012-01-01}, booktitle = {Proc. of the 2012 ACM Intl. Conf. on Intelligent User Interfaces}, pages = {325--326}, keywords = {}, pubstate = {published}, tppubtype = {inproceedings} } |
Publications
2016 |
Speaker-adapted confidence measures for speech recognition of video lectures Journal Article Computer Speech & Language, 37 , pp. 11–23, 2016, ISBN: 0885-2308. |
2015 |
The MLLP ASR Systems for IWSLT 2015 Inproceedings Proc. of 12th Intl. Workshop on Spoken Language Translation (IWSLT 2015), pp. 39–44, Da Nang (Vietnam), 2015. |
Efficient Generation of High-Quality Multilingual Subtitles for Video Lecture Repositories Inproceedings Proc. of 10th European Conf. on Technology Enhanced Learning (EC-TEL 2015), pp. 485–490, Toledo (Spain), 2015, ISBN: 978-3-319-24258-3. |
MLLP Transcription and Translation Platform Miscellaneous 2015, (Short paper for demo presentation accepted at 10th European Conf. on Technology Enhanced Learning (EC-TEL 2015), Toledo (Spain), 2015.). |
Evaluación de la revisión de transcripciones y traducciones automáticas de vídeos poliMedia Inproceedings Proc. of I Congreso Nacional de Innovación Educativa y Docencia en Red (IN-RED 2015), pp. 464–468, València (Spain), 2015. |
Efficiency and usability study of innovative computer-aided transcription strategies for video lecture repositories Journal Article Speech Communication, 74 , pp. 65–75, 2015, ISSN: 0167-6393. |
Supporting language diversity of European MOOCs with the EMMA platform Inproceedings Proc. of the European MOOC Stakeholder Summit EMOOCs 2015, pp. 157–165, Mons (Belgium), 2015. |
Window Repositioning for Printed Arabic Recognition Journal Article Pattern Recognition Letters, 51 , pp. 86–93, 2015, ISSN: 0167-8655. |
2014 |
Evaluación del proceso de revisión de transcripciones automáticas para vídeos Polimedia Inproceedings Proc. of I Jornadas de Innovación Educativa y Docencia en Red (IN-RED 2014), pp. 272–278, Valencia (Spain), 2014. |
Language model adaptation for lecture transcription by document retrieval Inproceedings Proc. of VIII Jornadas en Tecnología del Habla and IV Iberian SLTech Workshop (IberSpeech 2014), Las Palmas de Gran Canaria (Spain), 2014. |
Statistical text-to-speech synthesis of Spanish subtitles Inproceedings Proc. of VIII Jornadas en Tecnología del Habla and IV Iberian SLTech Workshop (IberSpeech 2014), Las Palmas de Gran Canaria (Spain), 2014. |
Using Automatic Speech Transcriptions in Lecture Recommendation Systems Inproceedings Proc. of VIII Jornadas en Tecnología del Habla and IV Iberian SLTech Workshop (IberSpeech 2014), Las Palmas de Gran Canaria (Spain), 2014. |
The transLectures-UPV toolkit Inproceedings Proc. of VIII Jornadas en Tecnología del Habla and IV Iberian SLTech Workshop (IberSpeech 2014), Las Palmas de Gran Canaria (Spain), 2014. |
Comparison of Data Selection Techniques for the Translation of Video Lectures Inproceedings Proc. of the Eleventh Biennial Conf. of the Association for Machine Translation in the Americas (AMTA-2014), pp. 193–207, Vancouver (Canada), 2014. |
Interactive Handwriting Recognition with Limited User effort Journal Article Intl. Journal on Document Analysis and Recognition (IJDAR), 17 , pp. 47–59, 2014. |
Handwriting word recognition using windowed Bernoulli HMMs Journal Article Pattern Recognition Letters, 35 (0), pp. 149–156, 2014, ISSN: 0167-8655, (Frontiers in Handwriting Processing). |
Discriminative Bernoulli HMMs for isolated handwritten word recognition Journal Article Pattern Recognition Letters, 35 (0), pp. 157–168, 2014, ISSN: 0167-8655, (Frontiers in Handwriting Processing). |
Effective balancing error and user effort in interactive handwriting recognition Journal Article Pattern Recognition Letters, 37 , pp. 135–142, 2014. |
Evaluating intelligent interfaces for post-editing automatic transcriptions of online video lectures Journal Article Open Learning: The Journal of Open, Distance and e-Learning, 29 (1), pp. 72–85, 2014. |
2013 |
The UPV Handwriting Recognition and Translation System for OpenHaRT 2013 Inproceedings Proc. of the NIST Open Handwriting Recognition and Translation Evaluation Workshop (OpenHaRT 2013), Washington DC (USA), 2013. |
A System Architecture to Support Cost-Effective Transcription and Translation of Large Video Lecture Repositories Inproceedings Proc. of the IEEE Intl. Conf. on Systems, Man, and Cybernetics SMC 2013 , pp. 3994-3999, Manchester (UK), 2013. |
Arabic Printed Word Recognition Using Windowed Bernoulli HMMs Inproceedings Proc. of the 17th Intl. Conf. on Image, Analysis and Processings (ICIAP 2013), pp. 330 – 339, Naples (Italy), 2013. |
Language model adaptation for video lectures transcription Inproceedings Proc. of the IEEE Intl. Conf. on Acoustics, Speech and Signal Processing ICASSP 2013, pp. 8450-8454, Vancouver (Canada), 2013. |
2012 |
transLectures Inproceedings Proceedings (Online) of IberSPEECH 2012, pp. 345–351, Madrid (Spain), 2012. |
Albayzin Evaluation: The PRHLT-UPV Audio Segmentation System Inproceedings Proceedings (Online) of IberSPEECH 2012, pp. 596-600, Madrid (Spain), 2012. |
Language Technology for Handwritten Text Recognition Incollection Advances in Speech and Language Technologies for Iberian Languages (IberSpeech 2012), 328 , pp. 178-186, Springer Berlin Heidelberg, 2012. |
Comparison of Bernoulli and Gaussian HMMs using a vertical repositioning technique for off-line handwriting recognition Inproceedings Proc. of the 2012 Intl. Conf. on Frontiers in Handwriting Recognition (ICFHR 2012), pp. 3 – 7, 2012. |
Guide to OCR for Arabic Scripts Book Chapter Märgner Volkerand El Abed, Haikal (Ed.): Chapter Arabic Handwriting Recognition Using Ber, pp. 255-272, Springer, 2012. |
Integrating a State-of-the-Art ASR System into the Opencast Matterhorn Platform Incollection Advances in Speech and Language Technologies for Iberian Languages (IberSpeech 2012), 328 , pp. 237-246, Springer Berlin Heidelberg, 2012, ISBN: 978-3-642-35291-1, (doi: 10.1007/978-3-642-35292-8_20). |
A Prototype for Interactive Speech Transcription Balancing Error and Supervision Effort Inproceedings Proc. of the 2012 ACM Intl. Conf. on Intelligent User Interfaces, pp. 325–326, 2012. |