2016 |
Valor Miró, Juan Daniel ; Turró, C; Civera, J; Juan, A Generación eficiente de transcripciones y traducciones automáticas en poliMedia Inproceedings Proc. of II Congreso Nacional de Innovación Educativa y Docencia en Red (IN-RED 2016), pp. 21–29, València (Spain), 2016. Abstract | Links | BibTeX | Tags: Docencia en Red, e-learning, transcription, translation, video @inproceedings{Valor2016, title = {Generación eficiente de transcripciones y traducciones automáticas en poliMedia}, author = {Valor Miró, Juan Daniel and Turró, C. and Civera, J. and Juan, A.}, url = {http://dx.doi.org/10.4995/INRED2016.2016.4276}, year = {2016}, date = {2016-01-01}, booktitle = {Proc. of II Congreso Nacional de Innovación Educativa y Docencia en Red (IN-RED 2016)}, pages = {21--29}, address = {València (Spain)}, abstract = {The use of educational videos in higher education has increased quickly for several educational applications, which leads to platforms and services such as poliMèdia at the Universitat Politècnica de València (UPV), which enables the creation, publication and dissemination of this educational multimedia content. Through various research projects, and specifically the EU project transLectures, the UPV has implemented a system that automatically generates subtitles in various languages for all poliMèdia videos. these subtitles are created by an automatic speech recognition and machine translation system that provides high accuracy in both recognition and translation of the main European languages. Transcriptions and translations are not only used to improve accessibility, but also enable the search and retrieval of video contents within the video portal. Thus, a user can locate the video, and the time within it, where a certain word is said for later viewing. In this article we also extend previous work in the assessment of the review process, including transcription of French and translation of Spanish into Catalan.}, keywords = {Docencia en Red, e-learning, transcription, translation, video}, pubstate = {published}, tppubtype = {inproceedings} } The use of educational videos in higher education has increased quickly for several educational applications, which leads to platforms and services such as poliMèdia at the Universitat Politècnica de València (UPV), which enables the creation, publication and dissemination of this educational multimedia content. Through various research projects, and specifically the EU project transLectures, the UPV has implemented a system that automatically generates subtitles in various languages for all poliMèdia videos. these subtitles are created by an automatic speech recognition and machine translation system that provides high accuracy in both recognition and translation of the main European languages. Transcriptions and translations are not only used to improve accessibility, but also enable the search and retrieval of video contents within the video portal. Thus, a user can locate the video, and the time within it, where a certain word is said for later viewing. In this article we also extend previous work in the assessment of the review process, including transcription of French and translation of Spanish into Catalan. |
del-Agua, Miguel Ángel; Martínez-Villaronga, Adrià; Giménez, Adrià; Sanchis, Alberto; Civera, Jorge; Juan, Alfons The MLLP system for the 4th CHiME Challenge Inproceedings Proc. of the 4th Intl. Workshop on Speech Processing in Everyday Environments (CHiME 2016), pp. 57–59, San Francisco (USA), 2016. Abstract | Links | BibTeX | Tags: @inproceedings{del-Aguadel-Agua2016, title = {The MLLP system for the 4th CHiME Challenge}, author = {Miguel Ángel del-Agua and Adrià Martínez-Villaronga and Adrià Giménez and Alberto Sanchis and Jorge Civera and Alfons Juan}, url = {http://www.mllp.upv.es/wp-content/uploads/2017/11/DelAgua2016-The_MLLP_system_for_the_4th_CHiME_Challenge.pdf http://hdl.handle.net/10251/177497 http://spandh.dcs.shef.ac.uk/chime_workshop/chime2016/chime2016proceedings.pdf}, year = {2016}, date = {2016-01-01}, booktitle = {Proc. of the 4th Intl. Workshop on Speech Processing in Everyday Environments (CHiME 2016)}, pages = {57--59}, address = {San Francisco (USA)}, abstract = {The MLLP's CHiME-4 system is presented in this paper. It has been built using the transLectures-UPV toolkit (TLK), developed by the MLLP research group, which makes use of state-of-the-art speech techniques. Our best system built for the CHiME-4 challenge consists on the combination of different sub-systems in order to deal with the variety of acoustic conditions. Each sub-system in turn, follows a hybrid approach with different acoustic models, such as Deep Neural Networks or BLSTM Networks.}, keywords = {}, pubstate = {published}, tppubtype = {inproceedings} } The MLLP's CHiME-4 system is presented in this paper. It has been built using the transLectures-UPV toolkit (TLK), developed by the MLLP research group, which makes use of state-of-the-art speech techniques. Our best system built for the CHiME-4 challenge consists on the combination of different sub-systems in order to deal with the variety of acoustic conditions. Each sub-system in turn, follows a hybrid approach with different acoustic models, such as Deep Neural Networks or BLSTM Networks. |
2015 |
del-Agua, Miguel Ángel; Martínez-Villaronga, Adrià; Piqueras, Santiago; Giménez, Adrià; Sanchis, Alberto; Civera, Jorge; Juan, Alfons The MLLP ASR Systems for IWSLT 2015 Inproceedings Proc. of 12th Intl. Workshop on Spoken Language Translation (IWSLT 2015), pp. 39–44, Da Nang (Vietnam), 2015. Abstract | Links | BibTeX | Tags: @inproceedings{delAgua15, title = {The MLLP ASR Systems for IWSLT 2015}, author = {Miguel Ángel del-Agua and Adrià Martínez-Villaronga and Santiago Piqueras and Adrià Giménez and Alberto Sanchis and Jorge Civera and Alfons Juan}, url = {https://aclanthology.org/2015.iwslt-evaluation.5/}, year = {2015}, date = {2015-12-03}, booktitle = {Proc. of 12th Intl. Workshop on Spoken Language Translation (IWSLT 2015)}, pages = {39--44}, address = {Da Nang (Vietnam)}, abstract = {This paper describes the Machine Learning and Language Processing (MLLP) ASR systems for the 2015 IWSLT evaluation campaing. The English system is based on the combination of five different subsystems which consist of two types of Neural Networks architectures (Deep feed-forward and Convolutional), two types of activation functions (sigmoid and rectified linear) and two types of input features (fMLLR and FBANK). All subsystems perform a speaker adaptation step based on confidence measures, the output of which is then combined with ROVER. This system achieves a Word Error Rate (WER) of 13.3% on the official IWSLT 2015 English test set.}, keywords = {}, pubstate = {published}, tppubtype = {inproceedings} } This paper describes the Machine Learning and Language Processing (MLLP) ASR systems for the 2015 IWSLT evaluation campaing. The English system is based on the combination of five different subsystems which consist of two types of Neural Networks architectures (Deep feed-forward and Convolutional), two types of activation functions (sigmoid and rectified linear) and two types of input features (fMLLR and FBANK). All subsystems perform a speaker adaptation step based on confidence measures, the output of which is then combined with ROVER. This system achieves a Word Error Rate (WER) of 13.3% on the official IWSLT 2015 English test set. |
Valor Miró, Juan Daniel ; Silvestre-Cerdà, Joan Albert ; Civera, Jorge ; Turró, Carlos ; Juan, Alfons Efficient Generation of High-Quality Multilingual Subtitles for Video Lecture Repositories Inproceedings Proc. of 10th European Conf. on Technology Enhanced Learning (EC-TEL 2015), pp. 485–490, Toledo (Spain), 2015, ISBN: 978-3-319-24258-3. Abstract | Links | BibTeX | Tags: Automatic Speech Recognition, Docencia en Red, Efficient video subtitling, Polimedia, Statistical machine translation, video lecture repositories @inproceedings{valor2015efficient, title = {Efficient Generation of High-Quality Multilingual Subtitles for Video Lecture Repositories}, author = {Valor Miró, Juan Daniel and Silvestre-Cerdà, Joan Albert and Civera, Jorge and Turró, Carlos and Juan, Alfons}, url = {http://link.springer.com/chapter/10.1007/978-3-319-24258-3_44 http://www.mllp.upv.es/wp-content/uploads/2016/03/paper.pdf }, isbn = {978-3-319-24258-3}, year = {2015}, date = {2015-09-17}, booktitle = {Proc. of 10th European Conf. on Technology Enhanced Learning (EC-TEL 2015)}, pages = {485--490}, address = {Toledo (Spain)}, abstract = {Video lectures are a valuable educational tool in higher education to support or replace face-to-face lectures in active learning strategies. In 2007 the Universitat Polit‘ecnica de Val‘encia (UPV) implemented its video lecture capture system, resulting in a high quality educational video repository, called poliMedia, with more than 10.000 mini lectures created by 1.373 lecturers. Also, in the framework of the European project transLectures, UPV has automatically generated transcriptions and translations in Spanish, Catalan and English for all videos included in the poliMedia video repository. transLectures’s objective responds to the widely-recognised need for subtitles to be provided with video lectures, as an essential service for non-native speakers and hearing impaired persons, and to allow advanced repository functionalities. Although high-quality automatic transcriptions and translations were generated in transLectures, they were not error-free. For this reason, lecturers need to manually review video subtitles to guarantee the absence of errors. The aim of this study is to evaluate the efficiency of the manual review process from automatic subtitles in comparison with the conventional generation of video subtitles from scratch. The reported results clearly indicate the convenience of providing automatic subtitles as a first step in the generation of video subtitles and the significant savings in time of up to almost 75% involved in reviewing subtitles.}, keywords = {Automatic Speech Recognition, Docencia en Red, Efficient video subtitling, Polimedia, Statistical machine translation, video lecture repositories}, pubstate = {published}, tppubtype = {inproceedings} } Video lectures are a valuable educational tool in higher education to support or replace face-to-face lectures in active learning strategies. In 2007 the Universitat Polit‘ecnica de Val‘encia (UPV) implemented its video lecture capture system, resulting in a high quality educational video repository, called poliMedia, with more than 10.000 mini lectures created by 1.373 lecturers. Also, in the framework of the European project transLectures, UPV has automatically generated transcriptions and translations in Spanish, Catalan and English for all videos included in the poliMedia video repository. transLectures’s objective responds to the widely-recognised need for subtitles to be provided with video lectures, as an essential service for non-native speakers and hearing impaired persons, and to allow advanced repository functionalities. Although high-quality automatic transcriptions and translations were generated in transLectures, they were not error-free. For this reason, lecturers need to manually review video subtitles to guarantee the absence of errors. The aim of this study is to evaluate the efficiency of the manual review process from automatic subtitles in comparison with the conventional generation of video subtitles from scratch. The reported results clearly indicate the convenience of providing automatic subtitles as a first step in the generation of video subtitles and the significant savings in time of up to almost 75% involved in reviewing subtitles. |
Pérez González de Martos, Alejandro ; Silvestre-Cerdà, Joan Albert ; Valor Miró, Juan Daniel ; Civera, Jorge ; Juan, Alfons MLLP Transcription and Translation Platform Miscellaneous 2015, (Short paper for demo presentation accepted at 10th European Conf. on Technology Enhanced Learning (EC-TEL 2015), Toledo (Spain), 2015.). Abstract | Links | BibTeX | Tags: Automatic Speech Recognition, Docencia en Red, Document translation, Efficient video subtitling, Machine Translation, MLLP, Post-editing, Video Lectures @misc{mllpttp, title = {MLLP Transcription and Translation Platform}, author = {Pérez González de Martos, Alejandro and Silvestre-Cerdà, Joan Albert and Valor Miró, Juan Daniel and Civera, Jorge and Juan, Alfons}, url = {http://hdl.handle.net/10251/65747 http://www.mllp.upv.es/wp-content/uploads/2015/09/ttp_platform_demo_ectel2015.pdf http://ectel2015.httc.de/index.php?id=722}, year = {2015}, date = {2015-09-16}, booktitle = {Tenth European Conference On Technology Enhanced Learning (EC-TEL 2015)}, abstract = {This paper briefly presents the main features of MLLP’s Transcription and Translation Platform, which uses state-of-the-art automatic speech recognition and machine translation systems to generate multilingual subtitles of educational audiovisual and textual content. It has proven to reduce user effort up to 1/3 of the time needed to generate transcriptions and translations from scratch.}, note = {Short paper for demo presentation accepted at 10th European Conf. on Technology Enhanced Learning (EC-TEL 2015), Toledo (Spain), 2015.}, keywords = {Automatic Speech Recognition, Docencia en Red, Document translation, Efficient video subtitling, Machine Translation, MLLP, Post-editing, Video Lectures}, pubstate = {published}, tppubtype = {misc} } This paper briefly presents the main features of MLLP’s Transcription and Translation Platform, which uses state-of-the-art automatic speech recognition and machine translation systems to generate multilingual subtitles of educational audiovisual and textual content. It has proven to reduce user effort up to 1/3 of the time needed to generate transcriptions and translations from scratch. |
Valor Miró, Juan Daniel ; Turró, C; Civera, J; Juan, A Evaluación de la revisión de transcripciones y traducciones automáticas de vídeos poliMedia Inproceedings Proc. of I Congreso Nacional de Innovación Educativa y Docencia en Red (IN-RED 2015), pp. 464–468, València (Spain), 2015. Links | BibTeX | Tags: Docencia en Red, evaluaciones con usuario, Polimedia, traducciones, transcripciones @inproceedings{Valor-InRed2015, title = {Evaluación de la revisión de transcripciones y traducciones automáticas de vídeos poliMedia}, author = {Valor Miró, Juan Daniel and Turró, C. and Civera, J. and Juan, A.}, url = {http://hdl.handle.net/10251/52755 http://www.mllp.upv.es/wp-content/uploads/2015/06/1574-3087-1-PB.pdf}, year = {2015}, date = {2015-06-30}, booktitle = {Proc. of I Congreso Nacional de Innovación Educativa y Docencia en Red (IN-RED 2015)}, pages = {464--468}, address = {València (Spain)}, keywords = {Docencia en Red, evaluaciones con usuario, Polimedia, traducciones, transcripciones}, pubstate = {published}, tppubtype = {inproceedings} } |
Valor Miró, Juan Daniel ; Silvestre-Cerdà, Joan Albert; Civera, Jorge; Turró, Carlos; Juan, Alfons Efficiency and usability study of innovative computer-aided transcription strategies for video lecture repositories Journal Article Speech Communication, 74 , pp. 65–75, 2015, ISSN: 0167-6393. Abstract | Links | BibTeX | Tags: Automatic Speech Recognition, Computer-assisted transcription, Interface design strategies, Usability study, video lecture repositories @article{Valor201565, title = {Efficiency and usability study of innovative computer-aided transcription strategies for video lecture repositories}, author = {Valor Miró, Juan Daniel and Joan Albert Silvestre-Cerdà and Jorge Civera and Carlos Turró and Alfons Juan}, url = {http://www.sciencedirect.com/science/article/pii/S0167639315001016 http://www.mllp.upv.es/wp-content/uploads/2016/03/paper1.pdf}, issn = {0167-6393}, year = {2015}, date = {2015-01-01}, journal = {Speech Communication}, volume = {74}, pages = {65--75}, abstract = {Abstract Video lectures are widely used in education to support and complement face-to-face lectures. However, the utility of these audiovisual assets could be further improved by adding subtitles that can be exploited to incorporate added-value functionalities such as searchability, accessibility, translatability, note-taking, and discovery of content-related videos, among others. Today, automatic subtitles are prone to error, and need to be reviewed and post-edited in order to ensure that what students see on-screen are of an acceptable quality. This work investigates different user interface design strategies for this post-editing task to discover the best way to incorporate automatic transcription technologies into large educational video repositories. Our three-phase study involved lecturers from the Universitat Politècnica de València (UPV) with videos available on the poliMedia video lecture repository, which is currently over 10,000 video objects. Simply by conventional post-editing automatic transcriptions users almost reduced to half the time that would require to generate the transcription from scratch. As expected, this study revealed that the time spent by lecturers reviewing automatic transcriptions correlated directly with the accuracy of said transcriptions. However, it is also shown that the average time required to perform each individual editing operation could be precisely derived and could be applied in the definition of a user model. In addition, the second phase of this study presents a transcription review strategy based on confidence measures (CM) and compares it to the conventional post-editing strategy. Finally, a third strategy resulting from the combination of that based on \\{CM\\} with massive adaptation techniques for automatic speech recognition (ASR), achieved to improve the transcription review efficiency in comparison with the two aforementioned strategies.}, keywords = {Automatic Speech Recognition, Computer-assisted transcription, Interface design strategies, Usability study, video lecture repositories}, pubstate = {published}, tppubtype = {article} } Abstract Video lectures are widely used in education to support and complement face-to-face lectures. However, the utility of these audiovisual assets could be further improved by adding subtitles that can be exploited to incorporate added-value functionalities such as searchability, accessibility, translatability, note-taking, and discovery of content-related videos, among others. Today, automatic subtitles are prone to error, and need to be reviewed and post-edited in order to ensure that what students see on-screen are of an acceptable quality. This work investigates different user interface design strategies for this post-editing task to discover the best way to incorporate automatic transcription technologies into large educational video repositories. Our three-phase study involved lecturers from the Universitat Politècnica de València (UPV) with videos available on the poliMedia video lecture repository, which is currently over 10,000 video objects. Simply by conventional post-editing automatic transcriptions users almost reduced to half the time that would require to generate the transcription from scratch. As expected, this study revealed that the time spent by lecturers reviewing automatic transcriptions correlated directly with the accuracy of said transcriptions. However, it is also shown that the average time required to perform each individual editing operation could be precisely derived and could be applied in the definition of a user model. In addition, the second phase of this study presents a transcription review strategy based on confidence measures (CM) and compares it to the conventional post-editing strategy. Finally, a third strategy resulting from the combination of that based on \{CM\} with massive adaptation techniques for automatic speech recognition (ASR), achieved to improve the transcription review efficiency in comparison with the two aforementioned strategies. |
Brouns, Francis; Serrano Martínez-Santos, Nicolás ; Civera, Jorge; Kalz, Marco; Juan, Alfons Supporting language diversity of European MOOCs with the EMMA platform Inproceedings Proc. of the European MOOC Stakeholder Summit EMOOCs 2015, pp. 157–165, Mons (Belgium), 2015. Abstract | Links | BibTeX | Tags: Automatic Speech Recognition, EMMA, Statistical machine translation @inproceedings{Brouns2015, title = {Supporting language diversity of European MOOCs with the EMMA platform}, author = {Francis Brouns and Serrano Martínez-Santos, Nicolás and Jorge Civera and Marco Kalz and Alfons Juan}, url = {http://www.emoocs2015.eu/node/55}, year = {2015}, date = {2015-01-01}, booktitle = {Proc. of the European MOOC Stakeholder Summit EMOOCs 2015}, pages = {157--165}, address = {Mons (Belgium)}, abstract = {This paper introduces the cross-language support of the EMMA MOOC platform. Based on a discussion of language diversity in Europe, we introduce the development and evaluation of automated translation of texts and subtitling of videos from Dutch into English. The development of an Automatic Speech Recognition (ASR) system and a Statistical Machine Translation (SMT) system is described. The resources employed and evaluation approach is introduced. Initial evaluation results are presented. Finally, we provide an outlook into future research and development.}, keywords = {Automatic Speech Recognition, EMMA, Statistical machine translation}, pubstate = {published}, tppubtype = {inproceedings} } This paper introduces the cross-language support of the EMMA MOOC platform. Based on a discussion of language diversity in Europe, we introduce the development and evaluation of automated translation of texts and subtitling of videos from Dutch into English. The development of an Automatic Speech Recognition (ASR) system and a Statistical Machine Translation (SMT) system is described. The resources employed and evaluation approach is introduced. Initial evaluation results are presented. Finally, we provide an outlook into future research and development. |
2014 |
Valor Miró, Juan Daniel ; Spencer, R N; Pérez González de Martos, A; Garcés Díaz-Munío, G; Turró, C; Civera, J; Juan, A Evaluación del proceso de revisión de transcripciones automáticas para vídeos Polimedia Inproceedings Proc. of I Jornadas de Innovación Educativa y Docencia en Red (IN-RED 2014), pp. 272–278, Valencia (Spain), 2014. Abstract | Links | BibTeX | Tags: ASR, Docencia en Red, evaluaciones, Polimedia, transcripciones @inproceedings{Valor14-InRed, title = {Evaluación del proceso de revisión de transcripciones automáticas para vídeos Polimedia}, author = {Valor Miró, Juan Daniel and Spencer, R.N. and Pérez González de Martos, A. and Garcés Díaz-Munío, G. and Turró, C. and Civera, J. and Juan, A.}, url = {http://hdl.handle.net/10251/40404 http://dx.doi.org/10.4995/INRED.2014 http://www.mllp.upv.es/wp-content/uploads/2015/04/paper1.pdf https://www.mllp.upv.es/wp-content/uploads/2019/09/poster.pdf}, year = {2014}, date = {2014-07-01}, booktitle = {Proc. of I Jornadas de Innovación Educativa y Docencia en Red (IN-RED 2014)}, pages = {272--278}, address = {Valencia (Spain)}, abstract = {[EN] Video lectures are a tool of proven value and wide acceptance in universities which is leading to video lecture platforms like poliMèdia (Universitat Politècnica de València). transLectures is an EU project that generates high-quality automatic transcriptions and translations for the poliMèdia platform, and improves them by using massive adaptation and intelligent interaction techniques. In this paper we present the evaluation with lecturers carried out under the Docència en Xarxa 2012-2013 action plan with the aim of studying the process of transcription post-editing, in contrast with transcribing from scratch. [ES] Los vídeos docentes son una herramienta de demostrada utilidad y gran aceptación en el mundo universitario que está dando lugar a plataformas de vídeos docentes como poliMèdia (Universitat Politècnica de València). transLectures es un proyecto europeo que genera transcripciones y traducciones automáticas de alta calidad para la plataforma poliMèdia, mediante técnicas de adaptación masiva e interacción inteligente. En este artículo presentamos la evaluación con profesores que se realizó en el marco del plan de acción Docència en Xarxa 2012-2013, con el objetivo de estudiar el proceso de supervisión de transcripciones, comparándolo con la obtención de la transcripción sin disponer de una transcripción automática previa. [CA] Els vídeos docents són una eina d'utilitat demostrada i amb gran acceptació en el món universitari que està donant lloc a plataformes de vídeos docents com poliMèdia (Universitat Politècnica de València). transLectures és un projecte europeu que genera transcripcions i traduccions automàtiques d'alta qualitat per a la plataforma poliMèdia, mitjançant tècniques d'adaptació massiva i interacció intel·ligent. En aquest article presentem l'avaluació amb professors que es realitzà en el marc del pla d'acció Docència en Xarxa 2012-2013, amb l'objectiu d'estudiar el procés de supervisió de transcripcions, comparant-lo amb l'obtenció de la transcripció sense disposar d'una transcripció automàtica prèvia.}, keywords = {ASR, Docencia en Red, evaluaciones, Polimedia, transcripciones}, pubstate = {published}, tppubtype = {inproceedings} } [EN] Video lectures are a tool of proven value and wide acceptance in universities which is leading to video lecture platforms like poliMèdia (Universitat Politècnica de València). transLectures is an EU project that generates high-quality automatic transcriptions and translations for the poliMèdia platform, and improves them by using massive adaptation and intelligent interaction techniques. In this paper we present the evaluation with lecturers carried out under the Docència en Xarxa 2012-2013 action plan with the aim of studying the process of transcription post-editing, in contrast with transcribing from scratch. [ES] Los vídeos docentes son una herramienta de demostrada utilidad y gran aceptación en el mundo universitario que está dando lugar a plataformas de vídeos docentes como poliMèdia (Universitat Politècnica de València). transLectures es un proyecto europeo que genera transcripciones y traducciones automáticas de alta calidad para la plataforma poliMèdia, mediante técnicas de adaptación masiva e interacción inteligente. En este artículo presentamos la evaluación con profesores que se realizó en el marco del plan de acción Docència en Xarxa 2012-2013, con el objetivo de estudiar el proceso de supervisión de transcripciones, comparándolo con la obtención de la transcripción sin disponer de una transcripción automática previa. [CA] Els vídeos docents són una eina d'utilitat demostrada i amb gran acceptació en el món universitari que està donant lloc a plataformes de vídeos docents com poliMèdia (Universitat Politècnica de València). transLectures és un projecte europeu que genera transcripcions i traduccions automàtiques d'alta qualitat per a la plataforma poliMèdia, mitjançant tècniques d'adaptació massiva i interacció intel·ligent. En aquest article presentem l'avaluació amb professors que es realitzà en el marc del pla d'acció Docència en Xarxa 2012-2013, amb l'objectiu d'estudiar el procés de supervisió de transcripcions, comparant-lo amb l'obtenció de la transcripció sense disposar d'una transcripció automàtica prèvia. |
Valor Miró, Juan Daniel ; Spencer, R N; Pérez González de Martos, A; Garcés Díaz-Munío, G; Turró, C; Civera, J; Juan, A Evaluating intelligent interfaces for post-editing automatic transcriptions of online video lectures Journal Article Open Learning: The Journal of Open, Distance and e-Learning, 29 (1), pp. 72–85, 2014. Abstract | Links | BibTeX | Tags: @article{doi:10.1080/02680513.2014.909722, title = {Evaluating intelligent interfaces for post-editing automatic transcriptions of online video lectures}, author = {Valor Miró, Juan Daniel and Spencer, R.N. and Pérez González de Martos, A. and Garcés Díaz-Munío, G. and Turró, C. and Civera, J. and Juan, A.}, url = {http://hdl.handle.net/10251/55925 http://dx.doi.org/10.1080/02680513.2014.909722 http://www.mllp.upv.es/wp-content/uploads/2015/04/author_version.pdf}, year = {2014}, date = {2014-01-01}, journal = {Open Learning: The Journal of Open, Distance and e-Learning}, volume = {29}, number = {1}, pages = {72--85}, abstract = {[EN] Video lectures are fast becoming an everyday educational resource in higher education. They are being incorporated into existing university curricula around the world, while also emerging as a key component of the open education movement. In 2007 the Universitat Politècnica de València (UPV) implemented its poliMèdia lecture capture system for the creation and publication of quality educational video content and now has a collection of over 10,000 video objects. In 2011 it embarked on the EU-subsidised transLectures project to add automatic subtitles to these videos in both Spanish and other languages. By doing so, it allows access to their educational content by non-native speakers and the deaf and hard-of-hearing, as well as enabling advanced repository management functions. In this paper, following a short introduction to poliMèdia, transLectures and Docència en Xarxa, the UPV's action plan to boost the use of digital resources at the university, we will discuss the three-stage evaluation process carried out with the collaboration of UPV lecturers to find the best interaction protocol for the task of post-editing automatic subtitles. [CA] "Avaluació d'interfícies intel·ligents per a la postedició de transcripcions automàtiques de vídeos docents en línia": Els vídeos docents s'estan convertint en un recurs d'ús quotidià en l'educació superior. Estan entrant en els plans d'estudis universitaris de tot el món, al mateix temps que es defineixen com un component clau del moviment de l'educació lliure. L'any 2007, la Universitat Politècnica de València (UPV) implementà el seu sistema d'enregistrament de classes poliMèdia, que permet la producció i publicació de continguts audiovisuals d'alta qualitat, i que ja acumula més de 10.000 vídeos. L'any 2011, la UPV va entrar en el projecte europeu transLectures per a afegir subtítols automàtics a aquests vídeos en castellà, català i altres llengües. Així, es facilita l'accés a aquests continguts educatius per part de parlants d'altres llengües i de persones sordes o amb dificultats auditives, i també es proporcionen funcions avançades de gestió del repositori. En aquest article, després de presentar poliMèdia, transLectures i Docència en Xarxa (el pla d'acció de la UPV per a impulsar l'ús de recursos digitals en la universitat), explicarem el procés d'avaluació en tres fases que s'ha realitzat amb la col·laboració de professors de la UPV per a trobar el millor protocol d'interacció per a la postedició de subtítols automàtics.}, keywords = {}, pubstate = {published}, tppubtype = {article} } [EN] Video lectures are fast becoming an everyday educational resource in higher education. They are being incorporated into existing university curricula around the world, while also emerging as a key component of the open education movement. In 2007 the Universitat Politècnica de València (UPV) implemented its poliMèdia lecture capture system for the creation and publication of quality educational video content and now has a collection of over 10,000 video objects. In 2011 it embarked on the EU-subsidised transLectures project to add automatic subtitles to these videos in both Spanish and other languages. By doing so, it allows access to their educational content by non-native speakers and the deaf and hard-of-hearing, as well as enabling advanced repository management functions. In this paper, following a short introduction to poliMèdia, transLectures and Docència en Xarxa, the UPV's action plan to boost the use of digital resources at the university, we will discuss the three-stage evaluation process carried out with the collaboration of UPV lecturers to find the best interaction protocol for the task of post-editing automatic subtitles. [CA] "Avaluació d'interfícies intel·ligents per a la postedició de transcripcions automàtiques de vídeos docents en línia": Els vídeos docents s'estan convertint en un recurs d'ús quotidià en l'educació superior. Estan entrant en els plans d'estudis universitaris de tot el món, al mateix temps que es defineixen com un component clau del moviment de l'educació lliure. L'any 2007, la Universitat Politècnica de València (UPV) implementà el seu sistema d'enregistrament de classes poliMèdia, que permet la producció i publicació de continguts audiovisuals d'alta qualitat, i que ja acumula més de 10.000 vídeos. L'any 2011, la UPV va entrar en el projecte europeu transLectures per a afegir subtítols automàtics a aquests vídeos en castellà, català i altres llengües. Així, es facilita l'accés a aquests continguts educatius per part de parlants d'altres llengües i de persones sordes o amb dificultats auditives, i també es proporcionen funcions avançades de gestió del repositori. En aquest article, després de presentar poliMèdia, transLectures i Docència en Xarxa (el pla d'acció de la UPV per a impulsar l'ús de recursos digitals en la universitat), explicarem el procés d'avaluació en tres fases que s'ha realitzat amb la col·laboració de professors de la UPV per a trobar el millor protocol d'interacció per a la postedició de subtítols automàtics. |
Piqueras, S; del-Agua, M A; Giménez, A; Civera, J; Juan, A Statistical text-to-speech synthesis of Spanish subtitles Inproceedings Proc. of VIII Jornadas en Tecnología del Habla and IV Iberian SLTech Workshop (IberSpeech 2014), Las Palmas de Gran Canaria (Spain), 2014. @inproceedings{PiqAgu14, title = {Statistical text-to-speech synthesis of Spanish subtitles}, author = {S. Piqueras and M. A. del-Agua and A. Giménez and J. Civera and A. Juan}, url = {http://www.mllp.upv.es/wp-content/uploads/2015/04/paper3.pdf http://link.springer.com/chapter/10.1007%2F978-3-319-13623-3_5}, year = {2014}, date = {2014-01-01}, booktitle = {Proc. of VIII Jornadas en Tecnología del Habla and IV Iberian SLTech Workshop (IberSpeech 2014)}, address = {Las Palmas de Gran Canaria (Spain)}, keywords = {}, pubstate = {published}, tppubtype = {inproceedings} } |
Pérez-González-de-Martos, A; Silvestre-Cerdá, J A; Rihtar, M; Juan, A; Civera, J Using Automatic Speech Transcriptions in Lecture Recommendation Systems Inproceedings Proc. of VIII Jornadas en Tecnología del Habla and IV Iberian SLTech Workshop (IberSpeech 2014), Las Palmas de Gran Canaria (Spain), 2014. @inproceedings{PerSil14, title = {Using Automatic Speech Transcriptions in Lecture Recommendation Systems}, author = {A. Pérez-González-de-Martos and J. A. Silvestre-Cerdá and M. Rihtar and A. Juan and J. Civera}, url = {http://www.mllp.upv.es/wp-content/uploads/2015/04/lavie_is2014_camready1.pdf}, year = {2014}, date = {2014-01-01}, booktitle = {Proc. of VIII Jornadas en Tecnología del Habla and IV Iberian SLTech Workshop (IberSpeech 2014)}, address = {Las Palmas de Gran Canaria (Spain)}, keywords = {}, pubstate = {published}, tppubtype = {inproceedings} } |
del-Agua, M A; Giménez, A; Serrano, N; Andrés-Ferrer, J; Civera, J; Sanchis, A; Juan, A The transLectures-UPV toolkit Inproceedings Proc. of VIII Jornadas en Tecnología del Habla and IV Iberian SLTech Workshop (IberSpeech 2014), Las Palmas de Gran Canaria (Spain), 2014. @inproceedings{AguGim14, title = {The transLectures-UPV toolkit}, author = {M. A. del-Agua and A. Giménez and N. Serrano and J. Andrés-Ferrer and J. Civera and A. Sanchis and A. Juan}, url = {http://www.mllp.upv.es/wp-content/uploads/2015/04/IberSpeech2014-TLK-camready1.pdf}, year = {2014}, date = {2014-01-01}, booktitle = {Proc. of VIII Jornadas en Tecnología del Habla and IV Iberian SLTech Workshop (IberSpeech 2014)}, address = {Las Palmas de Gran Canaria (Spain)}, keywords = {}, pubstate = {published}, tppubtype = {inproceedings} } |
Serrano, Nicolás; Civera, Jorge; Sanchis, Alberto; Juan, A Effective balancing error and user effort in interactive handwriting recognition Journal Article Pattern Recognition Letters, 37 , pp. 135–142, 2014. @article{Serrano14b, title = {Effective balancing error and user effort in interactive handwriting recognition}, author = {Nicolás Serrano and Jorge Civera and Alberto Sanchis and A. Juan}, url = {http://dx.doi.org/10.1016/j.patrec.2013.03.010}, year = {2014}, date = {2014-01-01}, journal = {Pattern Recognition Letters}, volume = {37}, pages = {135--142}, keywords = {}, pubstate = {published}, tppubtype = {article} } |
Serrano, Nicolás; Giménez, Adrià; Civera, Jorge; Sanchis, Alberto; Juan, Alfons Interactive Handwriting Recognition with Limited User effort Journal Article Intl. Journal on Document Analysis and Recognition (IJDAR), 17 , pp. 47–59, 2014. @article{Serrano14a, title = {Interactive Handwriting Recognition with Limited User effort}, author = {Nicolás Serrano and Adrià Giménez and Jorge Civera and Alberto Sanchis and Alfons Juan}, url = {http://dx.doi.org/10.1007/s10032-013-0204-5}, year = {2014}, date = {2014-01-01}, journal = {Intl. Journal on Document Analysis and Recognition (IJDAR)}, volume = {17}, pages = {47--59}, keywords = {}, pubstate = {published}, tppubtype = {article} } |
2013 |
Silvestre-Cerdà, Joan Albert; Pérez, Alejandro; Jiménez, Manuel; Turró, Carlos; Juan, Alfons; Civera, Jorge A System Architecture to Support Cost-Effective Transcription and Translation of Large Video Lecture Repositories Inproceedings Proc. of the IEEE Intl. Conf. on Systems, Man, and Cybernetics SMC 2013 , pp. 3994-3999, Manchester (UK), 2013. Abstract | Links | BibTeX | Tags: Accessibility, Automatic Speech Recognition, Education, Intelligent Interaction, Language Technologies, Machine Translation, Massive Adaptation, Multilingualism, Opencast Matterhorn, Video Lectures @inproceedings{Silvestre-Cerdà2013, title = {A System Architecture to Support Cost-Effective Transcription and Translation of Large Video Lecture Repositories}, author = {Joan Albert Silvestre-Cerdà and Alejandro Pérez and Manuel Jiménez and Carlos Turró and Alfons Juan and Jorge Civera}, url = {http://dx.doi.org/10.1109/SMC.2013.682}, year = {2013}, date = {2013-01-01}, booktitle = {Proc. of the IEEE Intl. Conf. on Systems, Man, and Cybernetics SMC 2013 }, pages = {3994-3999}, address = {Manchester (UK)}, abstract = {Online video lecture repositories are rapidly growing and becoming established as fundamental knowledge assets. However, most lectures are neither transcribed nor translated because of the lack of cost-effective solutions that can give accurate enough results. In this paper, we describe a system architecture that supports the cost-effective transcription and translation of large video lecture repositories. This architecture has been adopted in the EU project transLectures and is now being tested on a repository of more than 9000 video lectures at the Universitat Politecnica de Valencia. Following a brief description of this repository and of the transLectures project, we describe the proposed system architecture in detail. We also report empirical results on the quality of the transcriptions and translations currently being maintained and steadily improved.}, keywords = {Accessibility, Automatic Speech Recognition, Education, Intelligent Interaction, Language Technologies, Machine Translation, Massive Adaptation, Multilingualism, Opencast Matterhorn, Video Lectures}, pubstate = {published}, tppubtype = {inproceedings} } Online video lecture repositories are rapidly growing and becoming established as fundamental knowledge assets. However, most lectures are neither transcribed nor translated because of the lack of cost-effective solutions that can give accurate enough results. In this paper, we describe a system architecture that supports the cost-effective transcription and translation of large video lecture repositories. This architecture has been adopted in the EU project transLectures and is now being tested on a repository of more than 9000 video lectures at the Universitat Politecnica de Valencia. Following a brief description of this repository and of the transLectures project, we describe the proposed system architecture in detail. We also report empirical results on the quality of the transcriptions and translations currently being maintained and steadily improved. |
2012 |
Silvestre-Cerdà, Joan Albert ; Del Agua, Miguel ; Garcés, Gonçal; Gascó, Guillem; Giménez-Pastor, Adrià; Martínez, Adrià; Pérez González de Martos, Alejandro ; Sánchez, Isaías; Serrano Martínez-Santos, Nicolás ; Spencer, Rachel; Valor Miró, Juan Daniel ; Andrés-Ferrer, Jesús; Civera, Jorge; Sanchís, Alberto; Juan, Alfons transLectures Inproceedings Proceedings (Online) of IberSPEECH 2012, pp. 345–351, Madrid (Spain), 2012. Abstract | Links | BibTeX | Tags: Accessibility, Automatic Speech Recognition, Education, Intelligent Interaction, Language Technologies, Machine Translation, Massive Adaptation, Multilingualism, Opencast Matterhorn, Video Lectures @inproceedings{Silvestre-Cerdà2012b, title = {transLectures}, author = {Silvestre-Cerdà, Joan Albert and Del Agua, Miguel and Gonçal Garcés and Guillem Gascó and Adrià Giménez-Pastor and Adrià Martínez and Pérez González de Martos, Alejandro and Isaías Sánchez and Serrano Martínez-Santos, Nicolás and Rachel Spencer and Valor Miró, Juan Daniel and Jesús Andrés-Ferrer and Jorge Civera and Alberto Sanchís and Alfons Juan}, url = {http://hdl.handle.net/10251/37290 http://lorien.die.upm.es/~lapiz/rtth/JORNADAS/VII/IberSPEECH2012_OnlineProceedings.pdf https://web.archive.org/web/20130609073144/http://iberspeech2012.ii.uam.es/IberSPEECH2012_OnlineProceedings.pdf http://www.mllp.upv.es/wp-content/uploads/2015/04/1209IberSpeech.pdf}, year = {2012}, date = {2012-11-22}, booktitle = {Proceedings (Online) of IberSPEECH 2012}, pages = {345--351}, address = {Madrid (Spain)}, abstract = {[EN] transLectures (Transcription and Translation of Video Lectures) is an EU STREP project in which advanced automatic speech recognition and machine translation techniques are being tested on large video lecture repositories. The project began in November 2011 and will run for three years. This paper will outline the project's main motivation and objectives, and give a brief description of the two main repositories being considered: VideoLectures.NET and poliMèdia. The first results obtained by the UPV group for the poliMedia repository will also be provided. [CA] transLectures (Transcription and Translation of Video Lectures) és un projecte del 7PM de la Unió Europea en el qual s'estan posant a prova tècniques avançades de reconeixement automàtic de la parla i de traducció automàtica sobre grans repositoris digitals de vídeos docents. El projecte començà al novembre de 2011 i tindrà una duració de tres anys. En aquest article exposem la motivació i els objectius del projecte, i descrivim breument els dos repositoris principals sobre els quals es treballa: VideoLectures.NET i poliMèdia. També oferim els primers resultats obtinguts per l'equip de la UPV al repositori poliMèdia.}, keywords = {Accessibility, Automatic Speech Recognition, Education, Intelligent Interaction, Language Technologies, Machine Translation, Massive Adaptation, Multilingualism, Opencast Matterhorn, Video Lectures}, pubstate = {published}, tppubtype = {inproceedings} } [EN] transLectures (Transcription and Translation of Video Lectures) is an EU STREP project in which advanced automatic speech recognition and machine translation techniques are being tested on large video lecture repositories. The project began in November 2011 and will run for three years. This paper will outline the project's main motivation and objectives, and give a brief description of the two main repositories being considered: VideoLectures.NET and poliMèdia. The first results obtained by the UPV group for the poliMedia repository will also be provided. [CA] transLectures (Transcription and Translation of Video Lectures) és un projecte del 7PM de la Unió Europea en el qual s'estan posant a prova tècniques avançades de reconeixement automàtic de la parla i de traducció automàtica sobre grans repositoris digitals de vídeos docents. El projecte començà al novembre de 2011 i tindrà una duració de tres anys. En aquest article exposem la motivació i els objectius del projecte, i descrivim breument els dos repositoris principals sobre els quals es treballa: VideoLectures.NET i poliMèdia. També oferim els primers resultats obtinguts per l'equip de la UPV al repositori poliMèdia. |
Silvestre-Cerdà, Joan Albert; Giménez, Adrià; Andrés-Ferrer, Jesús; Civera, Jorge; Juan, Alfons Albayzin Evaluation: The PRHLT-UPV Audio Segmentation System Inproceedings Proceedings (Online) of IberSPEECH 2012, pp. 596-600, Madrid (Spain), 2012. Abstract | Links | BibTeX | Tags: @inproceedings{Silvestre-Cerdà2012c, title = {Albayzin Evaluation: The PRHLT-UPV Audio Segmentation System}, author = {Joan Albert Silvestre-Cerdà and Adrià Giménez and Jesús Andrés-Ferrer and Jorge Civera and Alfons Juan}, url = {http://hdl.handle.net/10251/53699 http://iberspeech2012.ii.uam.es/IberSPEECH2012_OnlineProceedings.pdf}, year = {2012}, date = {2012-11-22}, booktitle = {Proceedings (Online) of IberSPEECH 2012}, pages = {596-600}, address = {Madrid (Spain)}, abstract = {This paper describes the audio segmentation system developed by the PRHLT research group at the UPV for the Albayzin Audio Segmentation Evaluation 2012. The PRHLT-UPV audio segmentation system is based on a conventional GMM-HMM speech recognition approach in which the vocabulary set is defined by the power set of segment classes. MFCC features were extracted to represent the acoustic signal and the AK toolkit was used for both, training acoustic models and performing audio segmentation. Experimental results reveals that our system provides an excellent performance on speech detection, so it could be successfully employed to provide speech segments to a diarization or speech recognition system.}, keywords = {}, pubstate = {published}, tppubtype = {inproceedings} } This paper describes the audio segmentation system developed by the PRHLT research group at the UPV for the Albayzin Audio Segmentation Evaluation 2012. The PRHLT-UPV audio segmentation system is based on a conventional GMM-HMM speech recognition approach in which the vocabulary set is defined by the power set of segment classes. MFCC features were extracted to represent the acoustic signal and the AK toolkit was used for both, training acoustic models and performing audio segmentation. Experimental results reveals that our system provides an excellent performance on speech detection, so it could be successfully employed to provide speech segments to a diarization or speech recognition system. |
Turró, Carlos; Juan, Alfons; Civera, Jorge; Orliĉ, Davor; Jermol, Mitja transLectures: Transcription and Translation of Video Lectures Inproceedings Proc. of Cambridge 2012: Innovation and Impact - Openly Collaborating to Enhance Education, pp. 543-546, Cambridge (UK), 2012. Abstract | Links | BibTeX | Tags: Automatic Speech Recognition, Statistical machine translation @inproceedings{Turró2012, title = {transLectures: Transcription and Translation of Video Lectures}, author = {Carlos Turró and Alfons Juan and Jorge Civera and Davor Orliĉ and Mitja Jermol}, url = {http://oro.open.ac.uk/id/eprint/33640 http://hdl.handle.net/10251/54166}, year = {2012}, date = {2012-01-01}, booktitle = {Proc. of Cambridge 2012: Innovation and Impact - Openly Collaborating to Enhance Education}, pages = {543-546}, address = {Cambridge (UK)}, abstract = {transLectures is a FP7 project aimed at developing innovative, cost-effective solutions to produce accurate transcriptions and translations in large repositories of video lectures. This paper describes user requirements, first integration steps and evaluation plans at transLectures case studies, VideoLectures.NET and poliMedia.}, keywords = {Automatic Speech Recognition, Statistical machine translation}, pubstate = {published}, tppubtype = {inproceedings} } transLectures is a FP7 project aimed at developing innovative, cost-effective solutions to produce accurate transcriptions and translations in large repositories of video lectures. This paper describes user requirements, first integration steps and evaluation plans at transLectures case studies, VideoLectures.NET and poliMedia. |
del Agua, MiguelA.; Serrano, Nicolás ; Civera, Jorge ; Juan, Alfons Character-Based Handwritten Text Recognition of Multilingual Documents Incollection Advances in Speech and Language Technologies for Iberian Languages (IberSpeech 2012), 328 , pp. 187-196, Springer Berlin Heidelberg, 2012, ISBN: 978-3-642-35291-1, (doi: 10.1007/978-3-642-35292-8_20). @incollection{delAgua12, title = {Character-Based Handwritten Text Recognition of Multilingual Documents}, author = {MiguelA. del Agua and Nicolás Serrano and Jorge Civera and Alfons Juan}, url = {http://hdl.handle.net/10251/35180}, isbn = {978-3-642-35291-1}, year = {2012}, date = {2012-01-01}, booktitle = {Advances in Speech and Language Technologies for Iberian Languages (IberSpeech 2012)}, volume = {328}, pages = {187-196}, publisher = {Springer Berlin Heidelberg}, series = {Communications in Computer and Information Science}, note = {doi: 10.1007/978-3-642-35292-8_20}, keywords = {}, pubstate = {published}, tppubtype = {incollection} } |
Silvestre-Cerdà, Joan Albert; Andrés-Ferrer, Jesús; Civera, Jorge Explicit length modelling for statistical machine translation Journal Article Pattern Recognition, 45 (9), pp. 3183 - 3192, 2012, ISSN: 0031-3203. Abstract | Links | BibTeX | Tags: Length modelling, Log-linear models, Phrase-based models, Statistical machine translation @article{Silvestre-Cerdà2012a, title = {Explicit length modelling for statistical machine translation}, author = {Joan Albert Silvestre-Cerdà and Jesús Andrés-Ferrer and Jorge Civera}, url = {http://hdl.handle.net/10251/34996}, issn = {0031-3203}, year = {2012}, date = {2012-01-01}, journal = {Pattern Recognition}, volume = {45}, number = {9}, pages = {3183 - 3192}, abstract = {Explicit length modelling has been previously explored in statistical pattern recognition with successful results. In this paper, two length models along with two parameter estimation methods and two alternative parametrisation for statistical machine translation (SMT) are presented. More precisely, we incorporate explicit bilingual length modelling in a state-of-the-art log-linear SMT system as an additional feature function in order to prove the contribution of length information. Finally, a systematic evaluation on reference SMT tasks considering different language pairs prove the benefits of explicit length modelling.}, keywords = {Length modelling, Log-linear models, Phrase-based models, Statistical machine translation}, pubstate = {published}, tppubtype = {article} } Explicit length modelling has been previously explored in statistical pattern recognition with successful results. In this paper, two length models along with two parameter estimation methods and two alternative parametrisation for statistical machine translation (SMT) are presented. More precisely, we incorporate explicit bilingual length modelling in a state-of-the-art log-linear SMT system as an additional feature function in order to prove the contribution of length information. Finally, a systematic evaluation on reference SMT tasks considering different language pairs prove the benefits of explicit length modelling. |
Valor Miró, Juan Daniel ; Pérez González de Martos, Alejandro ; Civera, Jorge ; Juan, Alfons Integrating a State-of-the-Art ASR System into the Opencast Matterhorn Platform Incollection Advances in Speech and Language Technologies for Iberian Languages (IberSpeech 2012), 328 , pp. 237-246, Springer Berlin Heidelberg, 2012, ISBN: 978-3-642-35291-1, (doi: 10.1007/978-3-642-35292-8_20). Links | BibTeX | Tags: Google N-Gram, Language Modeling, Linear Combination, Opencast Matterhorn, Speech Recognition @incollection{Valor2012, title = {Integrating a State-of-the-Art ASR System into the Opencast Matterhorn Platform}, author = {Valor Miró, Juan Daniel and Pérez González de Martos, Alejandro and Civera, Jorge and Juan, Alfons}, url = {http://hdl.handle.net/10251/35190 http://www.mllp.upv.es/wp-content/uploads/2015/04/paper2.pdf}, isbn = {978-3-642-35291-1}, year = {2012}, date = {2012-01-01}, booktitle = {Advances in Speech and Language Technologies for Iberian Languages (IberSpeech 2012)}, volume = {328}, pages = {237-246}, publisher = {Springer Berlin Heidelberg}, series = {Communications in Computer and Information Science}, note = {doi: 10.1007/978-3-642-35292-8_20}, keywords = {Google N-Gram, Language Modeling, Linear Combination, Opencast Matterhorn, Speech Recognition}, pubstate = {published}, tppubtype = {incollection} } |
2011 |
Silvestre-Cerdà, Joan Albert; García-Martínez, Mercedes; Barrón-Cedeño, Alberto; Civera, Jorge; Rosso, Paolo Extracción de corpus paralelos de la Wikipedia basada en la obtención de alineamientos bilingües a nivel de frase Inproceedings Proceedings of the Workshop on Iberian Cross-Language Natural Language Processing Tasks (ICL 2011), pp. 14-21, CEUR-WS, 2011, ISSN: 1613-0073. Abstract | Links | BibTeX | Tags: Comparable Corpora, Parallel Sentences Extraction, Statistical machine translation @inproceedings{Silvestre-Cerdà2011b, title = {Extracción de corpus paralelos de la Wikipedia basada en la obtención de alineamientos bilingües a nivel de frase}, author = {Joan Albert Silvestre-Cerdà and Mercedes García-Martínez and Alberto Barrón-Cedeño and Jorge Civera and Paolo Rosso}, url = {http://hdl.handle.net/10251/27930}, issn = {1613-0073}, year = {2011}, date = {2011-01-01}, booktitle = {Proceedings of the Workshop on Iberian Cross-Language Natural Language Processing Tasks (ICL 2011)}, volume = {824}, pages = {14-21}, publisher = {CEUR-WS}, abstract = {This paper presents a proposal for extracting parallel corpora from Wikipedia on the basis of statistical machine translation techniques. We have used word-level alignment models from IBM in order to obtain phrase-level bilingual alignments between documents pairs. We have manually annotated a set of test English-Spanish comparable documents in order to evaluate the model. The obtained results are encouraging.}, keywords = {Comparable Corpora, Parallel Sentences Extraction, Statistical machine translation}, pubstate = {published}, tppubtype = {inproceedings} } This paper presents a proposal for extracting parallel corpora from Wikipedia on the basis of statistical machine translation techniques. We have used word-level alignment models from IBM in order to obtain phrase-level bilingual alignments between documents pairs. We have manually annotated a set of test English-Spanish comparable documents in order to evaluate the model. The obtained results are encouraging. |
Silvestre-Cerdà, Joan Albert; Andrés-Ferrer, Jesús ; Civera, Jorge Explicit Length Modelling for Statistical Machine Translation Incollection Vitrià, Jordi ; Sanches, JoãoMiguel ; Hernández, Mario (Ed.): Pattern Recognition and Image Analysis (IbPRIA 2011), 6669 , pp. 273-280, Springer Berlin Heidelberg, 2011, ISBN: 978-3-642-21256-7. Abstract | Links | BibTeX | Tags: Length modelling, Log-linear models, Phrase-based models, Statistical machine translation @incollection{Silvestre-Cerdà2011, title = {Explicit Length Modelling for Statistical Machine Translation}, author = { Joan Albert Silvestre-Cerdà and Jesús Andrés-Ferrer and Jorge Civera}, editor = {Vitrià, Jordi and Sanches, JoãoMiguel and Hernández, Mario}, url = {http://hdl.handle.net/10251/35749 http://dx.doi.org/10.1007/978-3-642-21257-4_34}, isbn = {978-3-642-21256-7}, year = {2011}, date = {2011-01-01}, booktitle = {Pattern Recognition and Image Analysis (IbPRIA 2011)}, volume = {6669}, pages = {273-280}, publisher = {Springer Berlin Heidelberg}, series = {Lecture Notes in Computer Science}, abstract = {Explicit length modelling has been previously explored in statistical pattern recognition with successful results. In this paper, two length models along with two parameter estimation methods for statistical machine translation (SMT) are presented. More precisely, we incorporate explicit length modelling in a state-of-the-art log-linear SMT system as an additional feature function in order to prove the contribution of length information. Finally, promising experimental results are reported on a reference SMT task.}, keywords = {Length modelling, Log-linear models, Phrase-based models, Statistical machine translation}, pubstate = {published}, tppubtype = {incollection} } Explicit length modelling has been previously explored in statistical pattern recognition with successful results. In this paper, two length models along with two parameter estimation methods for statistical machine translation (SMT) are presented. More precisely, we incorporate explicit length modelling in a state-of-the-art log-linear SMT system as an additional feature function in order to prove the contribution of length information. Finally, promising experimental results are reported on a reference SMT task. |
2010 |
González-Rubio, J; Civera, J; Juan, A; Casacuberta, F Saturnalia: A Latin-Catalan Parallel Corpus for Statistical MT Inproceedings Proc. of the 7th Language Resources and Evaluation Conf. (LREC 2010), pp. 3405–3408, Valleta (Malta), 2010. BibTeX | Tags: @inproceedings{Juan10_05b, title = {Saturnalia: A Latin-Catalan Parallel Corpus for Statistical MT}, author = {J. González-Rubio and J. Civera and A. Juan and F. Casacuberta}, year = {2010}, date = {2010-01-01}, booktitle = {Proc. of the 7th Language Resources and Evaluation Conf. (LREC 2010)}, pages = {3405--3408}, address = {Valleta (Malta)}, keywords = {}, pubstate = {published}, tppubtype = {inproceedings} } |
2009 |
Casacuberta, F; Civera, J; Cubel, E; Lagarda, A L; Lapalme, G; Macklovitch, E; Vidal, E Human interaction for high quality machine translation Journal Article Communications of the ACM, 52 (10), pp. 135-138, 2009. BibTeX | Tags: @article{Casacuberta08a, title = {Human interaction for high quality machine translation}, author = {F. Casacuberta and J. Civera and E. Cubel and A.L. Lagarda and G. Lapalme and E. Macklovitch and E. Vidal}, year = {2009}, date = {2009-01-01}, journal = {Communications of the ACM}, volume = {52}, number = {10}, pages = {135-138}, keywords = {}, pubstate = {published}, tppubtype = {article} } |
Barrachina, S; Bender, O; Casacuberta, F; Civera, J; Cubel, E; Khadivi, S; Ney, Lagarda A H; Tomás, J; Vidal, E Statistical approaches to computer-assisted translation Journal Article Computational Linguistics, 35 (1), pp. 3–28, 2009. BibTeX | Tags: @article{Barrachina08a, title = {Statistical approaches to computer-assisted translation}, author = {S. Barrachina and O. Bender and F. Casacuberta and J. Civera and E. Cubel and S. Khadivi and A. Lagarda H. Ney and J. Tomás and E. Vidal}, year = {2009}, date = {2009-01-01}, journal = {Computational Linguistics}, volume = {35}, number = {1}, pages = {3--28}, keywords = {}, pubstate = {published}, tppubtype = {article} } |
2008 |
Sanchis-Trilles, G; González, M T; Casacuberta, F; Vidal, E; Civera, J Introducing Additional Input Information into IMT Systems Incollection Proceedings of the 5th Joint Workshop on Multimodal Interaction and Related Machine Learning Algorithms. Lecture Notes in Computer Sciences, 5237 , pp. 284–295, Utecht, The Netherlands, 2008. BibTeX | Tags: @incollection{GSanchis08d, title = {Introducing Additional Input Information into IMT Systems}, author = {G. Sanchis-Trilles and M.T. González and F. Casacuberta and E. Vidal and J. Civera}, year = {2008}, date = {2008-01-01}, booktitle = {Proceedings of the 5th Joint Workshop on Multimodal Interaction and Related Machine Learning Algorithms. Lecture Notes in Computer Sciences}, volume = {5237}, pages = {284--295}, address = {Utecht, The Netherlands}, keywords = {}, pubstate = {published}, tppubtype = {incollection} } |
Romero, V; Toselli, A H; Civera, J; Vidal, E Improvements in the Computer Assisted Transciption System of Handwritten Text Images Inproceedings Proc. of the 8th Int. Workshop on Pattern Recognition in Information Systems (PRIS 2008), pp. 103–112, Barcelona (Spain), 2008. BibTeX | Tags: @inproceedings{VRomero08a, title = {Improvements in the Computer Assisted Transciption System of Handwritten Text Images}, author = {V. Romero and A. H. Toselli and J. Civera and E. Vidal}, year = {2008}, date = {2008-01-01}, booktitle = {Proc. of the 8th Int. Workshop on Pattern Recognition in Information Systems (PRIS 2008)}, pages = {103--112}, address = {Barcelona (Spain)}, keywords = {}, pubstate = {published}, tppubtype = {inproceedings} } |
Sanchis-Trilles, G; Ortiz-Martínez, Daniel; Civera, Jorge; Casacuberta, Francisco; Vidal, Enrique; Hoang, Hieu Improving Interactive Machine Translation via Mouse Actions Inproceedings EMNLP 2008: conference on Empirical Methods in Natural Language Processing, Waikiki, Honolulu, Hawaii, 2008. BibTeX | Tags: @inproceedings{GSanchis08e, title = {Improving Interactive Machine Translation via Mouse Actions}, author = {G. Sanchis-Trilles and Daniel Ortiz-Martínez and Jorge Civera and Francisco Casacuberta and Enrique Vidal and Hieu Hoang}, year = {2008}, date = {2008-01-01}, booktitle = {EMNLP 2008: conference on Empirical Methods in Natural Language Processing}, address = {Waikiki, Honolulu, Hawaii}, keywords = {}, pubstate = {published}, tppubtype = {inproceedings} } |
Publications
2016 |
Generación eficiente de transcripciones y traducciones automáticas en poliMedia Inproceedings Proc. of II Congreso Nacional de Innovación Educativa y Docencia en Red (IN-RED 2016), pp. 21–29, València (Spain), 2016. |
The MLLP system for the 4th CHiME Challenge Inproceedings Proc. of the 4th Intl. Workshop on Speech Processing in Everyday Environments (CHiME 2016), pp. 57–59, San Francisco (USA), 2016. |
2015 |
The MLLP ASR Systems for IWSLT 2015 Inproceedings Proc. of 12th Intl. Workshop on Spoken Language Translation (IWSLT 2015), pp. 39–44, Da Nang (Vietnam), 2015. |
Efficient Generation of High-Quality Multilingual Subtitles for Video Lecture Repositories Inproceedings Proc. of 10th European Conf. on Technology Enhanced Learning (EC-TEL 2015), pp. 485–490, Toledo (Spain), 2015, ISBN: 978-3-319-24258-3. |
MLLP Transcription and Translation Platform Miscellaneous 2015, (Short paper for demo presentation accepted at 10th European Conf. on Technology Enhanced Learning (EC-TEL 2015), Toledo (Spain), 2015.). |
Evaluación de la revisión de transcripciones y traducciones automáticas de vídeos poliMedia Inproceedings Proc. of I Congreso Nacional de Innovación Educativa y Docencia en Red (IN-RED 2015), pp. 464–468, València (Spain), 2015. |
Efficiency and usability study of innovative computer-aided transcription strategies for video lecture repositories Journal Article Speech Communication, 74 , pp. 65–75, 2015, ISSN: 0167-6393. |
Supporting language diversity of European MOOCs with the EMMA platform Inproceedings Proc. of the European MOOC Stakeholder Summit EMOOCs 2015, pp. 157–165, Mons (Belgium), 2015. |
2014 |
Evaluación del proceso de revisión de transcripciones automáticas para vídeos Polimedia Inproceedings Proc. of I Jornadas de Innovación Educativa y Docencia en Red (IN-RED 2014), pp. 272–278, Valencia (Spain), 2014. |
Evaluating intelligent interfaces for post-editing automatic transcriptions of online video lectures Journal Article Open Learning: The Journal of Open, Distance and e-Learning, 29 (1), pp. 72–85, 2014. |
Statistical text-to-speech synthesis of Spanish subtitles Inproceedings Proc. of VIII Jornadas en Tecnología del Habla and IV Iberian SLTech Workshop (IberSpeech 2014), Las Palmas de Gran Canaria (Spain), 2014. |
Using Automatic Speech Transcriptions in Lecture Recommendation Systems Inproceedings Proc. of VIII Jornadas en Tecnología del Habla and IV Iberian SLTech Workshop (IberSpeech 2014), Las Palmas de Gran Canaria (Spain), 2014. |
The transLectures-UPV toolkit Inproceedings Proc. of VIII Jornadas en Tecnología del Habla and IV Iberian SLTech Workshop (IberSpeech 2014), Las Palmas de Gran Canaria (Spain), 2014. |
Effective balancing error and user effort in interactive handwriting recognition Journal Article Pattern Recognition Letters, 37 , pp. 135–142, 2014. |
Interactive Handwriting Recognition with Limited User effort Journal Article Intl. Journal on Document Analysis and Recognition (IJDAR), 17 , pp. 47–59, 2014. |
2013 |
A System Architecture to Support Cost-Effective Transcription and Translation of Large Video Lecture Repositories Inproceedings Proc. of the IEEE Intl. Conf. on Systems, Man, and Cybernetics SMC 2013 , pp. 3994-3999, Manchester (UK), 2013. |
2012 |
transLectures Inproceedings Proceedings (Online) of IberSPEECH 2012, pp. 345–351, Madrid (Spain), 2012. |
Albayzin Evaluation: The PRHLT-UPV Audio Segmentation System Inproceedings Proceedings (Online) of IberSPEECH 2012, pp. 596-600, Madrid (Spain), 2012. |
transLectures: Transcription and Translation of Video Lectures Inproceedings Proc. of Cambridge 2012: Innovation and Impact - Openly Collaborating to Enhance Education, pp. 543-546, Cambridge (UK), 2012. |
Character-Based Handwritten Text Recognition of Multilingual Documents Incollection Advances in Speech and Language Technologies for Iberian Languages (IberSpeech 2012), 328 , pp. 187-196, Springer Berlin Heidelberg, 2012, ISBN: 978-3-642-35291-1, (doi: 10.1007/978-3-642-35292-8_20). |
Explicit length modelling for statistical machine translation Journal Article Pattern Recognition, 45 (9), pp. 3183 - 3192, 2012, ISSN: 0031-3203. |
Integrating a State-of-the-Art ASR System into the Opencast Matterhorn Platform Incollection Advances in Speech and Language Technologies for Iberian Languages (IberSpeech 2012), 328 , pp. 237-246, Springer Berlin Heidelberg, 2012, ISBN: 978-3-642-35291-1, (doi: 10.1007/978-3-642-35292-8_20). |
2011 |
Extracción de corpus paralelos de la Wikipedia basada en la obtención de alineamientos bilingües a nivel de frase Inproceedings Proceedings of the Workshop on Iberian Cross-Language Natural Language Processing Tasks (ICL 2011), pp. 14-21, CEUR-WS, 2011, ISSN: 1613-0073. |
Explicit Length Modelling for Statistical Machine Translation Incollection Vitrià, Jordi ; Sanches, JoãoMiguel ; Hernández, Mario (Ed.): Pattern Recognition and Image Analysis (IbPRIA 2011), 6669 , pp. 273-280, Springer Berlin Heidelberg, 2011, ISBN: 978-3-642-21256-7. |
2010 |
Saturnalia: A Latin-Catalan Parallel Corpus for Statistical MT Inproceedings Proc. of the 7th Language Resources and Evaluation Conf. (LREC 2010), pp. 3405–3408, Valleta (Malta), 2010. |
2009 |
Human interaction for high quality machine translation Journal Article Communications of the ACM, 52 (10), pp. 135-138, 2009. |
Statistical approaches to computer-assisted translation Journal Article Computational Linguistics, 35 (1), pp. 3–28, 2009. |
2008 |
Introducing Additional Input Information into IMT Systems Incollection Proceedings of the 5th Joint Workshop on Multimodal Interaction and Related Machine Learning Algorithms. Lecture Notes in Computer Sciences, 5237 , pp. 284–295, Utecht, The Netherlands, 2008. |
Improvements in the Computer Assisted Transciption System of Handwritten Text Images Inproceedings Proc. of the 8th Int. Workshop on Pattern Recognition in Information Systems (PRIS 2008), pp. 103–112, Barcelona (Spain), 2008. |
Improving Interactive Machine Translation via Mouse Actions Inproceedings EMNLP 2008: conference on Empirical Methods in Natural Language Processing, Waikiki, Honolulu, Hawaii, 2008. |