Área de Sistemas de la Información y las Comunicaciones

Loading...
OrgUnit Logo
Date established
City
Country
ID
Description

Publication Search Results

Now showing 1 - 10 of 12
  • Publication
    Efficient Generation of High-Quality Multilingual Subtitles for Video Lecture Repositories
    (Springer Verlag (Germany), 2015-11-26) Valor Miró, Juan Daniel; Silvestre Cerdà, Joan Albert; Civera Saiz, Jorge; Turró Ribalta, Carlos; Juan Císcar, Alfonso; Dpto. de Sistemas Informáticos y Computación; Área de Sistemas de la Información y las Comunicaciones; Escuela Politécnica Superior de Alcoy; Escuela Técnica Superior de Ingeniería Informática; Instituto Universitario Valenciano de Investigación en Inteligencia Artificial; European Commission; Ministerio de Economía y Competitividad
    Video lectures are a valuable educational tool in higher education to support or replace face-to-face lectures in active learning strategies. In 2007 the Universitat Politècnica de València (UPV) implemented its video lecture capture system, resulting in a high quality educational video repository, called poliMedia, with more than 10.000 mini lectures created by 1.373 lecturers. Also, in the framework of the European project transLectures, UPV has automatically generated transcriptions and translations in Spanish, Catalan and English for all videos included in the poliMedia video repository. transLectures’s objective responds to the widely-recognised need for subtitles to be provided with video lectures, as an essential service for non-native speakers and hearing impaired persons, and to allow advanced repository functionalities. Although high-quality automatic transcriptions and translations were generated in transLectures, they were not error-free. For this reason, lecturers need to manually review video subtitles to guarantee the absence of errors. The aim of this study is to evaluate the efficiency of the manual review process from automatic subtitles in comparison with the conventional generation of video subtitles from scratch. The reported results clearly indicate the convenience of providing automatic subtitles as a first step in the generation of video subtitles and the significant savings in time of up to almost 75 % involved in reviewing subtitles.
  • Publication
    Efficiency and usability study of innovative computer-aided transcription strategies for video lecture repositories
    (Elsevier, 2015-11) Valor Miró, Juan Daniel; Silvestre Cerdà, Joan Albert; Civera Saiz, Jorge; Turró Ribalta, Carlos; Juan Císcar, Alfonso; Dpto. de Sistemas Informáticos y Computación; Área de Sistemas de la Información y las Comunicaciones; Escuela Politécnica Superior de Alcoy; Escuela Técnica Superior de Ingeniería Informática; Instituto Universitario Valenciano de Investigación en Inteligencia Artificial; European Commission; Ministerio de Economía y Competitividad
    [EN] Video lectures are widely used in education to support and complement face-to-face lectures. However, the utility of these audiovisual assets could be further improved by adding subtitles that can be exploited to incorporate added-value functionalities such as searchability, accessibility, translatability, note-taking, and discovery of content-related videos, among others. Today, automatic subtitles are prone to error, and need to be reviewed and post-edited in order to ensure that what students see on-screen are of an acceptable quality. This work investigates different user interface design strategies for this post-editing task to discover the best way to incorporate automatic transcription technologies into large educational video repositories. Our three-phase study involved lecturers from the Universitat Polite`cnica de Vale`ncia (UPV) with videos available on the poliMedia video lecture repository, which is currently over 10,000 video objects. Simply by conventional post-editing automatic transcriptions users almost reduced to half the time that would require to generate the transcription from scratch. As expected, this study revealed that the time spent by lecturers reviewing automatic transcriptions correlated directly with the accuracy of said transcriptions. However, it is also shown that the average time required to perform each individual editing operation could be precisely derived and could be applied in the definition of a user model. In addition, the second phase of this study presents a transcription review strategy based on confidence measures (CM) and compares it to the conventional post-editing strategy. Finally, a third strategy resulting from the combination of that based on CM with massive adaptation techniques for automatic speech recognition (ASR), achieved to improve the transcription review efficiency in comparison with the two aforementioned strategies. 2015 Elsevier B.V. All rights reserved.
  • Publication
    Generación eficiente de transcripciones y traducciones automáticas en poliMedia
    (Editorial Universitat Politècnica de València, 2016-07-19) Valor Miró, Juan; Turró Ribalta, Carlos; Civera Saiz, Jorge; Juan Císcar, Alfons; Dpto. de Sistemas Informáticos y Computación; Área de Sistemas de la Información y las Comunicaciones; Escuela Técnica Superior de Ingeniería Informática; Instituto Universitario Valenciano de Investigación en Inteligencia Artificial
    [EN] Usage of educational videos in higher education has raised quickly for different educational applications, which leads to platforms and services such as poliMedia, of the Polytechnic University of Valencia (UPV), which enables the creation, publication and dissemination of this educational multimedia content. Through various research projects, and specifically the European project transLectures, UPV implemented a system that automatically generates subtitles in various languages for all poliMedia videos. Creation of these subtitles is made by an automatic speech recognition and machine translation system that provides high accuracy in both recognition and translation to the main European languages. Transcripts and translations are not only used to improve accessibility, but also allow search and retrieval of video contents within the video portal. So a user can locate the video, and the time within it, where a certain word is said for later viewing. In this article we also extend previous work in the assessment of the review process, including transcription of French and Spanish to Catalan translation.
  • Publication
    Evaluación del proceso de revisión de transcripciones automáticas para vídeos Polimedia
    (Editorial Universitat Politècnica de València, 2014-07-15) Valor Miró, Juan Daniel; Nadine Spencer, Rachel; Pérez González de Martos, Alejandro Manuel; Garcés Díaz-Munío, Gonzalo Vicente; Turró Ribalta, Carlos; Civera Saiz, Jorge; Juan Císcar, Alfonso; Dpto. de Sistemas Informáticos y Computación; Área de Sistemas de la Información y las Comunicaciones; Escuela Técnica Superior de Ingeniería Informática; Escuela de Doctorado; Instituto Universitario Valenciano de Investigación en Inteligencia Artificial; European Commission
    [EN] Video lectures are a tool of proven value and wide acceptance in universities that are leading to platforms like poliMedia. transLectures is a European project that generates automatic high-quality transcriptions and translations for the poliMedia platform, and improve them by using massive adaptation and intelligent interaction techniques. In this paper we present the evaluation with lecturers carried out under the Doc`encia en Xarxa 2012-2013 call, with the aim to study the process of supervise transcriptions, compared with to transcribe from scratch.
  • Publication
    Doblaje automático de vídeo-charlas educativas en UPV[Media]
    (Editorial Universitat Politècnica de València, 2023-01-09T07:40:08Z) Pérez González de Martos, Alejandro Manuel; Giménez Pastor, Adrián; Jorge Cano, Javier; Iranzo Sánchez, Javier; Silvestre Cerdà, Joan Albert; Garcés Díaz-Munío, Gonzalo Vicente; Baquero Arnal, Pau; Sanchis Navarro, José Alberto; Civera Saiz, Jorge; Juan Císcar, Alfonso; Turró Ribalta, Carlos; Dpto. de Sistemas Informáticos y Computación; Área de Sistemas de la Información y las Comunicaciones; Escuela Politécnica Superior de Alcoy; Escuela Técnica Superior de Ingeniería Informática; Escuela de Doctorado; Instituto Universitario Valenciano de Investigación en Inteligencia Artificial; Agencia Estatal de Investigación; European Commission
    [EN] More and more universities are banking on the production of digital contents to support online or blended learning in higher education. Over the last years, the MLLP research group has been working closely with the UPV’s ASIC media services in order to enrich educational multimedia resources through the application of natural language processing technologies including automatic speech recognition, machine translation and text-tospeech. In this work we present the steps that are being followed for the comprehensive translation of these materials, specifically through (semi-)automatic dubbing by making use of state-of-the-art speaker-adaptive text-to-speech technologies.
  • Publication
    A System Architecture to Support Cost-Effective Transcription and Translation of Large Video Lecture Repositories
    (IInstitute of Electrical and Electronics Engineers (IEEE), 2013) Silvestre Cerdà, Joan Albert; Pérez González de Martos, Alejandro Manuel; Jiménez López, Manuel; Turró Ribalta, Carlos; Juan Císcar, Alfonso; Civera Saiz, Jorge; Dpto. de Sistemas Informáticos y Computación; Área de Sistemas de la Información y las Comunicaciones; Escuela Politécnica Superior de Alcoy; Escuela Técnica Superior de Ingeniería Informática; Instituto Universitario Valenciano de Investigación en Inteligencia Artificial; European Commission; Ministerio de Educación, Cultura y Deporte
    [EN] Online video lecture repositories are rapidly growing and becoming established as fundamental knowledge assets. However, most lectures are neither transcribed nor translated because of the lack of cost-effective solutions that can give accurate enough results. In this paper, we describe a system architecture that supports the cost-effective transcription and translation of large video lecture repositories. This architecture has been adopted in the EU project transLectures and is now being tested on a repository of more than 9000 video lectures at the Universitat Politecnica de Valencia. Following a brief description of this repository and of the transLectures project, we describe the proposed system architecture in detail. We also report empirical results on the quality of the transcriptions and translations currently being maintained and steadily improved.
  • Publication
    Evaluación de la revisión de transcripciones y traducciones automáticas de vídeos poliMedia
    (Editorial Universitat Politècnica de València, 2015-06-30) Valor Miró, Juan Daniel; Turró Ribalta, Carlos; Civera Saiz, Jorge; Juan Císcar, Alfonso; Dpto. de Sistemas Informáticos y Computación; Área de Sistemas de la Información y las Comunicaciones; Escuela Técnica Superior de Ingeniería Informática; Instituto Universitario Valenciano de Investigación en Inteligencia Artificial; European Commission; Ministerio de Economía y Competitividad
    [ES] Los vídeos docentes son una herramienta de gran aceptación en el mundo universitario, que se integran en metodologías activas y innovadoras, dando pie a plataformas como poliMedia, plataforma de la Universitat Politècnica de València (UPV) que permite la creación, publicación y difusión de este tipo de contenido multimedia. En el marco del proyecto europeo transLectures, la UPV ha generado automáticamente transcripciones y traducciones en español, catalán e inglés para todos los vídeos incluidos en el repositorio poliMedia. Las transcripciones y traducciones automáticas generadas tienen una alta calidad. Sin embargo, estas transcripciones y traducciones poseen errores inherentes al proceso automático de subtitulado utilizado, por lo que en ocasiones puede ser necesaria una revisión manual posterior para garantizar la ausencia de errores. El objetivo de este trabajo es evaluar este proceso de revisión manual para poder compararlo en términos de coste temporal con un proceso de obtención completamente manual. Así pues, en el marco de las ayudas de la UPV de Docencia en Red 2013-2014, pusimos en marcha una campaña de evaluación del proceso de revisión de transcripciones y traducciones automáticas por parte del profesorado, cuyos resultados indican inequívocamente la eficacia de estas técnicas y el importante ahorro de tiempo que suponen.
  • Publication
    Hacia la traducción integral de vídeo charlas educativas
    (Editorial Universitat Politècnica de València, 2017-07-18) Piqueras, Santiago; Pérez González de Martos, Alejandro Manuel; Turró Ribalta, Carlos; Jimenez, Manuel; Sanchis Navarro, José Alberto; Civera Saiz, Jorge; Juan Císcar, Alfonso; Dpto. de Sistemas Informáticos y Computación; Área de Sistemas de la Información y las Comunicaciones; Escuela Técnica Superior de Ingeniería Informática; Instituto Universitario Valenciano de Investigación en Inteligencia Artificial; European Commission; Ministerio de Economía y Competitividad; Generalitat Valenciana
    [EN] More and more universities and educational institutions are banking on production of technological resources for different uses in higher education. The MLLP research group has been working closely with the ASIC at UPV in order to enrich educational multimedia resources through the use of machine learning technologies, such as automatic speech recognition, machine translation or text-to-speech synthesis. In this work, developed under the Plan de Docencia en Red 2016-17’s framework, we present the application of innovative technologies in order to achive the integral translation of educational videos.
  • Publication
    Towards cross-lingual voice cloning in higher education
    (Elsevier, 2021-10) Pérez-González de Martos, Alejandro Manuel; Garcés Díaz-Munío, Gonzalo Vicente; Giménez Pastor, Adrián; Silvestre Cerdà, Joan Albert; Sanchis Navarro, José Alberto; Civera Saiz, Jorge; Jiménez, Manuel; Turró Ribalta, Carlos; Juan Císcar, Alfonso; Dpto. de Sistemas Informáticos y Computación; Área de Sistemas de la Información y las Comunicaciones; Escuela Politécnica Superior de Alcoy; Escuela Técnica Superior de Ingeniería Informática; Escuela de Doctorado; Instituto Universitario Valenciano de Investigación en Inteligencia Artificial; AGENCIA ESTATAL DE INVESTIGACION; European Regional Development Fund; COMISION DE LAS COMUNIDADES EUROPEA; Universitat Politècnica de València
    [EN] The rapid progress of modern AI tools for automatic speech recognition and machine translation is leading to a progressive cost reduction to produce publishable subtitles for educational videos in multiple languages. Similarly, text-to-speech technology is experiencing large improvements in terms of quality, flexibility and capabilities. In particular, state-of-the-art systems are now capable of seamlessly dealing with multiple languages and speakers in an integrated manner, thus enabling lecturer¿s voice cloning in languages she/he might not even speak. This work is to report the experience gained on using such systems at the Universitat Politècnica de València (UPV), mainly as a guidance for other educational organizations willing to conduct similar studies. It builds on previous work on the UPV¿s main repository of educational videos, MediaUPV, to produce multilingual subtitles at scale and low cost. Here, a detailed account is given on how this work has been extended to also allow for massive machine dubbing of MediaUPV. This includes collecting 59 h of clean speech data from UPV¿s academic staff, and extending our production pipeline of subtitles with a state-of-the-art multilingual and multi-speaker text-to-speech system trained from the collected data. Our main result comes from an extensive, subjective evaluation of this system by lecturers contributing to data collection. In brief, it is shown that text-to-speech technology is not only mature enough for its application to MediaUPV, but also needed as soon as possible by students to improve its accessibility and bridge language barriers.
  • Publication
    Multilingual videos for MOOCs and OER
    (International Forum of Educational Technology & Society, 2018) Valor Miró, Juan Daniel; Baquero-Arnal, Pau; Civera Saiz, Jorge; Turró Ribalta, Carlos; Juan Císcar, Alfonso; Dpto. de Sistemas Informáticos y Computación; Área de Sistemas de la Información y las Comunicaciones; Escuela Técnica Superior de Ingeniería Informática; Instituto Universitario Valenciano de Investigación en Inteligencia Artificial; Ministerio de Economía y Empresa; European Commission
    [EN] Massive Open Online Courses (MOOCs) and Open Educational Resources (OER) are rapidly growing, but are not usually offered in multiple languages due to the lack of cost-effective solutions to translate the different objects comprising them and particularly videos. However, current state-of-the-art automatic speech recognition (ASR) and machine translation (MT) techniques have reached a level of maturity which opens the possibility of producing multilingual video subtitles of publishable quality at low cost. This work summarizes authors' experience in exploring this possibility in two real-life case studies: a MOOC platform and a large video lecture repository. Apart from describing the systems, tools and integration components employed for such purpose, a comprehensive evaluation of the results achieved is provided in terms of quality and efficiency. More precisely, it is shown that draft multilingual subtitles produced by domainadapted ASR/MT systems reach a level of accuracy that make them worth post-editing, instead of generating them ex novo, saving approximately 25%-75% of the time. Finally, the results reported on user multilingual data consumption reflect that multilingual subtitles have had a very positive impact in our case studies boosting student enrolment, in the case of the MOOC platform, by 70% relative.