• Login
    View Item 
    •   DSpace Home
    • Lecture Papers
    • Lecture Papers International Published Articles
    • View Item
    •   DSpace Home
    • Lecture Papers
    • Lecture Papers International Published Articles
    • View Item
    JavaScript is disabled for your browser. Some features of this site may not work without it.

    Design of Conversion Module from Korean Video to English Text for English Subtitle Generator Using Hidden Markov Model Algorithm

    Thumbnail
    View/Open
    Abstract (8.348Kb)
    Cover, TOC, Content (910.8Kb)
    Plagiarism (3.060Mb)
    Date
    2014
    Author
    Citra, Caecilia
    Yuwono, Elizabeth Irenne
    Metadata
    Show full item record
    Abstract
    Based on Youtube data on 2011, total views of Hallyu or Korean cultures related videos have reached 2.3 billion views, thrice compared to 2010 (800 million views). The rising popularity of Korean culture takes effect on the emergence of new English subtitle forums for Korean videos. These forums customers have to wait for minimum 2 days to get new video subtitle. In order to solve this issue, this research is held to build an automatic English subtitle generator for Korean video by combining and implementing today technologies called SubGen (the abbreviation of Subtitle Generator). By using SubGen, customers don’t have to wait days to get English subtitle for Korean videos. We hope this can give contribution for Korean natural language processing technology field. SubGen receives video as input and generate Subrip Text subtitle file (.srt) as output. SubRip Text file contains timing and text spoken for video subtitle. This research is focused on the Korean speech to text process by using the implementation of semi-continuous Hidden Markov Model (SCHMM) algorithm. Training process for this model use phoneme segmentation and forward-backward algorithm in HMM parameter estimation. Resources data for training and transcription are Korean pronunciation dictionary in Sphinx format using hangul syllable, 25 words including numeric and places in 225 sample audio and transcription texts. The training results in 36% accuracy for SCHMM implementation. The overall phases use 16 Korean video varied in content, speakers, and duration as first input. The test results for video to audio conversion, Korean to English translation and subtitle file creation tests is succeed. While speech to text process has 31% accuracy.
    URI
    http://dspace.uc.ac.id/handle/123456789/5874
    Collections
    • Lecture Papers International Published Articles

    Copyright©  2017 - LPPM & Library Of Universitas Ciputra
    »»» UC Town CitraLand, Surabaya - Indonesia 60219 «««
    Powered by : FreeBSD | DSpace | Atmire
     

     

    Browse

    All of DSpaceCommunities & CollectionsBy Issue DateAuthorsTitlesSubjectsThis CollectionBy Issue DateAuthorsTitlesSubjects

    My Account

    Login

    Copyright©  2017 - LPPM & Library Of Universitas Ciputra
    »»» UC Town CitraLand, Surabaya - Indonesia 60219 «««
    Powered by : FreeBSD | DSpace | Atmire