Abstract
In this paper, we introduced a system that extracts metadata by recognizing characters and objects in media using deep learning technology. In the field of broadcasting, multimedia contents such as video, audio, image, and text have been converted to digital contents for a long time, but the unconverted resources still remain vast. Building media archives requires a lot of manual work, which is time consuming and costly. Therefore, by implementing a deep learning-based metadata generation system, it is possible to save time and cost in constructing media archives. The whole system consists of four elements: training data generation module, object recognition module, character recognition module, and API server. The deep learning network module and the face recognition module are implemented to recognize characters and objects from the media and describe them as metadata. The training data generation module was designed separately to facilitate the construction of data for training neural network, and the functions of face recognition and object recognition were configured as an API server. We trained the two neural-networks using 1500 persons and 80 kinds of object data and confirmed that the accuracy is 98% in the character test data and 42% in the object data.
| Translated title of the contribution | Implementation of Character and Object Metadata Generation System for Media Archive Construction |
|---|---|
| Original language | Korean |
| Pages (from-to) | 1076-7084 |
| Number of pages | 9 |
| Journal | 방송공학회 논문지 |
| Volume | 24 |
| Issue number | 6 |
| DOIs | |
| State | Published - Nov 2019 |