Subtitling a video/song is a tedious task, not only one has to write the subtitle, but also one has to specify its timing (start and end times). ASTA project tries to automatically determine the subtitles timing based on the, possibly polyphonic, audio input. We found Marsyas to be the most suitable tool for both signal processing (i.e. feature extraction) and machine learning (i.e. training and classification). Beside its efficiency, it provides such a complete solution to audio-analysis that we didn't need any other library. We thank Marsyas team for open-sourcing such a great project.
http://sourceforge.net/projects/asta-annotator/
http://sourceforge.net/projects/asta-annotator/
