Abstract
The purpose of this project is to develop fundamental technologies for building a similarity-aware information environment in which people are able to know similarities among vast amounts of media content. This environment helps establish a “content-symbiotic society” in which media content such as music and video can be created and used in innovative, but ethical ways. Furthermore, by developing technologies for enhancing content appreciation and creation, we aim to promote a society in which people can actively engage in content appreciation and creation, and a content culture that respects past content and emphasizes experiencing emotion. We developed various types of technologies for supporting music appreciation and creation, such as Songle, Songrium, and TextAlive web services, and made those services open to the public.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
M.T. Cover, J.A. Thomas, Elements of Information Theory (Wiley, 2006)
Crypton Future Media. What is the HATSUNE MIKU movement? http://www.crypton.co.jp/download/pdf/info_miku_e.pdf, 2008
M.E.P. Davies, P. Hamel, K. Yoshii, M. Goto, AutoMashUpper: Automatic creation of multi-song music mashups. IEEE/ACM Trans. Audio Speech Lang. Process. 22(12), 1726–1737 (2014)
A. Dobashi, Y. Ikemiya, K. Itoyama, K. Yoshii, A music performance assistance system based on vocal, harmonic, and percussive source separation and content visualization for music audio signals, in Proceedings of SMC, pp. 99–104, 2015
H. Fujihara, M. Goto, T. Kitahara, H.G. Okuno, A modeling of singing voice robust to accompaniment sounds and its application to singer identification and vocal-timbre-similarity-based music information retrieval. IEEE Trans. Audio Speech Lang. Process. 18(3), 638–648 (2010)
H. Fujihara, M. Goto, J. Ogata, K. Komatani, T. Ogata, H.G. Okuno, Automatic synchronization between lyrics and music CD recordings based on Viterbi alignment of segregated vocal signals, in Proceedings of ISM, pp. 257–264, 2006
H. Fujihara, T. Kitahara, M. Goto, K. Komatani, T. Ogata, H.G. Okuno, F0 estimation method for singing voice in polyphonic audio signal based on statistical vocal model and Viterbi search, in Proceedings of ICASSP 2006, pp. V–253–256, 2006
S. Fukayama, M. Goto, Automated choreography synthesis using a gaussian process leveraging consumer-generated dance motions, in Proceedings of ACE, 2014
S. Fukayama, M. Goto, Music content driven automated choreography with beat-wise motion connectivity constraints, in Proceedings of SMC, pp. 177–183, 2015
M. Goto, A real-time music scene description system: predominant-F0 estimation for detecting melody and bass lines in real-world audio signals. Speech Commun. 43(4), 311–329 (2004)
M. Goto, A chorus-section detection method for musical audio signals and its application to a music listening station. IEEE Trans. Audio Speech Lang. Process. 14(5), 1783–1794 (2006)
M. Goto, Active music listening interfaces based on signal processing, in Proceedings of ICASSP, 2007
M. Goto, K. Yoshii, H. Fujihara, M. Mauch, T. Nakano, Songle: A web service for active music listening improved by user contributions, in Proceedings of ISMIR, pp. 311–316, 2011
M. Goto, K. Yoshii, T. Nakano, S. Widget, Making animation and physical devices synchronized with music videos on the web, in Proceedings of IEEE ISM, pp. 85–88, 2015
M. Hamasaki, M. Goto, Songrium: A music browsing assistance service based on visualization of massive open collaboration within music content creation community, in Proc. of the 9th International Symposium on Open Collaboration (WikiSym + OpenSym 2013), pp. 1–10, 2013
M. Hamasaki, M. Goto, T. Nakano, Songrium: A music browsing assistance service with interactive visualization and exploration of a web of music, in Proceedings of the 23rd International World Wide Web Conference (WWW 2014), pp. 523–528, 2014
M. Hamasaki, M. Goto, T. Nakano, Songrium: browsing and listening environment for music content creation community, in Proceedings of SMC, pp. 23–30, 2015
M. Hamasaki, H. Takeda, T. Hope, T. Nishimura, Network analysis of an emergent massively collaborative creation community: How can people create videos collaboratively without collaboration?, in Proceedings of ICWSM, pp. 222–225, 2009
M. Hamasaki, H. Takeda, T. Nishimura, Network analysis of massively collaborative creation of multimedia contents—case study of Hatsune Miku videos on Nico Nico Douga, in Proceedings of uxTV, pp. 165–168, 2008
T. Hirai, S. Nakamura, T. Yumura, S. Morishima, VRMixer: Mixing video and real world with video segmentation, in Proceedings of ACE, 2014
Y. Ikemiya, K. Yoshii, K. Itoyama, Singing voice analysis and editing based on mutually dependent f0 estimation and source separation, in Proceedings of IEEE ICASSP, pp. 574–578, 2015
J. Kato, T. Igarashi, M. Goto, Programming with examples to develop data-intensive user interfaces. IEEE Comput. 49(7), 34–42 (2016)
J. Kato, T. Nakano, M. Goto, TextAlive: Integrated design environment for kinetic typography, in Proceedings of ACM CHI, pp. 3403–3412, 2015
J. Kato, T. Nakano, M. Goto, TextAlive Online: Live programming of kinetic typography videos with online music, in Proceedings of ICLC, pp. 199–205, 2015
H. Kenmochi, H. Ohshita, Vocaloid—commercial singing synthesizer based on sample concatenation, in Proceedings of Interspeech, pp. 4010–4011, 2007
M. Mauch, S. Dixon, Approximate note transcription for the improved identification of difficult chords, in Proceedings of ISMIR, pp. 135–140, 2010
M. Mauch, S. Dixon, Simultaneous estimation of chords and musical context from audio. IEEE Trans. ASLP 18(6), 1280–1289 (2010)
M. McVicar, S. Fukayama, M. Goto, AutoGuitarTab: computer-aided composition of rhythm and lead guitar parts in the tablature space. IEEE/ACM Trans. Audio Speech Lang. Process. 23(7), 1105–1117 (2015)
D. Mochihashi, E. Sumita, The infinite Markov model, in Proceedings of Advances in Neural Information Processing Systems 20 (NIPS 2007), pp. 1017–1024, 2007
T. Nakamura, H. Kameoka, K. Yoshii, M. Goto, Timbre replacement of harmonic and drum components for music audio signals, in Proceedings of IEEE ICASSP, pp. 7520–7524, 2014
T. Nakano, M. Goto, VocaRefiner: an interactive singing recording system with integration of multiple singing recordings, in Proceedings of SMC, pp. 115–122, 2013
T. Nakano, M. Goto, LyricListPlayer: a consecutive-query-by-playback interface for retrieving similar word sequences from different song lyrics, in Proceedings of SMC, pp. 344–349, 2016
T. Nakano, J. Kato, M. Hamasaki, M. Goto, PlaylistPlayer: An interface using multiple criteria to change the playback order of a music playlist, in Proceedings of ACM IUI, pp. 186–190, 2016
T. Nakano, D. Mochihashi, K. Yoshii, M. Goto, Musical typicality: how many similar songs exist?, in Proceedings of ISMIR, pp. 695–701, 2016
T. Nakano, K. Yoshii, M. Goto, Vocal timbre analysis using latent dirichlet allocation and cross-gender vocal timbre similarity, in Proceedings of IEEE ICASSP, pp. 5239–5343, 2014
T. Nakano, K. Yoshii, M. Goto, Musical similarity and commonness estimation based on probabilistic generative models, in Proceedings of IEEE ISM, pp. 197–204, 2015
T. Nakano, K. Yoshii, M. Goto, Musical similarity and commonness estimation based on probabilistic generative models of musical elements. Int. J. Semant. Comput. (IJSC) 10(1), 27–52 (2016)
G. Percival, S. Fukayama, M. Goto, Song2Quartet: a system for generating string quartet cover songs from polyphonic audio of popular music, in Proceedings of ISMIR, pp. 114–120, 2015
S. Sasaki, K. Yoshii, T. Nakano, M. Goto, S. Morisihima, LyricsRadar: a lyrics retrieval system based on latent topics of lyrics, Proceedings of ISMIR, pp. 585–590, 2014
H. Sato, T. Hirai, T. Nakano, M. Goto, S. Morishima, A soundtrack generation system to synchronize the climax of a video clip with music, in Proceedings of IEEE ICME, 2016
J.B.L. Smith, G. Percival, J. Kato, M. Goto, S. Fukayama, Generating and unscrambling music mashups with real-time interactivity. CrossSong Puzzle, in Proceedings of SMC, pp. 61–67, 2015
K. Tsukuda, M. Goto, ExploratoryVideoSearch: a music video search system based on coordinate terms and diversification, in Proceedings of IEEE ISM, pp. 221–224, 2015
K. Tsukuda, M. Hamasaki, M. Goto, SmartVideoRanking: video search by mining emotions from time-synchronized comments, in Proceedings of IEEE ICDMW, 2016
K. Yoshii, M. Goto, A vocabulary-free infinity-gram model for nonparametric bayesian chord progression analysis, in Proceedings of ISMIR, pp. 645–650, 2011
Acknowledgements
This research was supported in part by CREST, JST. I would like to thank all those involved in the OngaCREST Project, especially co-researchers Shigeo Morishima, Satoshi Nakamura, and Kazuyoshi Yoshii, and the following who devoted their energies to developing the web services: Masahiro Hamasaki, Tomoyasu Nakano, Satoru Fukayama, Jun Kato, Kosetsu Tsukuda, Yuta Kawasaki, Keisuke Ishida, and Takahiro Inoue. I also would like to extend my appreciation to the Research Supervisor, Professor Toyoaki Nishida and Late Professor Yoh’ichi Tohkura, and Advisors of the research area of “Creation of Human-Harmonized information Technology for Convivial Society” of CREST, JST.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2017 Springer Japan KK
About this chapter
Cite this chapter
Goto, M. (2017). OngaCREST Project: Building a Similarity-Aware Information Environment for a Content-Symbiotic Society. In: Nishida, T. (eds) Human-Harmonized Information Technology, Volume 2. Springer, Tokyo. https://doi.org/10.1007/978-4-431-56535-2_1
Download citation
DOI: https://doi.org/10.1007/978-4-431-56535-2_1
Published:
Publisher Name: Springer, Tokyo
Print ISBN: 978-4-431-56533-8
Online ISBN: 978-4-431-56535-2
eBook Packages: Computer ScienceComputer Science (R0)