Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/1283880.1283887acmotherconferencesArticle/Chapter ViewAbstractPublication PagessadpiConference Proceedingsconference-collections
Article

Engineering multimedia applications on the basis of multi-structured descriptions of audiovisual contents

Published: 21 May 2007 Publication History
  • Get Citation Alerts
  • Abstract

    We focus our interest on the engineering of multimedia applications whose purpose is to exploit and make best use of the audiovisual heritage by means of prospective exploration of virtual access to audiovisual documents through multi-structured descriptions of these. Multi-structured descriptions are composed of multiple descriptors that are expressed using the FDL (Feria Description Language) object language whose expressive power is emphasized. FDL notably provides a multimedia developer with operations on descriptions and their inner descriptors, as well as temporal aggregation data types. An experimental multimedia application that makes extensive use of FDL concepts and mechanisms is outlined. It explores the use of syncing between the narrative structure of the text of a play and the narrative structure of different broadcasted performances of this play, at multiple granularity levels.

    References

    [1]
    A. Allauzen and J.-L. Gauvain. Adaptation automatique du modèle de langage d'un système de transcription de journaux parlés. Traitement Automatique du Langage, 44(1), 2003.
    [2]
    J. F. Allen. Maintaining knowledge about temporal intervals. Communications of the ACM, 26(11):832--843, 1983.
    [3]
    J. F. Allen. A general model of action and time. Artificial Intelligence, 23(2), 1984.
    [4]
    O. Aubert, P.-A. Champin, and Y. Prié. Integration of semantic web technology in an annotation-based hypervideo system. In Workshop on Semantic Web Annotations for Multimedia (SWAMM'06), Edinburgh, UK, 2006.
    [5]
    O. Aubert and Y. Prié. Advene: active reading through hypervideo. In 16th ACM Conf. on Hypertext and Hypermedia, pages 235--244, Salzburg, Austria, 2005.
    [6]
    G. Auffret, J. Carrive, O. Chevet, and T. Dechilly. Audiovisual-based hypermedia authoring: using structured rerpresentations for efficient access to av documents. In ACM Hypertext '99, Darmstadt, Germany, 1999.
    [7]
    F. Baader, D. Calvanese, D. McGuiness, D. Nardi, and P. Patel-Schneider, editors. The Description Logic Handbook. Cambridge University Press, 2002.
    [8]
    W. Bailer and P. Schallauer. The detailed audiovisual profile: Enabling interoperability between mpeg-7 based systems. In Proceedings of 12th Multimedia Modellinig Conference, pages 217--224, Beijing, China, 2006.
    [9]
    V. Beaudoin. Mètres et Rythmes du Vers Classique. Corneille, Racine. Champion, 2002.
    [10]
    V. Beaudoin and F. Yvon. The metrometer: a tool for analysing french verse. Literary and Linguistic Computing, 11(1), 1996.
    [11]
    H. Bowman, H. Cameron, P. King, and S. Thompson. Mexitl: Multimedia in executable interval temporal logic. Technical report, Computing Laboratory, University of Kent, 1997.
    [12]
    V. Brunie, J. Carrive, and L. Vinet. Ingénierie des documents audiovisuels : le projet feria. TSI, 25(4):469 -- 496, Mai 2006.
    [13]
    M. Caillet. Un système expert d'aide à la classification taxonomique de classes de descripteurs. In Ingénierie des Connaissances, Grenoble, France, July 2007.
    [14]
    J. Carrive, F. Pachet, and R. Ronfard. Clavis - a temporal reasoning system for classification of audiovisual sequences. In Proceedings of RIAO, Paris, France, 2000.
    [15]
    L. Daigle, D. van Gulik, R. Iannella, and P. Faltstrom. Uniform Resource Names (URN) Namespace Definition Mechanisms, October 2002.
    [16]
    R. Deltour and C. Roisin. The limsee3 multimedia authoring model. In DocEng '06, Amsterdam, The Netherlands, 2006.
    [17]
    J. Glass, T. Hazen, S. Cyphers, I. Malioutov, and R. Barzilay. Progress in spoken lecture processing. In Int. Conf. on Spoken Language Processing, Pittsburgh, U.S.A., 2006.
    [18]
    R. Goularte, E. dos Santos Moreira, and M. da Graça C. Pimentel. Structuring interactive tv documents. In DocEng '03, Grenoble, France, 2003.
    [19]
    G. Gravier, F. Yvon, B. Jacob, and F. Bimbot. Sirocco, un système ouvert de reconnaissance de la parole. In XXIVe Journées d'Études sur la Parole (JEP'02), Nancy, France, 2002.
    [20]
    L. Hardman, D. C. A. Bulterman, and G. van Rossum. The amsterdam hypermedia model: adding time and context to the dexter model. Communications if the ACM, 37(2):50--62, 1994.
    [21]
    I. Horrocks, P. F. Patel-Schneider, and F. van Harmelen. From shiq and rdf to owl: The making of a web ontology language. Journal of Web Semantics, 1(1):7--26, 2003.
    [22]
    J. Hunter. Adding multimedia to the semantic web - building an mpeg-7 ontology. In 1st Int. Semantic Web Working Symposium SWWS'01, Stanford, California, USA, 2001.
    [23]
    M. Jourdan, N. Layaïda, C. Roisin, L. Sabry-Ismaïl, and L. Tardif. Madeus, an authoring environment for interactive multimedia documents. In ACM Multimedia '98, Bristol, UK, 1998.
    [24]
    P. King, H. Cameron, H. Bowman, and S. Thompson. Synchronization in multimedia documents. LNCS, 1998.
    [25]
    J. Lewis. Automated lip-sync: Backgrounds and techniques. Visualization and Computer Animation, 2, 1991.
    [26]
    K.-Y. Liu and H.-Y. Chen. Exploring media correlation and synchronization for navigated hypermedia documents. In 13th annual ACM international conference on Multimedia, Singapore, 2005.
    [27]
    C. L. Madhwacharyula, M. Davis, P. Mulhem, and M. S. Kankanhalli. Metadata handling: A video perspective. ACM Transactions on Multimedia Computing, Communications and Applications, 2(4), 2006.
    [28]
    J. M. Martínez, R. Koenen, and F. Pereira. Mpeg-7: The generic multimedia content description standard, part 1. IEEE Multimedia, 9(2):78--87, Avril-Juin 2002.
    [29]
    J. M. Martínez, R. Koenen, and F. Pereira. Mpeg-7: The generic multimedia content description standard, part 2. IEEE Multimedia, 9(3):83--93, Juillet-Septembre 2002.
    [30]
    D. L. McGuinness and F. van Harmelen. Owl web ontology language overview, http://www.w3.org/tr/owl-features/, 2004.
    [31]
    M. Mohri, F. C. N. Pereira, and M. Riley. Weighted finite-state transducers in speech recognition. Computer, Speech and Language, 16(1):69--88, 2002.
    [32]
    R. Ronfard and T. T. Thuong. A framework for aligning and indexing movies with their script. In IEEE Int. Conf. on Multimedia and Expo, Baltimore, Maryland, 2003.
    [33]
    T. K. Shih, L.-J. Hwang, and J.-Y. Tsai. Formal model of temporal properties underlying multimedia presentations. In Multimedia Modeling, 1996.
    [34]
    R. Troncy, W. Bailer, M. Hausenblas, and R. Schlatte. Enabling multimedia metadata interoperability by defining formal semantics of mpeg-7 profiles. In 1st Int. Conf. on Semantic and Digital Media Technologies (SAMT'06), pages 41--55, Athens, Greece, December 2006.
    [35]
    R. Troncy, J. Carrive, S. Lalande, and J.-P. Poli. A motivating scenario for designing an extensible audio-visual description language. In CORIMEDIA'04, Sherbrooke, Canada, Octobre 2004.

    Cited By

    View all
    • (2018)Multimedia applications for playing with digitized theater performancesMultimedia Tools and Applications10.1007/s11042-013-1651-173:3(1777-1793)Online publication date: 31-Dec-2018

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Other conferences
    SADPI '07: Proceedings of the 2007 international workshop on Semantically aware document processing and indexing
    May 2007
    111 pages
    ISBN:9781595936684
    DOI:10.1145/1283880
    • General Chairs:
    • Marc Nanard,
    • Peter King
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    In-Cooperation

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 21 May 2007

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. audiovisual
    2. automatic speech recognition
    3. multi-structured descriptions
    4. multimedia
    5. temporal constraints

    Qualifiers

    • Article

    Conference

    SADPI07

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)1
    • Downloads (Last 6 weeks)0
    Reflects downloads up to

    Other Metrics

    Citations

    Cited By

    View all
    • (2018)Multimedia applications for playing with digitized theater performancesMultimedia Tools and Applications10.1007/s11042-013-1651-173:3(1777-1793)Online publication date: 31-Dec-2018

    View Options

    Get Access

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media