skip to main content
Article

Early versus late fusion in semantic video analysis

Published: 06 November 2005 Publication History
  • Get Citation Alerts
  • Abstract

    Semantic analysis of multimodal video aims to index segments of interest at a conceptual level. In reaching this goal, it requires an analysis of several information streams. At some point in the analysis these streams need to be fused. In this paper, we consider two classes of fusion schemes, namely early fusion and late fusion. The former fuses modalities in feature space, the latter fuses modalities in semantic space. We show by experiment on 184 hours of broadcast video data and for 20 semantic concepts, that late fusion tends to give slightly better performance for most concepts. However, for those concepts where early fusion performs better the difference is more significant.

    References

    [1]
    A. Amir et al. IBM research TRECVID-2003 video retrieval system. In Proc. TRECVID Workshop, Gaithersburg, USA, 2003.
    [2]
    J. Gauvain, L. Lamel, and G. Adda. The LIMSI broadcast news transcription system. Speech Communication, 37(1--2):89--108, 2002.
    [3]
    G. Iyengar, H. Nock, and C. Neti. Discriminative model fusion for semantic concept detection and annotation in video. In ACM Multimedia, pages 255--258, Berkeley, USA, 2003.
    [4]
    NIST. TREC Video Retrieval Evaluation, 2004. http://www-nlpir.nist.gov/projects/trecvid/.
    [5]
    J. Platt. Probabilities for SV machines. In Advances in Large Margin Classifiers, pages 61--74. MIT Press, 2000.
    [6]
    C. Snoek et al. The MediaMill TRECVID 2004 semantic video search engine. In Proc. TRECVID Workshop, Gaithersburg, USA, 2004.
    [7]
    S. Tsekeridou and I. Pitas. Content-based video parsing and indexing based on audio-visual interaction. IEEE Trans. CSVT, 11(4):522--535, 2001.
    [8]
    V. Vapnik. The Nature of Statistical Learning Theory. Springer-Verlag, NY, USA, 2th edition, 2000.
    [9]
    T. Westerveld et al. A probabilistic multimedia retrieval model and its evaluation. EURASIP JASP, (2):186--197, 2003.
    [10]
    Y. Wu, E. Chang, K.-C. Chang, and J. Smith. Optimal multimodal fusion for multimedia data analysis. In ACM Multimedia, New York, USA, 2004.

    Cited By

    View all
    • (2024)Deep multimodal spatio-temporal Harris Hawk Optimized Pose Recognition framework for self-learning fitness exercisesJournal of Intelligent & Fuzzy Systems10.3233/JIFS-23328646:4(9783-9805)Online publication date: 18-Apr-2024
    • (2024)MIFI: MultI-Camera Feature Integration for Robust 3D Distracted Driver Activity RecognitionIEEE Transactions on Intelligent Transportation Systems10.1109/TITS.2023.330431725:1(338-348)Online publication date: Jan-2024
    • (2024)AM-Bi-LSTM: Adaptive Multi-Modal Bi-LSTM for Sequential RecommendationIEEE Access10.1109/ACCESS.2024.335554812(12720-12733)Online publication date: 2024
    • Show More Cited By

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Conferences
    MULTIMEDIA '05: Proceedings of the 13th annual ACM international conference on Multimedia
    November 2005
    1110 pages
    ISBN:1595930442
    DOI:10.1145/1101149
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Sponsors

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 06 November 2005

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. early fusion
    2. late fusion
    3. multimedia understanding
    4. semantic concept detection

    Qualifiers

    • Article

    Conference

    MM05

    Acceptance Rates

    MULTIMEDIA '05 Paper Acceptance Rate 49 of 312 submissions, 16%;
    Overall Acceptance Rate 995 of 4,171 submissions, 24%

    Upcoming Conference

    MM '24
    The 32nd ACM International Conference on Multimedia
    October 28 - November 1, 2024
    Melbourne , VIC , Australia

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)214
    • Downloads (Last 6 weeks)33

    Other Metrics

    Citations

    Cited By

    View all
    • (2024)Deep multimodal spatio-temporal Harris Hawk Optimized Pose Recognition framework for self-learning fitness exercisesJournal of Intelligent & Fuzzy Systems10.3233/JIFS-23328646:4(9783-9805)Online publication date: 18-Apr-2024
    • (2024)MIFI: MultI-Camera Feature Integration for Robust 3D Distracted Driver Activity RecognitionIEEE Transactions on Intelligent Transportation Systems10.1109/TITS.2023.330431725:1(338-348)Online publication date: Jan-2024
    • (2024)AM-Bi-LSTM: Adaptive Multi-Modal Bi-LSTM for Sequential RecommendationIEEE Access10.1109/ACCESS.2024.335554812(12720-12733)Online publication date: 2024
    • (2024)Emotion recognition in user‐generated videos with long‐range correlation‐aware networkIET Image Processing10.1049/ipr2.13174Online publication date: 10-Jul-2024
    • (2024)Fusing linguistic and acoustic information for automated forensic speaker comparisonScience & Justice10.1016/j.scijus.2024.07.001Online publication date: Jul-2024
    • (2024)A multiscale neural architecture search framework for multimodal fusionInformation Sciences10.1016/j.ins.2024.121005(121005)Online publication date: Jun-2024
    • (2024)Determining the onset of driver’s preparatory action for take-over in automated driving using multimodal dataExpert Systems with Applications10.1016/j.eswa.2024.123153(123153)Online publication date: Jan-2024
    • (2024)Deep multimodal fusion for 3D mineral prospectivity modeling: Integration of geological models and simulation data via canonical-correlated joint fusion networksComputers & Geosciences10.1016/j.cageo.2024.105618188(105618)Online publication date: Jun-2024
    • (2024)A Study of Multimodal Sentiment Analysis and Design of an ArchitectureData Management, Analytics and Innovation10.1007/978-981-97-3242-5_43(659-670)Online publication date: 23-Jul-2024
    • (2024)Multimodal Fusion of Echocardiography and Electronic Health Records for the Detection of Cardiac AmyloidosisArtificial Intelligence in Medicine10.1007/978-3-031-66535-6_25(227-237)Online publication date: 25-Jul-2024
    • Show More Cited By

    View Options

    Get Access

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media