Big Data technology has been one of key engines driving the new industrial revolution. However, the majority of current Big Data research efforts have been devoted to single-modal data analysis, which leads to a huge gap in performance when algorithms are carried out separately. Although significant progress has been made, single-modal data is often insufficient to derive accurate and robust models in many applications.

Multimodal is the most general form for information representation and delivery in a real world. Using multimodal data is natural for humans to make accurate perceptions and decisions. In fact, our digital world is multimodal, combining different modalities of data such as text, audio, images, videos, animations, drawings, depth, 3D, biometrics, interactive content, etc. Multimodal data analytics algorithms often outperform single modal data analytics in many real-world problems.

Multi-sensor information fusion has also been a topic of great interest in industry nowadays. In particular, such companies working on automotive, drone vision, surveillance or robotics have grown exponentially. They are attempting to automate processes by using a wide variety of control signals from various sources.

With the rapid development of Big Data technology and its remarkable applications to many fields, multimodal Big Data is a timely topic. This workshop aims to generate momentum around this topic of growing interest, and to encourage interdisciplinary interaction and collaboration between Natural Language Processing (NLP), computer vision, machine learning, multimedia, robotics, Human-Computer Interaction (HCI), cloud computing, Internet of Things (IoT), and geospatial communities. It serves as a forum to bring together active researchers and practitioners from academia and industry to share their recent advances in this promising area.

Topics

This is an open call for papers, which solicits original contributions considering recent findings in theory, methodologies, and applications in the field of multimodal Big Data. The list of topics includes, but not limited to:

  • Multimodal learning
  • Cross-modal learning
  • Multimodal big data analytics
  • Multimodal big data infrastructure and management
  • Multimodal scene understanding
  • Cross-modal adaptation
  • Multimodal data fusion and data representation
  • Multimodal perception and interaction
  • Multi-modal benchmark datasets and evaluations
  • Multimodal information tracking, retrieval and identification
  • Multimodal object detection, classification, recognition and segmentation
  • Language and vision (e.g., image/video searching and captioning, visual question answering, visual scene understanding, etc.)
  • Biometrics and big data (e.g., face recognition, behavior recognition, eye retina and movement, palm vein and print, etc.)
  • Multimodal applications (e.g., autonomous driving, robotic vision, smart cities, industrial inspection, medical diagnosis, social media, etc.)


Important Dates

  • First Deadline:
    • Oct. 4, 2021: Submission of full papers (7-10 pages) and short papers (5-6 pages)
    • Nov 1, 2021: Notification of paper acceptance
    • Nov 19, 2021: Accepted Papers & Program Schedule
  • Second Deadline (General Workshop Deadline of IEEE Big Data 2021):
    • Oct. 27, 2021: Submission of full papers (7-10 pages) and short papers (5-6 pages)
    • Nov 10, 2021: Notification of paper acceptance
    • Nov 19, 2021: Accepted Papers & Program Schedule
  • IEEE Big Data 2021 Important Deadlines:
    • Nov 21, 2021: Camera-ready of accepted papers
    • Nov 24, 2021: Presentation video record uploading
  • Dec. 15-18, 2021: IEEE Big Data Conference & Workshops (Virtually)


Submission

Please submit to IEEE Big Data 2021 paper submission site.


Program Committee

Program Chair

Lindi Liao, George Mason University, USA

Program Committee members

  • Kaiqun Fu, South Dakota State University, USA
  • Fanchun Jin, Google, USA
  • Ge Jin, Purdue University, USA
  • Emanuela Marasco, George Mason University, USA
  • Jundong Li, University of Virginia, USA
  • Michael D. Porter, University of Virginia, USA
  • Chen Shen, National Institute of Standards and Technology (NIST), USA
  • Ying-Qing Xu, Tsinghua University, China
  • Kai Xing, University of Science and Technology of China