Skip to main content
Log in

Multimedia Tools and Applications - Call for Papers: Recent Advances in AI-Powered Multimedia Visual Computing and Multimodal Signal Processing for Metaverse Era [1247]

Aims and Scope

In the new Era of Metaverse—a collective, virtual shared space created by the convergence of virtually enhanced physical reality and physically persistent virtual reality—understanding and developing techniques in visual computing and multimodal signal processing have never been more crucial. Visual computing makes use of computing abilities of smart devices to process multimedia data such as images and videos. It allows the creation of immersive environments, lifelike avatars, and interactive virtual objects, making the Metaverse experience as real as possible. On the other hand, multimodal signal processing enables the integration and analysis of diverse modality data sources, like cameras, infrared cameras, and LiDAR sensors to enhance features and ensure seamless and intuitive interactions. The integration of Multimedia Visual Computing and Multimodal Signal Processing enables the immersive experiences by delivering interactions that will mutually reflect the real and virtual world.

This special issue is devoted to present the latest developments and breakthroughs in visual computing and multimodal signal processing as they apply to the creation, manipulation, and interaction within the Metaverse. We aim to invite researchers, technologists, and industry professionals from around the globe to contribute high-quality, unpublished, and original research papers on these timely and important topics. 

Topics of interest include, but are not limited to:

  • Data acquisition and processing of immersive media formats, e.g., light field, 360 video, point cloud, multiview, diffusion and AI based generated content (AIGC).
  • Multimedia data application, e.g., NeRF-based rendering, view synthesis, video compression, point cloud reconstruction, and quality assessment, etc.
  • Deep learning techniques for visual computing, e.g., object detection and tracking, and human object interaction, etc.
  • Multimodal signal processing, e.g., RGB, Spectral, LiDAR data registration, segmentation, fusion, and reconstruction, CLIP-based video-image-text modality connecting, etc.
  • Promotable Large Language/Vision Model driven scene understanding and manipulation, e.g., ChatGPT, Llama, Segment Anything Model (SAM), etc.
  • Emerging new trends in VR/AR/MR and multimedia system and their applications.


Important Dates:
Submissions Open: Immediately 
Submission Deadline: January15, 2024


Guest Editors: 

Kejun Wu - Nanyang Technological University

You Yang - Huazhong University of Science and Technology

Suchen Wang - Amazon

Yi Wang - The Hong Kong Polytechnic University

Qiong Liu - Huazhong University of Science and Technology

Liang Liao -  Nanyang Technological University

Kim-Hui Yap - Nanyang Technological University

Lap-Pui Chau - The Hong Kong Polytechnic University


Submission Guidelines:

Authors should prepare their manuscript according to the Instructions for Authors available from the Multimedia Tools and Applications website (this opens in a new tab). Authors should submit through the online submission site at https://www.editorialmanager.com/mtap/default.aspx (this opens in a new tab) and select “SI 1247- Recent Advances in AI-Powered Multimedia Visual Computing and Multimodal Signal Processing for Metaverse Era” when they reach the “Article Type” step in the submission process. Submitted papers should present original, unpublished work, relevant to one of the topics of the special issue. All submitted papers will be evaluated on the basis of relevance, significance of contribution, technical quality, scholarship, and quality of presentation, by at least three independent reviewers. It is the policy of the journal that no submission, or substantially overlapping submission, be published or be under review at another journal or conference at any time during the review process. Please note that the authors of selected papers presented at IEEE MMSP 2023 are invited to submit an extended version of their contributions by taking into consideration both the reviewers’ comments on their conference paper, and the feedback received during presentation at the conference. It is worth clarifying that the extended version is expected to contain a substantial scientific contribution, e.g., in the form of new algorithms, experiments or qualitative/quantitative comparisons, and that neither verbatim transfer of large parts of the conference paper nor reproduction of already published figures will be tolerated. The extended versions of IEEE MMSP 2023 papers will undergo the standard, rigorous journal review process and be accepted only if well-suited to the topic of this special issue and meeting the scientific level of the journal. Final decisions on all papers are made by the Editor in Chief.


Navigation