Meta has announced an open-source AI model called ImageBind that connects strings of data, including text, audio, visual data, temperature and movement readings. The core concept of research is linking data into a single multi-dimensional index, a report by The Verge revealed.

This comes after the social media company announced the expansion of ads on its Reels monetisation program to pay creators on Facebook based on the performance of their Reels.

Meta said in a blog post that six types of data included in its new model are: visual (in the form of both image and video); thermal (infrared images); text; audio; depth information; and movement readings generated by an inertial measuring unit, or IMU. 

The social media giant added that the model helps advance AI by enabling machines to analyse many forms of information together. For example, using ImageBind, Meta’s Make-A-Scene could create images from audio, such as creating an image based on the sounds of a rain forest or a bustling market. While Make-A-Scene can generate images by using text prompts, ImageBind could upgrade it to generate images using audio sounds, such as rain.

ImageBind will open access for researchers to develop new holistic systems, such as combining 3D and IMU sensors to design or experience immersive, virtual worlds.

comment COMMENT NOW