SAM 2 from Meta revolutionizes image and video editing with AI

SAM 2 and SA-V at a glance: Meta simplifies visual editing

On your own blog explains Metawhat the new generation of the Segment Anything Model has to offer. Above all, the increased precision in object marking in images and videos stands out.

The object marking is more precise with SAM 2 than with SAM, © Meta

Meta particularly wanted to optimise the application possibilities in the moving image sector:

SAM was able to learn a general notion of what objects are in images. However, images are only a static snapshot of the dynamic real world in which visual segments can exhibit complex motion. Many important real-world use cases require accurate object segmentation in video data, for example in mixed reality, robotics, autonomous vehicles, and video editing. We believe that a universal segmentation model should be applicable to both images and video (…).

In parallel with SAM 2, Meta launched the SA-V dataset with over 51,000 real-world videos. This is available under the CC BY 4.0 license and offers around 53 times more annotations than the largest video dataset for object marking to date.

On the Meta Blog you can find out more about the technical details, application options, but also limitations of SAM 2. In fast-paced scenarios, object marking can sometimes get lost.

One potential application scenario, however, is the integration into AI Glasses – which Meta itself also offers – whereby individual objects in the user’s view could be immediately recognized and classified.

Object recognition thanks to Meta AI, also in the AI ​​Glasses context, © Meta, GIFs, individual objects such as bowls or a sieve in the kitchen are recognized and marked, partly with text
Object recognition thanks to Meta AI, also in the AI ​​Glasses context, © Meta

After increase in sales and user growth:

Meta wants to push AI in all services

© Penfer - Unsplash, meta lettering, neon pink, blue background
© Penfer – Unsplash



Source: onlinemarketing.de