Very good piece, with as you might expect, good demonstration video. This takes such methods beyond captioning and close to what we continually do as humans, visually interpret and understand a contextually changing view. A sort of visual scene understanding. Sensory AI at its most useful.
Helping AI master video understanding By Dan Gutfreund
Video Analytics Scientist, IBM Research AI
I am part of the team at the MIT IBM Watson AI Lab that is carrying out fundamental AI research to push the frontiers of core technologies that will advance the state-of-the-art in AI video comprehension. This is just one example of joint research we’re pursuing together to produce innovations in AI technology that solve real business challenges.
Great progress has been made and I am excited to share that we are releasing the Moments in Time Dataset, a large-scale dataset of one million three-second annotated video clips for action recognition to accelerate the development of technologies and models that enable automatic video understanding for AI.
A lot can happen in a moment of time: a girl kicking a ball, behind her on the path a woman walks her dog, on a park bench nearby a man is reading a book and high above a bird flies in the sky. Humans constantly absorb such moments through their senses and process them swiftly and effortlessly. When asked to describe such a moment, a person can quickly identify objects (girl, ball, bird, book), the scene (park) and the actions that are taking place (kicking, walking, reading, flying). ... "
Subscribe to:
Post Comments (Atom)
No comments:
Post a Comment