visual cognition - http://archive.pkmital.com https://archive.pkmital.com computational audiovisual augmented reality research Mon, 09 Feb 2015 20:12:21 +0000 en-US hourly 1 https://wordpress.org/?v=6.9.4 C.A.R.P.E. version 0.1.1 release https://archive.pkmital.com/2015/02/09/c-a-r-p-e-version-0-1-1-release/ https://archive.pkmital.com/2015/02/09/c-a-r-p-e-version-0-1-1-release/#respond Mon, 09 Feb 2015 20:01:12 +0000 http://pkmital.com/home/?p=1887 Screen Shot 2015-02-09 at 2.59.14 PM

I’ve updated C.A.R.P.E., a graphical tool for visualizing eye-movements and processing audio/video, to include a graphical timeline (thanks to ofxTimeline by James George/YCAM), support for audio playback/scrubbing (using pkmAudioWaveform), audio saving, and various bug fixes. This release has changed some parameters of the XML file and added others. Please refer to this example XML file for how to setup your own data:

See my previous post for information on the initial release.

Please fill out the following form if you’d like to use C.A.R.P.E..:

Continue reading...

The post C.A.R.P.E. version 0.1.1 release first appeared on http://archive.pkmital.com.

]]>
https://archive.pkmital.com/2015/02/09/c-a-r-p-e-version-0-1-1-release/feed/ 0
Toolkit for Visualizing Eye-Movements and Processing Audio/Video https://archive.pkmital.com/2015/02/06/toolkit-for-visualizing-eye-movements-and-processing-audio-video/ https://archive.pkmital.com/2015/02/06/toolkit-for-visualizing-eye-movements-and-processing-audio-video/#comments Fri, 06 Feb 2015 23:53:18 +0000 http://pkmital.com/home/?p=1852 Screen Shot 2015-02-06 at 6.24.27 PM

Original video still without eye-movements and heatmap overlay copyright Dropping Knowledge Video Republic.

From 2008 – 2010, I worked on the Dynamic Images and Eye-Movements (D.I.E.M.) project, led by John Henderson, with Tim Smith and Robin Hill. We worked together to collect nearly 200 participants eye-movements on nearly 100 short films from 30 seconds to 5 minutes in length. The database is freely available and covers a wide range of film styles form advertisements, to movie and music trailers, to news clips. During my time on the project, I developed an open source toolkit, C.A.R.P.E. to complement D.I.E.M., or Computational Algorithmic Representation and Processing of Eye-movements (Tim’s idea!), for visualizing and processing the data we collected, and used it for writing up a journal paper describing a strong correlation between tightly clustered eye-movements and the motion in a scene. We also output visualizations of our entire corpus on our Vimeo channel. The project came to a halt and so did the visualization software. I’ve since picked up the ball and re-written it entirely from the ground up.

The image below shows how you can represent the movie, the motion in the scene of the movie (represented in … Continue reading...

The post Toolkit for Visualizing Eye-Movements and Processing Audio/Video first appeared on http://archive.pkmital.com.

]]>
https://archive.pkmital.com/2015/02/06/toolkit-for-visualizing-eye-movements-and-processing-audio-video/feed/ 4
Intention in Copyright https://archive.pkmital.com/2012/06/29/intention-in-copyright/ https://archive.pkmital.com/2012/06/29/intention-in-copyright/#respond Fri, 29 Jun 2012 14:49:27 +0000 http://pkmital.com/home/?p=1106 The following article is written for the LUCID Studio for Speculative Art based in India.

Introduction

My work in audiovisual resynthesis aims to create models of how humans represent and attend to audiovisual scenes. Using pattern recognition of both audio and visual material, these models use large corpora of learned audiovisual material which can be matched to ongoing streams of incoming audio or visual material. The way audio and visual material is stored and segmented within the model is based heavily on neurobiology and behavioral evidence (the details are saved for another post). I have called the underlying model Audiovisual Content-based Information Description/Distortion (or ACID for short).

As an example, a live stream of audio may be matched to a database of learned sounds from recordings of nature, creating a re-synthesis of the audio environment at present using only pre-recorded material from nature itself. These learned sounds may be fragments of a bird chirping, or the sound of footsteps. Incoming sounds of someone talking may then be synthesized using the closest sounding material to that person talking, perhaps a bird chirp or a footstep. Instead of a live stream, one can also re-synthesize a pre-recorded stream. Consider using a database … Continue reading...

The post Intention in Copyright first appeared on http://archive.pkmital.com.

]]>
https://archive.pkmital.com/2012/06/29/intention-in-copyright/feed/ 0
Concatenative Video Synthesis (or Video Mosaicing) https://archive.pkmital.com/2011/10/08/concatenative-video-synthesis-or-video-mosaicing/ https://archive.pkmital.com/2011/10/08/concatenative-video-synthesis-or-video-mosaicing/#comments Sat, 08 Oct 2011 10:08:47 +0000 http://pkmital.com/home/?p=830 prototype

Working closely with my adviser Mick Grierson, I have developed a way to resynthesize existing videos using material from another set of videos. This process starts by learning a database of objects that appear in the set of videos to synthesize from. The target video to resynthesize is then broken into objects in a similar manner, but also matched to objects in the database. What you get is a resynthesis of the video that appears as beautiful disorder. Here are two examples, the first using Family Guy to resynthesize The Simpsons. And the second using Jan Svankmajer’s Food to resynthesize Jan Svankmajer’s Dimensions of Dialogue.

Continue reading...

The post Concatenative Video Synthesis (or Video Mosaicing) first appeared on http://archive.pkmital.com.

]]>
https://archive.pkmital.com/2011/10/08/concatenative-video-synthesis-or-video-mosaicing/feed/ 2
Tim J Smith guest blogs for David Bordwell https://archive.pkmital.com/2011/02/20/tim-j-smith-guest-blogs-for-david-bordwell/ https://archive.pkmital.com/2011/02/20/tim-j-smith-guest-blogs-for-david-bordwell/#respond Sun, 20 Feb 2011 03:43:36 +0000 http://pkmital.com/home/?p=540 Tim J Smith, expert in scene perception and film cognition, and of The DIEM project [1] recently starred as a guest blogger for David Bordwell, a leading film theorist with an impressive list of books and publications widely used in film cognition/film art research/studies [2]. In his article featured on David’s site, Tim expands on his research on film cognition including continuity editing [3], attentional synchrony [4], and the project we worked on in 2008-2010 as part of The DIEM Project. Since Tim’s feature on David Bordwell’s blog, The DIEM Project saw a surge of publicity and our vimeo video loads going higher than 200,000 in a single day and features on dvice, slashfilm, gizmodo, Rogert Ebert’s facebook/twitter, and the front page of imbd.com.

Not to mention, our tools and visualizations are finally reaching an audience with interests in film, photography, and cognition. If you haven’t yet seen some of our videos, please head on over to our vimeo page, where you can see a range of videos embedded with eye-tracking of participants and many different visualizations of models of eye-movements using machine learning, or start by reading Tim’s post on Continue reading...

The post Tim J Smith guest blogs for David Bordwell first appeared on http://archive.pkmital.com.

]]>
https://archive.pkmital.com/2011/02/20/tim-j-smith-guest-blogs-for-david-bordwell/feed/ 0
6DOF Head Tracking https://archive.pkmital.com/2010/11/18/6dof-head-tracking/ https://archive.pkmital.com/2010/11/18/6dof-head-tracking/#respond Thu, 18 Nov 2010 01:41:26 +0000 http://pkmital.com/home/?p=420 The following demo works with SeeingMachines FaceAPI in openFrameworks controlling a Mario avatar.  It also has some really poor gesture recognition (and learning but it’s not shown here), though a threshold on the rotation DOF would have produced better results for the simple task of looking up/down left/right gestures.

6DOF Head Tracking from pkmital on Vimeo.

interfacing seeingmachines faceapi with openFrameworks to control a 3D mario avatar

This is just with the non-commercial license. The full commercial license (~$3000?) gives you access to lip/mouth tracking and eye-brows, as well as much more flexibility in how to use their api with different/multiple cameras and accessing image data.

Of course, there are other initiatives at producing similar results. Mutual information based template trackers, for instance, seem to be state-of-art. Take a look at recent work by Panin and Knoll using OpenTL:

 

I imagine a lot of people would like this technology.… Continue reading...

The post 6DOF Head Tracking first appeared on http://archive.pkmital.com.

]]>
https://archive.pkmital.com/2010/11/18/6dof-head-tracking/feed/ 0
Facebook Graph API https://archive.pkmital.com/2010/07/19/facebook-graph-api/ https://archive.pkmital.com/2010/07/19/facebook-graph-api/#respond Sun, 18 Jul 2010 23:25:37 +0000 http://pkmital.com/home/?p=352 If you are one of the +500 million users of facebook, and you know your user id, try plugging it in here: http://zesty.ca/facebook/

This uses the Facebook Graph API to get information about Facebook users in a very accessible manner.  Of course, it is only your “public” information that is accessible without authorization.  But once you “allow” an application to access your information, you’re allowing access to EVERYTHING.

Generally, these items are publicly known:

{
   "id": "0123456789",
   "name": "Parag K Mital",
   "first_name": "Parag",
   "middle_name": "K",
   "last_name": "Mital",
   "locale": "en_US"
}

and also your Profile picture.

Check out a montage of the first 3600 Facebook user’s profile pictures, obtained just by using the public url: http://graph.facebook.com/USER_ID/picture


And an image of the average of all 3600 profile images:

Continue reading...

The post Facebook Graph API first appeared on http://archive.pkmital.com.

]]>
https://archive.pkmital.com/2010/07/19/facebook-graph-api/feed/ 0
Dynamic Scene Perception Eye-Movement Data Videos and Analysis https://archive.pkmital.com/2010/05/21/carpe-diem/ https://archive.pkmital.com/2010/05/21/carpe-diem/#respond Fri, 21 May 2010 11:56:22 +0000 http://pkmital.com/home/?p=293 Over the past 2 years, I have been working under the direction of Prof John M Henderson together with Dr Tim J Smith and Dr Robin Hill on the DIEM project (Dynamic Images and Eye-Movements). Our project has focused on investigating active visual cognition by eye-tracking numerous participants watching a wide-variety of short videos.

We are in the process of making all of our data freely available for research use. As well, we have also worked on tools for analyzing eye-movements during such dynamic scenes.

CARPE, or more bombastically known as Computational Algorithmic Representation and Processing of Eye-movements, allows one to begin visualizing eye-movement data together with the video data it was tracked with in a number of ways. It currently supports low-level feature visualizations, clustering of eye-movements, model selection, heat-map visualizations, blending, contour visualizations, peek-through visualizations, movie output, binocular data input, and more. The videos shown above on our Vimeo page were all created using this tool. Head over to Google code to check out the source code or download the binary. We are still in the process of stream-lining this process by creating manuals for new users and uploading more of the eye-tracking and video data so … Continue reading...

The post Dynamic Scene Perception Eye-Movement Data Videos and Analysis first appeared on http://archive.pkmital.com.

]]>
https://archive.pkmital.com/2010/05/21/carpe-diem/feed/ 0
DIEM Website https://archive.pkmital.com/2009/09/16/diem-website/ https://archive.pkmital.com/2009/09/16/diem-website/#comments Wed, 16 Sep 2009 16:26:00 +0000 http://pkmital.com/home/?p=50 The DIEM Project (Dynamic Images and Eye Movements) has a sleek new website which you can check out here: http://www.psy.ed.ac.uk/diemContinue reading...

The post DIEM Website first appeared on http://archive.pkmital.com.

]]>
https://archive.pkmital.com/2009/09/16/diem-website/feed/ 1
CLOSE-UP 2 https://archive.pkmital.com/2009/06/18/close-up-2/ https://archive.pkmital.com/2009/06/18/close-up-2/#respond Thu, 18 Jun 2009 01:49:00 +0000 http://pkmital.com/home/?p=42 An event organized by the new Center for Film, Performance, and Media Arts (CFPMA) was held today in Edinburgh University discussing recent topics in… film, performance, and media arts.

It was an interesting group of people that strangely somehow all had much in common. I had the fortune of presenting my research as it relates to DIEM in the place of Tim J. Smith.

The schedule:

Close-Up 2: Schedule for Wednesday 17th June 2009

10am coffee and tech checks (G.11, William Robertson Building)

10.30 Welcome, Annette Davison (Music, ACE and Director, Cfpma), Martine Beugnet (LLC, Convener of Film Studies)

Who is who, where is what? People and resources for which the Cfpma will provide a point of convergence.

11am Individual Presentations (MAX 10 mins each):

Andrew Lawrence (African Studies, SSPS) — Difficult satire under austerity: the films of Sissako and Amoussou

Martine Beugnet (Film, LLC) — “The Wounded Screen”

Richard Williams (Architecture, ACE) — “The Modernist City on Film”

Stephen Cairns (Architecture, ACE) — “Cultures of Legibility: Emergent Urban Landscapes in Southeast Asia”

Simon Frith and Annette Davison (Music, ACE) — “The Role of Cinemas in the History of Live Music”

Mary Fogarty (Music, ACE) — “The … Continue reading...

The post CLOSE-UP 2 first appeared on http://archive.pkmital.com.

]]>
https://archive.pkmital.com/2009/06/18/close-up-2/feed/ 0