computer vision - http://archive.pkmital.com http://archive.pkmital.com computational audiovisual augmented reality research Sun, 08 Feb 2015 21:36:43 +0000 en-US hourly 1 https://wordpress.org/?v=6.9.4 Toolkit for Visualizing Eye-Movements and Processing Audio/Video http://archive.pkmital.com/2015/02/06/toolkit-for-visualizing-eye-movements-and-processing-audio-video/ http://archive.pkmital.com/2015/02/06/toolkit-for-visualizing-eye-movements-and-processing-audio-video/#comments Fri, 06 Feb 2015 23:53:18 +0000 http://pkmital.com/home/?p=1852 Screen Shot 2015-02-06 at 6.24.27 PM

Original video still without eye-movements and heatmap overlay copyright Dropping Knowledge Video Republic.

From 2008 – 2010, I worked on the Dynamic Images and Eye-Movements (D.I.E.M.) project, led by John Henderson, with Tim Smith and Robin Hill. We worked together to collect nearly 200 participants eye-movements on nearly 100 short films from 30 seconds to 5 minutes in length. The database is freely available and covers a wide range of film styles form advertisements, to movie and music trailers, to news clips. During my time on the project, I developed an open source toolkit, C.A.R.P.E. to complement D.I.E.M., or Computational Algorithmic Representation and Processing of Eye-movements (Tim’s idea!), for visualizing and processing the data we collected, and used it for writing up a journal paper describing a strong correlation between tightly clustered eye-movements and the motion in a scene. We also output visualizations of our entire corpus on our Vimeo channel. The project came to a halt and so did the visualization software. I’ve since picked up the ball and re-written it entirely from the ground up.

The image below shows how you can represent the movie, the motion in the scene of the movie (represented in … Continue reading...

The post Toolkit for Visualizing Eye-Movements and Processing Audio/Video first appeared on http://archive.pkmital.com.

]]>
http://archive.pkmital.com/2015/02/06/toolkit-for-visualizing-eye-movements-and-processing-audio-video/feed/ 4
Handwriting Recognition with LSTMs and ofxCaffe http://archive.pkmital.com/2015/02/06/handwriting-recognition-with-lstms-and-ofxcaffe/ http://archive.pkmital.com/2015/02/06/handwriting-recognition-with-lstms-and-ofxcaffe/#respond Fri, 06 Feb 2015 04:41:49 +0000 http://pkmital.com/home/?p=1836

Long Short Term Memory (LSTM) is a Recurrent Neural Network (RNN) architecture designed to better model temporal sequences (e.g. audio, sentences, video) and long range dependencies than conventional RNNs [1]. There is a lot of excitement in the machine learning communities with LSTMs (and Deep Minds’s counterpart, “Neural Turing Machines” [2], or Facebook’s, “Memory Networks” [3]) as they overcome a fundamental limitation to conventional RNNs and are able to achieve state-of-the-art benchmark performances on a number of tasks [4,5]:

  • Text-to-speech synthesis (Fan et al., Microsoft, Interspeech 2014)
  • Language identification (Gonzalez-Dominguez et al., Google, Interspeech 2014)
  • Large vocabulary speech recognition (Sak et al., Google, Interspeech 2014)
  • Prosody contour prediction (Fernandez et al., IBM, Interspeech 2014)
  • Medium vocabulary speech recognition (Geiger et al., Interspeech 2014)
  • English to French translation (Sutskever et al., Google, NIPS 2014)
  • Audio onset detection (Marchi et al., ICASSP 2014)
  • Social signal classification (Brueckner & Schulter, ICASSP 2014)
  • Arabic handwriting recognition (Bluche et al., DAS 2014)
  • TIMIT phoneme recognition (Graves et al., ICASSP 2013)
  • Optical character recognition (Breuel et al., ICDAR 2013)
  • Image caption generation (Vinyals et al., Google, 2014)
  • Video to textual description (Donahue et al., 2014)

The current dynamic state … Continue reading...

The post Handwriting Recognition with LSTMs and ofxCaffe first appeared on http://archive.pkmital.com.

]]>
http://archive.pkmital.com/2015/02/06/handwriting-recognition-with-lstms-and-ofxcaffe/feed/ 0
Real-Time Object Recognition with ofxCaffe http://archive.pkmital.com/2015/01/04/real-time-object-recognition-with-ofxcaffe/ http://archive.pkmital.com/2015/01/04/real-time-object-recognition-with-ofxcaffe/#comments Sun, 04 Jan 2015 03:53:48 +0000 http://pkmital.com/home/?p=1764 Screen Shot 2015-01-03 at 12.57.23 PM

I’ve spent a little time with Caffe over the holiday break to try and understand how it might work in the context of real-time visualization/object recognition in more natural scenes/videos. Right now, I’ve implemented the following Deep Convolution Networks using the 1280×720 resolution webcamera on my 2014 Macbook Pro:

The above image depicts the output from an 8×8 grid detection showing brighter regions as higher probabilities of the class “snorkel” (automatically selected by the network from 1000 possible classes as the highest probability).

So far I have spent some time understanding how Caffe keeps each layer’s data during a forward/backward pass, and how the deeper layers could be “visualized” in a … Continue reading...

The post Real-Time Object Recognition with ofxCaffe first appeared on http://archive.pkmital.com.

]]>
http://archive.pkmital.com/2015/01/04/real-time-object-recognition-with-ofxcaffe/feed/ 4
YouTube’s “Copyright School” Smash Up http://archive.pkmital.com/2012/11/09/youtubes-copyright-school/ http://archive.pkmital.com/2012/11/09/youtubes-copyright-school/#respond Fri, 09 Nov 2012 15:19:51 +0000 http://pkmital.com/home/?p=1303 Ever wonder what happens when you’ve been accused of violating copyright multiple times on YouTube? First, you get a redirect to YouTube’s “Copyright School” whenever you visit YouTube, forcing you to watch a cartoon of Happy Tree Friends where the main character is dressed as an actual pirate:

Second, I’m guessing, your account will be banned. Third, you cry and wonder why you ever violated copyright in the first place.

In my case, I’ve disputed every one of the 4 copyright violation notices that I’ve received under grounds of Fair Use and Fair Dealing. Here’s what happens when you file a dispute using YouTube’s online form (click for high-res):






3 of the 4 have been dropped after I’ve filed disputes, though I’m still waiting to hear about the response to the above dispute. Read the dispute letter to Sony ATV and UPMG Publishers in full here.

The picture above shows a few stills from what my Smash Ups look like. The process described in greater detail on createdigitalmotion.com is part of my ongoing research into how existing content can be transformed into artistic styles reminiscent of analytic cubist, figurative, and futurist paintings. The process to create the videos … Continue reading...

The post YouTube’s “Copyright School” Smash Up first appeared on http://archive.pkmital.com.

]]>
http://archive.pkmital.com/2012/11/09/youtubes-copyright-school/feed/ 0
An open letter to Sony ATV and UMPG http://archive.pkmital.com/2012/11/09/an-open-letter-to-sony-atv-and-umpg/ http://archive.pkmital.com/2012/11/09/an-open-letter-to-sony-atv-and-umpg/#respond Fri, 09 Nov 2012 00:55:57 +0000 http://pkmital.com/home/?p=1291

Dear Sony ATV Publishing, UMPG Publishing, and other concerned parties,

I ask you to please withdraw your copyright violation notice on my video, “PSY – GANGNAM STYLE (?????) M/V (YouTube SmashUp)” as I believe my use of any copyrighted material is protected under Fair Use or Fair Dealing. This video was created by an automated process as part of an art project developed during my PhD at Goldsmiths, University of London: http://archive.pkmital.com/projects/visual-smash-up/ and http://archive.pkmital.com/projects/youtube-smash-up/

The process which creates the audio and video is entirely automated meaning the accused video is created by an algorithm. This algorithm begins by first creating a large database of tiny fragments of audio and video (less than 1 second of audio per fragment) using 9 videos from YouTube’s top 10 list. From this database, the tiny fragments of video and audio are stored as unrelated pieces of information and described only by a short series of 10-15 numbers. These numbers represent low-level features describing the texture and shape of the fragment of audio or video. These tiny fragments are then matched to the tiny fragments of audio and video detected within the target for resynthesis, in this case the number one YouTube video … Continue reading...

The post An open letter to Sony ATV and UMPG first appeared on http://archive.pkmital.com.

]]>
http://archive.pkmital.com/2012/11/09/an-open-letter-to-sony-atv-and-umpg/feed/ 0
Copyright Violation Notice from “Rightster” http://archive.pkmital.com/2012/10/16/copyright-violation-notice-from-rightster/ http://archive.pkmital.com/2012/10/16/copyright-violation-notice-from-rightster/#comments Tue, 16 Oct 2012 11:49:38 +0000 http://pkmital.com/home/?p=1246 I’ve been working on an art project which takes the top 10 videos in YouTube and tries to resynthesize the #1 video in YouTube using the remaining 9 videos. The computational model is based on low-level human perception and uses only very abstract features such as edges, textures, and loudness. I’ve created a new synthesis each week using the top 10 of the week in the hopes that, one day, I will be able to resynthesize my own video in the top 10. It is a viral algorithm essentially but it is not proven if it will succeed or not.

The database of content used in the recreation of the above video comes from the following videos:
#2 News Anchor FAIL Compilation 2012 || PC
#3 Flo Rida – Whistle [Official Video]
#4 Carly Rae Jepsen – Call Me Maybe
#5 Jennifer Lopez – Goin’ In ft. Flo Rida
#6 Taylor Swift – We Are Never Ever Getting Back Together
#7 will.i.am – This Is Love ft. Eva Simons
#8 Call Me Maybe – Carly Rae Jepsen (Chatroulette Version)
#9 Justin Bieber – As Long As You Love Me ft. Big Sean
#10 Rihanna – Where Have You Been

It … Continue reading...

The post Copyright Violation Notice from “Rightster” first appeared on http://archive.pkmital.com.

]]>
http://archive.pkmital.com/2012/10/16/copyright-violation-notice-from-rightster/feed/ 2
Concatenative Video Synthesis (or Video Mosaicing) http://archive.pkmital.com/2011/10/08/concatenative-video-synthesis-or-video-mosaicing/ http://archive.pkmital.com/2011/10/08/concatenative-video-synthesis-or-video-mosaicing/#comments Sat, 08 Oct 2011 10:08:47 +0000 http://pkmital.com/home/?p=830 prototype

Working closely with my adviser Mick Grierson, I have developed a way to resynthesize existing videos using material from another set of videos. This process starts by learning a database of objects that appear in the set of videos to synthesize from. The target video to resynthesize is then broken into objects in a similar manner, but also matched to objects in the database. What you get is a resynthesis of the video that appears as beautiful disorder. Here are two examples, the first using Family Guy to resynthesize The Simpsons. And the second using Jan Svankmajer’s Food to resynthesize Jan Svankmajer’s Dimensions of Dialogue.

Continue reading...

The post Concatenative Video Synthesis (or Video Mosaicing) first appeared on http://archive.pkmital.com.

]]>
http://archive.pkmital.com/2011/10/08/concatenative-video-synthesis-or-video-mosaicing/feed/ 2
Facial Appearance Modeling/Tracking http://archive.pkmital.com/2011/05/26/facial-appearance-modelingtracking/ http://archive.pkmital.com/2011/05/26/facial-appearance-modelingtracking/#comments Thu, 26 May 2011 22:19:10 +0000 http://pkmital.com/home/?p=602
I’ve been working on developing a method for automatic head-pose tracking, and along the way have come to model facial appearances. I start by initializing a facial bounding box using the Viola-Jones detector, a well known and robust detector used for training objects. This allows me to centralize the face. Once I know where the 2D plane of the face is in an image, I can register an Active Shape Model like so:

After multiple views of the possible appearance variations of my face, including slight rotations, I construct an appearance model.

The idea I am working with is using the first components of variations of this appearance model for determining pose. Here I show the first two basis vectors and the images they reconstruct:

As you may notice, these two basis vectors very neatly encode rotation. By looking at the eigenvalues of the model, you can also interpret pose.… Continue reading...

The post Facial Appearance Modeling/Tracking first appeared on http://archive.pkmital.com.

]]>
http://archive.pkmital.com/2011/05/26/facial-appearance-modelingtracking/feed/ 2
Tim J Smith guest blogs for David Bordwell http://archive.pkmital.com/2011/02/20/tim-j-smith-guest-blogs-for-david-bordwell/ http://archive.pkmital.com/2011/02/20/tim-j-smith-guest-blogs-for-david-bordwell/#respond Sun, 20 Feb 2011 03:43:36 +0000 http://pkmital.com/home/?p=540 Tim J Smith, expert in scene perception and film cognition, and of The DIEM project [1] recently starred as a guest blogger for David Bordwell, a leading film theorist with an impressive list of books and publications widely used in film cognition/film art research/studies [2]. In his article featured on David’s site, Tim expands on his research on film cognition including continuity editing [3], attentional synchrony [4], and the project we worked on in 2008-2010 as part of The DIEM Project. Since Tim’s feature on David Bordwell’s blog, The DIEM Project saw a surge of publicity and our vimeo video loads going higher than 200,000 in a single day and features on dvice, slashfilm, gizmodo, Rogert Ebert’s facebook/twitter, and the front page of imbd.com.

Not to mention, our tools and visualizations are finally reaching an audience with interests in film, photography, and cognition. If you haven’t yet seen some of our videos, please head on over to our vimeo page, where you can see a range of videos embedded with eye-tracking of participants and many different visualizations of models of eye-movements using machine learning, or start by reading Tim’s post on Continue reading...

The post Tim J Smith guest blogs for David Bordwell first appeared on http://archive.pkmital.com.

]]>
http://archive.pkmital.com/2011/02/20/tim-j-smith-guest-blogs-for-david-bordwell/feed/ 0
Responsive Ecologies Documentation http://archive.pkmital.com/2011/02/09/responsive-ecologies-documentation/ http://archive.pkmital.com/2011/02/09/responsive-ecologies-documentation/#respond Wed, 09 Feb 2011 14:34:40 +0000 http://pkmital.com/home/?p=479

As part of a system of numerous dynamic connections and networks, we are reactive and deterministic to a complex system of cause and effect. The consequence of our actions upon our selves, the society we live in and the broader natural world is conditioned by how we perceive our involvement. The awareness of how we have impacted on a situation is often realised and processed subconsciously, the extent and scope of these actions can be far beyond our knowledge, our consideration, and importantly beyond our sensory reception. With this in mind, how can we associate our actions, many of which may be overlooked as customary, with for instance, the honey bee depopulation syndrome or the declining numbers of Siberian Tigers.

Responsive Ecologies is part of an ongoing collaboration with ZSL London Zoo and Musion Academy. Collectively we have been exploring innovative means of public engagement, to generate an awareness and understanding of nature and the effects of climate change. All of the contained footage has come from filming sessions within the Zoological Society; this coincidentally has raised some interesting questions on the spectacle of captivity, a issue which we have tried to reflect upon in the construction and presentation of … Continue reading...

The post Responsive Ecologies Documentation first appeared on http://archive.pkmital.com.

]]>
http://archive.pkmital.com/2011/02/09/responsive-ecologies-documentation/feed/ 0