Beamatron is an example of the trend Microsoft is exploring that blends the physical world with the virtual world. It’s augmented reality concept that combines a projector and Kinect sensor on a pan tilt moving head – of the kind you may find in a nightclub. The setup utilizes KinectFusion to build a 3D model of a space and enables projected graphics to react in physically appropriate ways. For example a virtual car can be driven around the floor of the room bumping into actual obstacles and running over real ramps.
Arturo Castro says:
This is a technical demo for face substitution technique. The application works in real time and it’s developed using the opensource framework for creative coding openFrameworks: openFrameworks.cc
Most of the “magic” happens thanks to Jason Saragih’s c++ library for face tracking web.mac.com/jsaragih/FaceTracker/FaceTracker.html. The face tracking library returns a mesh that matches the contour of the eyes, nose, mouth and other facial features.
That way the mesh obtained from a photo is matched to my own face in the video. Applying some color interpolation algorithms from Kevin Atkinson’s image clone code: methodart.blogspot.com/ gives it the blending effect that can be seen in the final footage.
I’m also using Kyle McDonald’s ofxFaceTracker addon for openframeworks github.com/kylemcdonald/ofxFaceTracker which wraps Jason’s library for easier use.
Kyle has uploaded another video giving a try at a different blending algorithm, which get rids of the artifacts and an even creepier look: vimeo.com/29348533
Thesis Prize Winner – Harvard Graduate School of Design 2011
by Keiichi Matsuda
The architecture of the contemporary city is no longer simply about the physical space of buildings and landscape, more and more it is about the synthetic spaces created by the digital information that we collect, consume and organise; an immersive interface may become as much part of the world we inhabit as the buildings around us.
Augmented Reality (AR) is an emerging technology defined by its ability to overlay physical space with information. It is part of a paradigm shift that succeeds Virtual Reality; instead of disembodied occupation of virtual worlds, the physical and virtual are seen together as a contiguous, layered and dynamic whole. It may lead to a world where media is indistinguishable from ‘reality’. The spatial organisation of data has important implications for architecture, as we re-evaluate the city as an immersive human-computer interface.
Jiim is a system developed at the Australian Centre for Visual technologies which allows the user to create 3D models while they capture video. This has a number of applications, but perhaps most interestingly means that a user might be able to perform 3D modelling within an Augmented Reality system.
This would mean that real and synthetic (augmented) geometry would be able to interact, without the need for pre-existing 3D models of the scene.
The video shows a few examples of this application.
A sketch of how we may some day be able to shop with the help of spatial computing. Spatial computing is the experience of digital 3d objects inhabiting real 3d space in a way that is interactive and intuitive.More at: http://www.phedhex.com/blog/?p=173
Pattie Maes & Pranav Mistry: Unveiling the “Sixth Sense,” game-changing wearable tech
This demo — from Pattie Maes’ lab at MIT, spearheaded by Pranav Mistry — was the buzz of TED. It’s a wearable device with a projector that paves the way for profound interaction with our environment. Imagine “Minority Report” and then some.
Pranav Mistry is the genius behind Sixth Sense, a wearable device that enables new interactions between the real world and the world of data.