User:Mihail Bakalov/Trimester 2/lENS BASED PROTOTYPING/KINECT PROJECT
< User:Mihail Bakalov | Trimester 2/lENS BASED PROTOTYPING
Revision as of 15:36, 7 September 2014 by Mihail Bakalov (talk | contribs)
MIRROR OF VIRTUAL TRUTH
An installation reflecting on internet trackers in a closed environment. It consists of a projection of virtual reality which shows how data about ones doings online is being tracked A display acting as a mirror shows visual presentation of how a persons data is spread around the web, displaying a conversation between the trackers with the data being pulled about a persons location.
CODE HERE:
import SimpleOpenNI.*;
import java.awt.Color;
import java.util.Iterator;
import gifAnimation.*;
SingleUserKinect kinect;
PVector userPosRealWorld = new PVector(); // 3d user position
PVector userPosProjective = new PVector(); // user screen position
float comX; // Center of Mass X
float maxX = 500; // 1/2 meter (0 is in the middle)
float minX = -500; // -1/2 meter (0 is in the middle)
/* STUDENT: Adjust this value to the right number of images you have */
int numImages = 17; // total number of images
PImage[] images = new PImage[numImages]; // the images will be stored in this list
PImage agent2;
SimpleOpenNI context;
PImage[] animation;
PImage[] animation2;
PImage[] animation3;
PVector prevUserPosRealWorld = new PVector();
float moveDist = 4;
PFont font;
float[] kordinati;
void setup() {
// size of the window
//size(400,400);// use size "(displayWidth, displayHeight)" for fullscreen
size(displayWidth, displayHeight, P3D);
// set frame background to black (this is the background you'll see in Present Mode)
frame.setBackground(new Color(0, 0, 0));
// user SingleUserKinect for tracking.
kinect = new SingleUserKinect(this);
kinect.context.enableRGB();
agent2 = loadImage("data/agent_half.gif");
animation = Gif.getPImages(this, "data/binaryflow2.gif");
animation2 = Gif.getPImages(this, "data/agent_half2.gif");
animation3 = Gif.getPImages(this, "data/agent_half3.gif");
font = createFont("Arial", 14);
textFont(font);
}
void draw() {
// update user position
kinect.update();
background(255);
// set center of mass to the maximum value
float comX = maxX;
// if we are tracking a user, get the position of his Center of Mass
if (kinect.trackedUserId != 0) {
kinect.getCoM(userPosRealWorld);
comX = userPosRealWorld.x;
if (comX > maxX) {
comX = maxX;
}
else if (comX < minX) {
comX = minX;
}
}
float maxRange = maxX - minX; // total range
float x = map(comX, minX, maxX, 0, maxRange); // map center of mass to a positive range
float d = dist(userPosRealWorld.x, userPosRealWorld.y ,userPosRealWorld.z, prevUserPosRealWorld.x, prevUserPosRealWorld.y, prevUserPosRealWorld.z);
if (userPosRealWorld == new PVector() || prevUserPosRealWorld == new PVector() ) {
d = 0.0;
}
kinect.context.convertRealWorldToProjective(userPosRealWorld, userPosProjective);
float screenX = map(userPosProjective.x, 0, 640, 0, width);
float screenY = map(userPosProjective.y, 0, 640, 0, width);
float screenZ = map(userPosProjective.z, 0, 640, 0, width);
//println(userPosRealWorld.x + " --- " + userPosProjective.x + " --- " + screenX);
//kordinati[] = float[screenX] + float[screenY] + float[screenZ];
// draw the image
image(kinect.context.rgbImage(), 0, 0, width, height);
//image(kinect.context.rgbImage(), kinect.context.depthWidth(), 0, displayWidth, displayHeight);
// image(animation2[(int) (animation2.length / (float) (width) * screenX)], (screenX) - 150, screenY - 500, 114, 155);
// image(animation3[(int) (animation3.length / (float) (width) * screenX)], (screenX) + 100, screenY - 500, 114, 155);
image(animation2[(int) (animation2.length / (float) (width) * screenX)], (screenX) - 150, screenY - 500, 200 - screenZ/60, 240 - screenZ/60);
image(animation3[(int) (animation3.length / (float) (width) * screenX)], (screenX) + 100, screenY - 500, 200 - screenZ/60, 240 - screenZ/60);
println("distance moved " + d);
if (d > moveDist) {
text( userPosRealWorld.x, (screenX) - 10, screenY - 450);
text( userPosRealWorld.y, (screenX) - 10, screenY - 440);
text( userPosRealWorld.z, (screenX) - 10, screenY - 430);
//image(animation[(int) (animation.length / (float) (width) * screenX)], (screenX) - 130, screenY - 400, 74, 42);
//image(animation[(int) (animation.length / (float) (width) * screenX)], (screenX) + 120, screenY - 400, 74, 42);
image(animation[(int) (animation.length / (float) (width) * screenX)], (screenX) + 120, screenY - 400, 160 - screenZ/60, 130 - screenZ/60);
image(animation[(int) (animation.length / (float) (width) * screenX)], (screenX) - 130, screenY - 400, 160 - screenZ/60, 130 - screenZ/60);
}
prevUserPosRealWorld = new PVector(userPosRealWorld.x, userPosRealWorld.y, userPosRealWorld.z);
} // end draw
// -----------------------------------------------------------------
// SimpleOpenNI user events
// -----------------------------------------------------------------
// onNewUser is triggered when the kinect registers a new user
void onNewUser(SimpleOpenNI curContext, int userId)
{
// let our SingleUserKinect Class take care of this
kinect.registerNewUser(curContext, userId);
}
// onLostUser is triggered when the kinect deregisters a user
void onLostUser(SimpleOpenNI curContext, int userId)
{
// let our SingleUserKinect Class take care of this
kinect.deRegisterUser(curContext, userId);
}