This is how it started, with the most simplified flocking algorithm; had only attraction towards target: Here, flock is conducted with Leap motion, using hand tracking.
processing sketch : here :
However, after the user testing, I’ve decided to use the Kinect for motion tracing and conducting to utilize more of the body parts for more complex motions, instead of just one hand.
1. Testing a responsive processing sketch
Hoping to see how people respond to what they see. Before that, how noticeable it is. Whether they will see it or pass by. If they do, how interesting it is – will they stop, will they try to interact? How many times do they need to come by it to get intrigued and interact.
How long they interact? What is the interaction like? Will they come back or how they act next time if they are already familiar?
Testing was planned to be executed using one of ITP screens, to run the Processing sketch on the screen and attach leap motion next to it. Ideally, I would get the screen on the 1st floor of Tisch building, one that is inside but facing street. So the random passengers would be users. However, that was impossible to do, as the someone else’s project is running until May 1st.
After that, I got permission to use one of the screens on the floor, in the hallway, which was satisfactory enough. Again, I’ve encounter difficulties. I have spent more then 2 hours trying to get the sketch running, with the ending conclusion to ask for software updates on screen computers, and permission for that too. As that also will not be done in time, I used projector for testing. This, as a major change, has affected results of the test (as I knew it would), still I was hoping to get at least some feedback.
2. Interviewed persons info:
Random ITPers passing by, as they would every day. Testing day is Tuesday, when mostly 2nd years are on the floor, doing their thesis. Not as much students as during other weekdays. (so many factors)
– Most of them just passed by, with a glimpse but walked on.
– 20% actually took a look for few seconds to see what is going on. Here, not only on projection, but all around, projector, leap, cables. More likely they were interested in what is going on in general then on the screen. Of course as projector, cables and leap were improvised and so obvious.
– And those 10%, that were walking close by enough to trigger the sketch, and intrigued enough to try to interact, spent not more then 20-30 seconds. One reason is that projector was projecting into their back. (only one crunched to try to play a bit more). That again, lasted not more 30 seconds, putting hand right in front of the leap.
3. overall interpretation
Screen, screen, screen, it has to be on the screen. Even if it was short throw projector, it is different when facing a wall with something behind than the screen with something ‘in it’. *unless this is to be installed in subway, projecting on the other side’s platform, when we want people to look behind, onto another side.
Replace Leap with Kinect. Leap has to short range, so significant number of potential interactions is lost. Much more than I thought. Also when placing the sensor, leap or Kinect, should make sure it is well hidden or at least be placed as it is part of the screen. The aim to make people interact with the thing on the screen, not that obviously with sensor itself.
More important, Kinect will enable more interesting (possible) movements/actions, than just hand movements, which are not interesting in amazingly short time.
Also the sketch itself should be changed in a way to poses more different behaviors, so it can build more complex interactions. Counting on that more complex means more engaging.
The state of the sketch, when idle and when ‘is after someone’, should be more distinctive.