Monthly Archives: October 2015

Virtual-Reality Headset, bonus

This brief post is to show you an old version of the modern HMD!

It seems like the cardboards and the oculus are just as new as this:



Virtual-Reality Headset, part 2

Hello! I will now continue with the showcase of my HMD prototypes, with the hope that this could help anyone out there, trying to make something similar. In the previous post I was very generic, here instead I will list some key steps that need to be considered while building a prototype of such devices:

  • Screen size: the difference here is between big and small.
    First you need to decide the ratio of our screen, it can be anything you can imagine/found: if your target is mostly movies and games, as they have a 16/9 aspect ratio (and we need one frame for each eye), you want a screen that has a (16*2) 32/9 ratio. Instead, if like me you want something for augmented or virtual reality, then you just want to have a reasonable surface to dedicate at each eye.
    Second you need to figure out the main dimension, let’s say the width. If as before you want to use for media-content, then as most of that material is supposed to have the user looking mainly at the center of the screen, you want that the distance from each eye-center is equal to half of the width of the entire screen. The images below explain it easily:
    suppose we have two monitor with 2560×1440 pixels of resolution. One (on the left) with a width of 14cm and one on the right with a width of 30cm. As shown in the images, if the user (like me) has a interpupillary distance of 70mm, the monitor on the left will be able to show the same content (the greenish box) using more pixels. On the right instead, to show the content big enough, they are overlapped causing a cut-off of the area.
    Some consideration: everything said here is intending lenses to focus to monitor have the optical axes perpendicular to it. This is required as making them not perpendicular (and thus being able to change the focus point on the monitor) will cause heavy-focus related problems. Solution? Use two monitor and put them at different angles, as StarVR did.
  • Optical system: as said before, if you are using a single straight display, you want to lenses optical axes to be perfectly perpendicular to the monitor. Also, you need the at the right distance from it: to know the exact distance for your lens, i suggest you to follow how explained in this video. Basically you make a straight laser go trough your lens and see where the lines intersects on the other side: be carefull, the laser must be parallel to the lens optical axis…
    Another important aspect is th
  • Building: now that monitor and lenses are chosen, to put everything together is really up to you with your personal skill. This process is both fun and annoying, only suggestion I can give you is to start with an easy shape and material, and gradually improve it.

Now that I told you a bit of my experience, I show you some of my prototypes:

IMG-20150407-WA0002 - CopyIMG_20150914_015231540_HDR

And my final result (waiting for 3D printed parts), which has:

  • 2560×1440@60hx display with micro-HDMI board
  • Creative Senz3D ToF sensor with a Wide-Angle lens for the RGB camera
  • 9-DOF IMU
  • a modified ColorCross holder for two 51mm aspheric lenses

IMG-20151010-WA0005 IMG-20151010-WA0003

If you also would like to see some of the software I develop for this device, and try it on yours, I will release it soon for everyone to download. Here some screenshot!

new_mode1  new_mode3new_mode2

And some young-user giving their best feedback at a local presentation!

IMG_20150925_204438890 IMG_20150925_204449171 IMG-20150925-WA0002

Virtual-Reality Headset, part 1

Hello! In this post I will share some test I have done to build my own virtual-reality headset. The amount of small-decision that must be made to go from a sketch to a working prototype are the ones that will make your build good or not: a lot depends on the materials you are using, and I believe 3D printing is seriously the best choice, but it has not been a possibility for me so far.

The first attempts that I have made where with cardboard, and some very-wide magnification lenses from some scrap hardware, and as the material is very easy to work with, it turn out pretty good and robust. With this cardboard frame, I used my phone as monitor and I was able to enjoy some games and videos, but the best experience was using it for movies while lying in bed! (always considering if you have the right-sized display on your phone for your eyes!)


As the need for a better support was needed, I went  with hard-plastic, but it turn out pretty heavy.It had a 7-inch display (in which I could crop the image via software, as it was used as a secondary screen attached to a PC) with a resolution of 1280*720 pixels and refresh rate not more than 60Hz. The lenses were a couple of the famous 51mm aspheric lens that I found while looking for suggestion by other people. I also calibrated them using a small chessboard, as described in my thesis (A. Memo, “Multi-Modal Head Mounted Vision System with Gesture Interface”, University of Padua, 2015) and the result is a pretty decent image quality, but chromatic-aberration also needs to be addressed.


End of part 1, soon the rest!

HandGenerator, a synthetic hand-pose dataset creation library

Hello, In this post I will give a brief presentation of the so
bothftware HandGenerator, that I created to generate the synthetic dataset used in my publication (A.
Memo, L. Minto and P. Zanuttigh, “Exploiting Silhouette Descriptors and Synthetic Data for Hand Gesture Recognition”, STAG: Smart Tools & Apps for Graphics, 2015).

To download and test the library, to see it’s capabilities, please follow this link to the LTTM Laboratory at the University of Padua, HandGenerator@LTTM.

The library enables the user to have an intuitive GUI to set up the hand pose, and then recursively generate a large dataset with its parameters. The possibility for customization allow for a wide field of usage, from color-based datasets to depth-based (as the one used in the paper above): this is done exploiting the possibility of using custom shaders for the OpenGL pipeline that render the 3D model, and in this case the results are very good as shown below.

g13s g6s g1s