Jean-Baptiste Boin

Ph.D. candidate in Electrical Engineering at Stanford University


Current work

Art++

Art++ is an interactive museum guide using augmented reality on mobile devices. The augmented reality component uses state-of-the-art image retrieval techniques combined with real-time tracking and rendering. The project led to an exhibition at the Cantor Arts Center over the summer 2016.

Art++ was funded by a Magic Grant from the Brown Institute for Media Innovation.

This work got media coverage in the New York Times [1] [2], the Stanford Daily [3] [4], the Palo Alto Weekly [5], The Creators Project [6] and the San Francisco Chronicle [7].

Summer 2014 - Summer 2016.

Art++ is an interactive museum guide using augmented reality on mobile devices. The augmented reality component uses state-of-the-art image retrieval techniques combined with real-time tracking and rendering. The project led to an exhibition at the Cantor Arts Center over the summer 2016.

Art++ was funded by a Magic Grant from the Brown Institute for Media Innovation.

This work got media coverage in the New York Times [1] [2], the Stanford Daily [3] [4], the Palo Alto Weekly [5], The Creators Project [6] and the San Francisco Chronicle [7].

Summer 2014 - Summer 2016.

Effective Fisher Vector Aggregation for 3D Object Retrieval

Research project. Collaboration with André Araujo, Lamberto Ballan and Bernd Girod.

Winter/Spring 2016.

Depth Augmented Stereo Panorama

Research project. Collaboration with Jayant Thatte, Haricharan Lakshman, Gordon Wetzstein and Bernd Girod.

This project led to two publications. The conference paper "Depth Augmented Stereo Panorama For Cinematic Virtual Reality With Focus Cues" received the Conference Best Paper Award for Industry at ICIP 2016.

Summer - Fall 2015.

Research project. Collaboration with Jayant Thatte, Haricharan Lakshman, Gordon Wetzstein and Bernd Girod.

This project led to two publications. The conference paper "Depth Augmented Stereo Panorama For Cinematic Virtual Reality With Focus Cues" received the Conference Best Paper Award for Industry at ICIP 2016.

Summer - Fall 2015.

Previous projects

AI Agent for Light Cycle Racing
AI Agent for Light Cycle Racing in Tron

Course project for CS 221 (Artificial Intelligence: Principles and Techniques).

Fall 2015.

Dissection of a CNN
Tiny ImageNet Challenge - Dissection of a convolutional neural network

Course project for CS 231N (Convolutional Neural Networks for Visual Recognition).

Spring 2015.

Feature Tracking on Android
Improving Feature Tracking Using Motion Sensors on Android

Course project for CS 231M (Mobile Computer Vision).

Spring 2014.

Multi-resolution image recognition
Multi-resolution image recognition

Research project. Collaboration with Roland Angst, David Chen and Bernd Girod.

Summer/Fall 2014.

Lost Flock
Lost Flock

Android game designed as a course project for CS 248 (Interactive Computer Graphics). Collaboration with CJ Cullen, Ranjay Krishna and Jeff Stone.

Spring 2014.

Plane Rectification on Android
Plane Rectification in Real Time on an Android Device

Research project. Also presented as a course project for CS 231A (Computer Vision: From 3D Reconstruction to Recognition).

Fall 2013/Winter 2014.

Polyphonic Piano Transcription
Polyphonic Piano Transcription

Course project for CS 229 (Machine Learning). Collaboration with Mohammad Sadegh Ebrahimi.

Fall 2013.

Blood Vessel Segmentation
Blood Vessel Segmentation in Retinal Fundus Images

Course project for EE 368 (Digital Image Processing). Collaboration with Karianne Bergen and Sahinaz Safari Sanjani.

Spring 2013.

Camera Forensics
Camera Forensics

Course project for PSYCH 221 (Applied Vision and Image Systems). Collaboration with Tiffany Jou.

Winter 2013.

Personal projects

Wobble
Wobble

Parallax is a strong cue for depth perception. This project was inspired by gifs such as these ones, where a strong sense of depth is given by alternating between nearby viewpoints. In this demo, I wanted to explore this effect interactively in a virtual 3D world. The focus is automatically locked to the object at the center of the screen, and the camera frequency can be manually adjusted.

This demo was built using Unity.

3D From Head Tracking
3D From Head Tracking

This project was inspired by this video from 2007, which uses head tracking to increase depth perception. Here, we detect the position of the head relative to the screen by using the webcam, assumed to be located above the screen. We then use the clmtrackr library to extract the position of the eyes in real time. This allows us to render the scene as it would be seen from these eye positions if it was in the real 3D world.

This demo was built using three.js and was tested on Chrome and Firefox.

Teaching

EE 368: Digital Image Processing
EE 368: Digital Image Processing

Teaching assistant (head teaching assistant in Fall 2016).

Spring 2015, Fall 2016.

EE 261: The Fourier Transform
EE 261: The Fourier Transform and its Applications

Head teaching assistant.

Summer 2013.

Languages