2013 Israel Computer Vision Day
Monday, January 20, 2014
Vision Day Schedule
Time |
Speaker
and Collaborators |
Affiliation |
Title |
08:50-09:20 |
Gathering |
||
09:20-09:40 |
Yohay
Swirski, Yoav
Schechner |
Technion |
|
9:45-10:05 |
Yonatan
Aflalo Ron
Kimmel |
Technion |
|
10:10-10:30 |
Tal
Hassner |
Open
U. |
|
10:35-10:55 |
Yehonatan
Goldman Ilan
Shimshoni Ehud
Rivlin |
Technion |
|
11:00-11:30 |
Coffee Break |
||
11:30-11:50 |
Simon
Korman Daniel
Reichman Gilad Tsur Shai
Avidan |
TAU Weizmann |
|
11:55-12:15 |
Elhanan
Elboher Michael Werman Yacov Hel-Or |
HUJI IDC |
The Generalized Laplacian
Distance and its Applications for Visual Matching |
12:20-12:40 |
Alexandra
Gilinsky Lihi
Zelnik-Manor |
Technion |
SIFTpack:
a Compact Representation for Efficient SIFT matching |
12:45-13:05 |
Ofir Pele Ben Taskar |
Ariel |
|
13:05-14:10 |
Lunch
|
||
14:10-14:20 |
"Intermezzo" |
||
14:20-14:40 |
Raja
Giryes Michael
Elad |
Technion |
|
14:45-15:05 |
Tomer
Michaeli Michal
Irani |
Weizmann |
|
15:10-15:30 |
Dror
Sholomon Omid
David Nathan
S. Netanyahu |
BIU |
A Genetic Algorithm-Based Solver for Very
Large Jigsaw Puzzles |
15:35-15:55 |
Dan
Levi Shai
Silberstein Aharon Bar-Hillel |
GM |
|
16:00-16:20 |
Coffee
Break |
||
16:20-16:40 |
Shahar
Gino Orly
Goitein Eli
Konen Hedva
Spitzer |
TAU |
Video Stabilization and Region-Of-Interest
tracking in non-rigid object Cardiac MRI |
16:45-17:05 |
Yair Hanani Lior Wolf Tal Hassner |
TAU |
|
17:10-17:30 |
Shimrit
Haber Yossi
Keller |
BIU |
A probabilistic graph-based framework for multi-cue visual
tracking |
The
Computer Vision Day is sponsored by:
Abstracts |
|
Yohay Swirski and Yoav Schechner -
Technion |
|
Yonatan Aflalo and
Ron Kimmel - Technion |
|
Tal Hassner – Open University |
|
Yehonatan
Goldman (Technion), Ilan Shimshoni (Haifa) and Ehud Rivlin (Technion) |
|
Simon Korman
(TAU), Daniel Reichman (Weizmann), Gilad Tsur (TAU) and Shai Avidan (TAU) |
|
The Generalized Laplacian Distance and its
Applications for Visual Matching |
Elhanan Elboher (HUJI), Michael Werman (HUJI),
Yacov Hel-Or (IDC) |
|
SIFTpack: a Compact Representation for Efficient
SIFT Matching |
Alexandra Gilinsky and Lihi Zelnik-Manor - Technion |
Computing distances between large
sets of SIFT descriptors is a basic step in numerous algorithms in computer
vision. When the number of descriptors is large, as is often the case,
computing these distances can be extremely time consuming. In this paper we
propose the SIFTpack: a compact way of storing SIFT descriptors, which
enables significantly faster calculations between sets of SIFTs than the
current solutions. SIFTpack can be used to represent SIFTs densely extracted
from a single image or sparsely from multiple different images. We show that
the SIFTpack representation saves both storage space and run time, for both
finding nearest neighbors and for computing all distances between all
descriptors. The usefulness of SIFTpack is also demonstrated as an
alternative implementation for K-means dictionaries of visual words.
|
Ofir Pele and Ben Taskar – Ariel |
We present a new histogram distance,
the Tangent Earth Mover’s Distance (TEMD). The TEMD is a generalization of
the Earth Mover’s Distance (EMD) that is invariant to some global
transformations. Thus, like the EMD it is robust to local deformations.
Additionally, it is robust to global transformations such as global
translations and rotations of the whole image. The TEMD is formulated as a
linear program which allows efficient computation. Additionally, previous
works about the efficient computation of the EMD that reduced the number of
variables in the EMD linear program can be used to accelerate also the TEMD
computation. We present results for image retrieval using the Scale Invariant
Feature Transform (SIFT) and color image descriptors. We show that the new
TEMD outperforms state of the art distances. |
Raja Giryes and
Michael Elad - Technion |
|
Tomer
Michaeli and Michal Irani - Weizmann |
|
A Genetic Algorithm-Based Solver for Very
Large Jigsaw Puzzles |
Dror Sholomon, Omid David, and Nathan S. Netanyahu – Bar Ilan |
|
Dan Levi, Shai Silberstein, Aharon
Bar-Hillel – General Motors |
|
Video
Stabilization and Region-Of-Interest tracking in non-rigid object Cardiac MRI |
Shahar Gino, Orly Goitein, Eli Konen and
Hedva Spitzer – TAU |
Several Cardiac-MRI (CMRI) sequences,
such as the perfusion series, are influenced by diaphragm and cardiac motion
throughout the respiratory and cardiac cycles [1]. Perfusion is a sequence of
Cardiac-MRI (CMRI), which is a non-invasive tool to assess myocardial
abnormalities, such ischemia. Myocardial first-pass perfusion schemes track
the contrast agent changes (Gadolinium) passage through the heart. This
perfusion imaging is used as a key component of most clinical cardiac MRI
exams. Stabilizing these videos expected to allow a significant improvement
in medical diagnosis. Video-stabilization and ROI-tracking are well-known
problems in computer-vision, with many practical applications [2]-[3].
However these two problems become even more challenging for medical gray
videos, in which separating ROI from its background at varying texture
conditions makes it hard. We suggest a novel algorithm for CMRI tracking and
stabilization, inspired by human visual system (VS) mechanisms. It combines
information from both edge and region pathways and adaptively weights them
according to ROI state. The algorithm applies cortical receptive fields for
the contour (edge) detection and contour completion of VS mechanism for
region base pathway. The ROI motion is then estimated by common
linear-approximation for stabilization. The Video-stabilization is obtained
by solving the ROI-tracking problem, and keeping its initial position fixed.
The proposed algorithm was tested on several CMRI videos and appears to
achieve promising results. It is autonomous, self-adapting and requires no
user-interference. It is robust to image type and highly-sensitive to objects
motion. Moreover, it handles occlusions and deformations and runs in a
reasonable complexity. Finally we suggest a method for measuring a given
video ROI-stability, which has been used for estimating our
quality-of-results (QoR). We are using both objective and clinical approaches
for estimating our results. The objective approach is based on
Inter-Frame-Similarity (ITF) and Structural Similarity (SSIM) metrics. The
clinical approach is based on statistical experiment done with radiologists.
We perform cooperation comparison according unique measures which compare our
video results with video input and to state-of-art competitors’ algorithms.
All of the video results are being ranked by the radiologists with a 1-5
scale. Preliminary results are quite encouraging in the sense of
object-tracking and video-stabilization. It appears to be successful for
track moving and deforming objects with high-sensitivity, which allows a
promising video-stabilization. Stabilizing perfusion CMRI slice by heart
tracking seems well for long burst of frames. This should allow better
medical diagnosis. |
Yair Hanani, Lior Wolf and Tal Hassner - TAU |
|
A probabilistic graph-based framework for
multi-cue visual tracking |
Shimrit Haber and Yossi Keller –
Bar-Ilan University |
Object tracking is a fundamental task
in computer vision. Varying tracking scenarios require the use of multiiple
cues such as color, texture, motion detection, template matching, object
detection and Kalman filtering, to name a few. In this talk with discuss
recent results on multi-cue tracking, that is formulated as a probabilistic
inference problem. An image (video frame) is represented by a set of image
patches denoted as superpixels, and the tracking is formulated as the
classification of these superpixels to either foreground/background. The
inference is is computed by representing the superpixels as graph nodes that
are matched to a binary (foreground/background) state graph. We derive a
computationally efficient inference scheme based on spectral graph matching.
This formulation allows to adaptively utilize multiple cues simultaneously,
and is exemplified by applying it to surveillance video segments.. |