15th Israel Machine Vision Conference (IMVC) 2026

Monday|April 27, 2026

Pavilion 10, EXPO Tel Aviv

Agenda

IMVC 2026 will feature presentations by leading researchers in AI, with a focus on image and video processing, computer vision, machine learning, and deep learning. Attendees can expect visionary insights and cutting-edge developments from both academia and industry, showcasing the latest trends in artificial intelligence and its applications.

 

Exhibition

IMVC is the premier platform for companies shaping the future of AI. Discover the latest advancements in machine vision and machine learning, and connect with experts, entrepreneurs, developers, and engineers. Join us in Tel Aviv to forge collaborations, explore ongoing trends, and witness new applications in the field.

Topics

Core Computer Vision & AI Technologies | Advanced AI Methodologies | Emerging Technologies & Applications | Specialized Application Domains | Technical Foundations & Optimization | Cutting-Edge Research Areas

...and many more

 

Keynote Speaker

Eyal Enav

Nvidia

Bio:

Ayellet Tal is a professor and the Alfred and Marion Bär Chair in Engineering at the Technion's Department of Electrical and Computer Engineering. She holds a Ph.D. in Computer Science from Princeton University and a B.Sc degree (Summa cum Laude) in Mathematics and Computer Science from Tel Aviv University. Among Prof. Tal’s accomplishments are the Rechler Prize for Excellence in Research, the Henry Taub Prize for Academic Excellence, and the Milton and Lillian Edwards Academic Lectureship. Prof. Tal has chaired several conferences on computer graphics, shape modeling, and computer vision, including the upcoming ICCV.

Title:

Build Vision AI Agents With NVIDIA Cosmos Reason VLM and Video Analytics Blueprint

Abstract:

A point cloud, which is a set of 3D positions, is a simple, efficient, and versatile representation of 3D data. Given a point cloud and a viewpoint, which points are visible from that viewpoint? Since points themselves do not occlude one another, the real question becomes: which points would be visible if the surface they were sampled from were known? In this talk we will explore why point visibility is important, not only in computer vision but also beyond, how it can be determined, and in particular, how it can be addressed within optimization or deep learning frameworks.

Iris Barshack

ProfessorSheba Medical Center & Ariel University

Bio:

Title:

Abstract:

Idan Bassouk

Aidoc

Bio:

Title:

Abstract:

Ilan Tsarfaty

ProfessorTel Aviv University

Bio:

Title:

Abstract:

Tal Drori

IBM Research

Bio:

Title:

Abstract:

Speakers

Ofir Lindenbaum

Assistant professor Bar Ilan University

Bio:

Ofir Lindenbaum is a Senior Lecturer (Assistant Professor) in the Faculty of Engineering at Bar-Ilan University. He completed a postdoctoral fellowship at Yale University in the Applied Mathematics Program with Prof. Ronald Coifman and Prof. Yuval Kluger, and earned his Ph.D. in Electrical Engineering from Tel Aviv University under Prof. Arie Yeredor and Prof. Amir Averbuch. His research develops interpretable, efficient machine learning methods for scientific discovery, focusing on high-dimensional tabular data, multimodal learning, sparsification, optimization, and representation learning, aiming to design principled, reliable, and data-efficient algorithms for real-world scientific data.  

Title:

COPER: Correlation-based Permutations for Multi-View Clustering

Abstract:

Combining data from multiple sources often leads to better insights, yet many existing multi-view clustering methods are tailored to specific domains or require complex, multi-stage pipelines. We present a practical end-to-end deep learning framework that works across diverse data types, including images and tabular data. Our approach learns unified representations that capture shared structure across sources and enables consistent grouping without manual labels. The method is scalable, robust to noise, and supported by both theoretical insights and extensive experiments on ten benchmark datasets, demonstrating strong and reliable performance across varied real-world settings.

Michal Holtzman Gazit

AI research scientistEarth Dynamics AI

Bio:

Michal Holtzman Gazit is a Computer Vision and AI Researcher at Earth Dynamics AI with over 25 years of expertise in image processing, computer vision and deep learning. Her career evolved from a foundation in medical imaging to sophisticated 2D and 3D structural analysis. She holds a BSc and MSc in electrical engineering, a PhD in computer science Technion and performed post-doctoral research in inverse problems at the University of British Columbia. Michal specializes in leading the transition of advanced research to production-ready systems. Currently, she develops Geoscience Foundation Models for mineral exploration, utilizing AI to decode Earth’s 3D structures and revolutionize resource discovery

Title:

Multi-Modal Geologic Intelligence: 3D Inversion and Map Synthesis via Generative Foundation Models

Abstract:

The integration of generative foundation models into geoscientific workflows represents a transformative shift in solving complex inverse problems. We explore advanced architectures for map synthesis via Conditional Flow Matching and volumetric inversion via 3D VAEs, leveraging magnetic, gravity, and drilling data. By constraining multi-modal generative priors with physical laws, we synthesize high-fidelity geologic insights from sparse, unorganized measurements. This approach accelerates mineral exploration by significantly reducing the cost and uncertainty of targeting subsurface anomalies. This synergy of cross-modal generative processes and potential field theory defines a new era of geologic intelligence.

Lorina Dascal

Principal Image Processing SpecialistAbbott Laboratories

Bio:

Lorina Dascal is a principal computer vision and image processing specialist at Abbott Labs. Her research interests include    deep learning for image/ video understanding, 3D medical shapes, multimodal fusion of imaging and neural partial differential equation in vision. She has authored 14 published papers and has earned 11 patents. She holds a   PhD in Applied Mathematics from Tel-Aviv University, she was a postdoctoral fellow and a research  assistant   in the Computer Science Department at the Technion. 

Title:

Automatic 3D Surface Reconstruction of the Left Atrium from Unorganized Contours

Abstract:

ICE (intracardiac echocardiography) is a valuable tool in cardiac catheterization and electrophysiology (EP) procedures, assisting physicians in visualizing anatomical details and in monitoring procedures like catheter ablation, septal defect closure, left atrial appendage occlusion, and valve implantation.  Our aim is to automatically create an accurate three-dimensional surface model of Left atrium from automatic segmented boundaries of ICE images.  We propose a modified Poisson reconstruction method with additional geometric constraints, which enables the creation of accurate, highly detailed and computationally efficient surfaces from diverse sets of unorganized and sparse contours.

Hani Bezalel

Tel Aviv University

Bio:

Hana Bezalel holds an M.Sc. from Tel Aviv University, supervised by Hadar Averbuch Elor. Her CVPR 2025 publication focuses on relative, in-the-wild, pose estimation in extreme settings. Previously Lead Computer Vision Engineer at Rafael, she currently serves as an Algorithm Developer at Mobileye, where her work centers on geometric computer vision and spatial understanding.

Title:

Extreme Rotations Estimation In The Wild

Abstract:

We present a technique and benchmark dataset for estimating the relative 3D orientation between a pair of Internet images captured in an extreme setting, where the images have limited or non-overlapping field of views. Prior work targeting extreme rotation estimation assume constrained 3D environments and emulate perspective images by cropping regions from panoramic views. However, real images captured in the wild are highly diverse, exhibiting variation in both appearance and camera intrinsics. In this work, we propose a Transformer-based method for estimating relative rotations in extreme real-world settings, and contribute the ExtremeLandmarkPairs dataset, assembled from scene-level Internet photo collections. Our evaluation demonstrates that our approach succeeds in estimating the relative rotations in a wide variety of extreme-view Internet image pairs, outperforming various baselines, including dedicated rotation estimation techniques and contemporary 3D reconstruction methods.

Or Kozlovsky

Senior AI Applied Researcher Bluewhite Robotics

Bio:

Or Kozlovsky is a Senior AI Applied Researcher at Bluewhite Robotics and was recently a Student Researcher at Google. His work and research focus on generative AI, spatial AI, and real-time computer vision in both 2D and 3D domains. With a strong record of bridging cutting-edge research with real-world computer vision applications across a broad range of areas, including medical, space, entertainment, and robotics.  

Currently, Or is an M.Sc. student at Tel Aviv University under the supervision of Prof. Amit Bermano, and holds dual B.Sc. degrees in Electrical Engineering and Economics from the Technion. 

Title:

BINA: Bootstrapped Intelligence for Novel Adaptation

Abstract:

Robotic systems in real-world environments face conditions unseen during development, and while foundation models promise better generalization, integrating them under real-time onboard constraints remains challenging. We introduce BINA , a deployment-driven framework for online perceptual adaptation. BINA leverages online sparse supervision from a VLM to incrementally distil semantic knowledge into an onboard perception module. Beyond single-robot learning, BINA supports fleet-level knowledge aggregation, enabling scalable adaptation to new environments. Demonstrated on off-road traversability estimation, BINA rapidly converges from zero prior knowledge through operator-guided driving. Although demonstrated on traversability, BINA is task-agnostic and applicable to other perception and autonomy tasks. 

Moshe Mandel

AI ResearcherThe Hebrew University of Jerusalem

Bio:

Earned an MSc in Computer Science from the Hebrew University of Jerusalem (HUJI) under the supervision of Dr. Yossi Adi. Research bridges audio and visual domains, grounded in deep learning methodologies. Work centers on creative generative AI at the intersection of machine learning and artistic expression.

Title:

Latent Space JAM: Layout-Guided Video Generation

Abstract:

Controlling video generation is commonly achieved through training-based methods such as fine-tuning or adding control modules, which require extra optimization, data, or architectural changes. We propose a training-free approach that leverages pretrained video diffusion models to control object layout and motion using coarse spatio-temporal layouts. Our method operates in two passes: first, it steers spatial placement and temporal evolution through prompt-guided cross-attention to produce a coarse visual guide; second, this guide conditions the same model to generate a high-quality, layout-consistent video. The approach enables structured, coordinated motion with strong temporal consistency, supporting complex trajectories and multi-object interactions without additional training.

Asaf Joseph

R&DLightricks

Bio:

Hold MS.c in computer science for HUJI under Prof. Shmuel Peleg supervision.Currently works at Lightricks; main research interest is Video Generation.

 

Title:

Latent Space JAM: Layout-Guided Video Generation

Abstract:

Controlling video generation is commonly achieved through training-based methods such as fine-tuning or adding control modules, which require extra optimization, data, or architectural changes. We propose a training-free approach that leverages pretrained video diffusion models to control object layout and motion using coarse spatio-temporal layouts. Our method operates in two passes: first, it steers spatial placement and temporal evolution through prompt-guided cross-attention to produce a coarse visual guide; second, this guide conditions the same model to generate a high-quality, layout-consistent video. The approach enables structured, coordinated motion with strong temporal consistency, supporting complex trajectories and multi-object interactions without additional training.

Zvi Stein

Sr. Algorithm DeveloperAlign Technology, Inc.

Bio:

Zvi Stein is an Algorithms Engineer at Align Technology, working on computer vision and 3D geometry pipelines for multi-view scanning. His work focuses on surface reconstruction, mesh refinement, and performance-critical implementations with GPU acceleration. He has experience building end-to-end systems, from image-based inference to real-time processing and quality evaluation, aimed at improving surface accuracy and robustness in challenging acquisition conditions.

Title:

Mesh Refinement from Multi-View RGB Using Image-Predicted Surface Normals

Abstract:

Accurate surface refinement in regions with fine geometric detail remains challenging in practical 3D acquisition pipelines, where reconstructed meshes are often limited by scan resolution and noise. Although many scanning systems capture high-resolution multi-view RGB imagery, exploiting these images for metric geometry refinement is difficult due to scale ambiguity and perspective effects inherent to wide field-of-view 2D projections.

We present a geometry-refinement pipeline that converts multi-view RGB observations into a consistent surface normal field and integrates it to deform an initial mesh toward a refined surface. The central approach is to use image-predicted surface normals as the primary refinement signal, providing scale-consistent geometric constraints that are not directly available from intensity values alone. Input views are selected and scored based on geometric visibility and viewpoint diversity to ensure robust coverage and stable convergence across the surface. To mitigate projection-induced distortions, images are undistorted and re-parameterized into locally aligned patches, with corresponding rotations applied to the predicted normals.

A U-Net model trained from scratch predicts normal maps on a dedicated network surface, while deformation is applied on a separate, explicitly upsampled sampling surface designed to absorb high-frequency detail beyond the resolution of the original reconstruction; an additional simplified surface supports efficient view selection and scoring. The refined normal field is fused by solving a Poisson formulation to recover metrically consistent vertex displacements. Experimental results demonstrate improved reconstruction fidelity in high-curvature and detail-critical regions, recovering subtle structures that are commonly smoothed or missing in scan-resolution-limited meshes.

Rafael Ben Ari

Software Tech Lead, Senior AI EngineerGeneral Motors

Bio:

Rafael is a Software Tech Lead with a passion for AI Engineering at General Motors, where his focus is on making AI systems reliable, efficient, and production-ready. His work centers on integrating AI into core systems, and optimizing the tradeoffs between cost, latency, and result quality, using engineering to give AI exactly the context it needs to perform well. Rafael builds agents, sandboxes, benchmarks, and metrics to deeply understand system behavior and design AI solutions that actually work in the real world.

Title:

Introduction to Multi-Agent Architecture Patterns

Abstract:

A practical 15-minute technical session exploring the core principles of agentic workflows and how to move beyond single-agent limitations into coordinated, multi-agent AI systems. Attendees will learn key orchestration patterns, when to apply each, task decomposition strategies, overhead considerations, and an overview of leading frameworks. Participants will leave with a clear blueprint for designing robust multi-agent workflows and applying the right architecture for their AI applications.

A Taste from IMVC 2025