Skip to content
View bnavard's full-sized avatar
🎯
Focusing
🎯
Focusing

Block or report bnavard

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Maximum 250 characters. Please don't include any personal information such as legal names or email addresses. Markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
bnavard/README.md

πŸ‘‹ Hi, I'm Pouyan Navard

I'm a Computer Vision Engineer focused on building intelligent systems that see, learn, and adapt β€” both in pixels and in physical space. My journey started with classical 3D reconstruction and has since evolved into designing deep learning systems for complex perception tasks in medical imaging and robotics.

🧭 My Journey So Far

I began my career fascinated by how machines perceive the 3D world β€” starting with point clouds, stereo vision, and Structure-from-Motion techniques. This interest eventually took me to The Ohio State University for a PhD, where I explored self-supervised learning on 3D volumetric medical images. Working with noisy and low-resolution data pushed me to think deeply about representation learning, robustness, and the limits of generalization.

As I dove deeper into research, I became increasingly drawn to generative models β€” how they can synthesize, imagine, and even control the visual world. This passion led me to Path Robotics Inc., where I now build photorealistic 3D object generators, diffusion models with fine-grained control, and world models that help robots learn from their environments.

🧠 What I Work On

  • Generative AI for Robotics: Creating realistic, controllable environments and objects for training autonomous systems.
  • Self-supervised 3D Vision: Learning robust features from sparse, noisy, or unlabeled data.
  • Representation Learning: Compressing complex visual input into meaningful, task-aware latent spaces.
  • MLOps for Research: Scaling experimentation with clean, automated pipelines for training and evaluation.

🌱 What Drives Me

I'm passionate about bridging the gap between academic research and real-world deployment β€” especially where models move beyond screen-based outputs and into the physical world. Whether it’s helping a robot weld with more precision or enabling ultrasound systems to make sense of blurry 3D scans, I’m driven by a single question:

"How can we build models that learn from the world, not just datasets?"

πŸ”— Where to Find Me


TL;DR: PhD hacker turning pixels into intelligence.

Pinned Loading

  1. OSUPCVLab/SegFormer3D OSUPCVLab/SegFormer3D Public

    Official Implementation of SegFormer3D: an Efficient Transformer for 3D Medical Image Segmentation (CVPR/W 2024)

    Python 179 27

  2. aminK8/KnobGen aminK8/KnobGen Public

    CVPR 2025 Workshop on CVEU.

    Python 42 5

  3. OSUPCVLab/ERDES OSUPCVLab/ERDES Public

    ERDES: A Benchmark Video Dataset for Retinal Detachment and Macular Status Classification in Ocular Ultrasound (Nature Scientific Data 2025)

    Python 2