DreamEdit3D: Personalization of Multi-View Diffusion Models for 3D Editing (2025)
This project presents DreamEdit3D, a novel framework for personalized 3D scene editing by leveraging multi-view diffusion models. The approach enables users to edit 3D scenes with fine-grained control through personalized text-driven modifications while maintaining multi-view consistency.
Technologies: Multi-View Diffusion, 3D Gaussian Splatting, Personalization, Deep Learning, PyTorch
- Multi-View Consistency: Ensures coherent edits across all viewpoints by personalizing multi-view diffusion models, avoiding the inconsistencies common in single-view editing approaches.
- Personalized Editing: Enables subject-driven 3D editing by fine-tuning diffusion models on user-provided reference images, allowing precise insertion and modification of objects in 3D scenes.
- 3D Reconstruction Integration: Combines edited multi-view outputs with 3D Gaussian Splatting for high-quality, real-time renderable 3D scene reconstruction.
Paper under submission to ECCV 2026 — GIFs and demo video coming soon.
GenSMPL: Generative Skinned Multi-Person Linear Model (2025)
GenSMPL is a data-driven framework for updating and extending the SMPL body model to better represent diverse body shapes and structural variations of children not captured by the original model. Our approach focuses on modifying key components such as the identity PCA space to adapt to entirely new categories.
Technologies: SMPL, PCA, 3D Body Modeling, Dense Registration, Synthetic Data Generation, PyTorch
- Controlled 3D Data Generation: Generating controlled 3D children's body data to serve as training input for learning new shape priors.
- Dense Registration: Performing dense registration to the SMPL template, ensuring consistent mesh topology across all generated samples.
- Shape Prior Learning: Learning new shape priors from aligned meshes via PCA, extending the generalization capability of SMPL to children categories.
Validated on synthetic and real-world datasets, achieving improved reconstruction accuracy and more realistic shape representation.
Paper under submission — visuals and demo coming soon.
MMOM System Fullstack Development (2023 to 2024)
- This project represents the culmination of 15 months of dedicated work during my tenure on the TOA team.
- Due to confidentiality clauses and agreements with TUM, I am unable to provide specific details or share any gifs, videos, or pictures.