About me
My name is Boxiang Rong(荣波翔), I am currently a Master student of Robotics, Systems, and Control (RSC) in ETH Zurich. I obtained my Bachelor degree in Tianjin University, and I was supervised by Prof. Dr. Yuxiong Wang for my Bachelor thesis.
My interests lie in the fields of computer graphics and computer vision. I am fascinated with creating immersive virtual world and committed to working in the areas of rendering, character motion, and physics-based simulation.
Education
- 2022 - Present, MSc, Robotics Systems and Contorl, ETH Zurich, Switzerland.
- 2018 - 2022, B.E., Communication Engineering, Tianjin University, China.
Publication
4D-DRESS: A 4D Dataset of Real-World Human Clothing With Semantic Annotations
Wenbo Wang*, Hsuan-I Ho*, Chen Guo, Boxiang Rong, Artur Grigorev, Jie Song, Juan Jose Zarate, Otmar Hilliges
Conference on Computer Vision and Pattern Recognition (CVPR), 2024
[Project Page][Arxiv][Video]
Selected Projects
Nori CPU Render
Boxiang Rong*, Ziyao Shang*
Course project of Computer Graphics 2022 in ETH Zurich
[Website][Present][Code(not available)]
We built our own offline physically-based renderer on Nori – an educational ray tracing framework, written in C++. Our final image is selected for presentation in ETH 2022 Rendering Competition.
Theme for this year’s rendering competition is “Out Of Place”. Our work is inspired by Chinese anime movie Big Fish & Begonia and ancient literature A Happy Excursion(逍遥游). In which we built huge fishes(named K’un) flying in the universe, and voyagers sailing on clouds towards the northernmost of the world.
In the project, I built the whole scene in blender and implemented rendering functions, like Multi-Importance Sampling, Path Tracing, Photon Mapping, Environment Map, Disney BSDF, Procedural Volume, Texture Mapping and Low-Discrepancy Sampling .etc
Position-based Fluid and Foam Simulation
Boxiang Rong*, Longteng Duan*, Guo Han*
Course project of Physically-Based Simulation 2023 in ETH Zurich
[Slide][Demo][Code]
“Bunny in the water” – We implemented the 3D position-based fluids simulation using Taichi language. And we handled collisions of simple static rigid body and used AABB to facilitate. We reimplemented foam simulation from SPlisHSPlasH in Taichi to simulate white particles: spray, foam and bubbles. Thanks to the CPU/GPU parallelization empowered by Taichi, our whole pipeline can be simulated in real-time(~25FPS).
For the rendering part, we used splashsurf
to reconstruct fluid surface and built the whole scene in Blender. We used bpy
to instantiate models, materials and finish final rendering.
Motion Matching for Responsive Animation for Digital Humans
Boxiang Rong*, Longteng Duan*, Guo Han*, Hang Yin*
Course project of Digital Humans 2023 in ETH Zurich
[Report][Code][Demos][Slides]
Motion matching is a simple yet effective character animation tool for synthesizing motions that follow a user’s high-level command input. In this project, we implement a motion matching pipeline from scratch by utilizing publicly available motion capture dataset (LaFan1). Checkout the real-time (game-like video!) demo of responsive character animation.
To summarize the pipeline: After loading human skeleton motion data, we constructed 2d and 3d feature base for all frames of motions. Then we use spring damper system to generate smooth trajectories, this information will be combined with current states to match next motion. After each matching, inertialization is implemented to get a smooth pose transformation.
Apart from the pipeline, we achieved multiple means of control: keyboard, painting and real-time human pose. We collected a pose dataset to turn pose into command by classification, and leveraged FastPose to capture poses in real-time.
Rendering Trees in Mixed Reality
Boxiang Rong*, Marius Debussche*, David Kamm*, Adrien Lanne*
Course project of Mixed Reality 2023 in ETH Zurich
[Report][Video][Poster]
We developed an MR program in Unity for MagicLeap2 device, where you can select different types of trees and plant them on the reconstructed ground.
We built upon Unity’s URP rendering pipeline, and achieved realistic rendering of trees by custimizing Shader Graph. Our pipeline supports billboard-based models, RGB and normal mapping, environment map. Additionally, we employ Perlin Noise to create dynamic leaf oscillation animations. Leveraging the Mixed Reality Toolkit (MRTK), we customize the user interface for enhanced interactivity, elevating the tree planting experience.
Head-Worn Camera Image Stabilization using Neural Radiance Field
Boxiang Rong, Zilong Deng, Ziyao Shang, Minjing Shi
Course project of 3D Vision 2023 in ETH Zurich
[Report][Code1][Code2][Demos][Poster]
Motion blur happens with fast head-camera movement and long exposure times. In this project, we proposed to reconstruct the indoor environment in advance and render clear images to replace blurry camera views to achieve the goal of image stabilization. We used four pipelines, including traditional reconstruction and NeRF-based methods (Depth-Supervised NeRF, Deblur NeRF, Instant NGP). We did experiments to test the stabilizing performance and analyze the pros and cons of each pipeline. We also proposed adding camera motion information to Deblur NeRF for better deblurring and verified the performance of our modification.
Main Contributions:
- Build mesh reconstruction pipeline with Open3d, including aligning depth to RGB images, Point-Cloud reconstruction and stitching, Poisson surface reconstruction, and Color Map Optimization
- Modified Deblur Nerf to load our dataset and add trajectory information to better model blurry pattern
Priority Labeling in the Scenarios of Self-Driving
Boxiang Rong, Hang Zhao
Research project in Shanghai Qizhi AI Lab 2022
[Code][Slides]
Street scenarios are generally complex, where numbers of pedestrains, vehicles and obstacles interact with each other. However, we don’t need to pay equal attention to all of the visible targets. Instead, similar to human drivers, we only care about the important target that are most likely to appear in our way and influence our driving.
Our work is done on on Nuscenes dataset. We arranged human drivers to manually annotate priority levels to all visible targets. And we also designed rule-based algorithm to automatically annotate priorities, which considers information, like distance and relative velocity between ego and targets, trajectory intersections, and map-based terrain information.