kth.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Registering Neural 4D Gaussians for Endoscopic Surgery
The Chinese University of Hong Kong (CUHK), Department of Electronic Engineering, Hong Kong, China; Shenzhen Research Institute, CUHK, Shenzhen, China.
The Chinese University of Hong Kong (CUHK), Department of Electronic Engineering, Hong Kong, China; Shenzhen Research Institute, CUHK, Shenzhen, China.
KTH, School of Electrical Engineering and Computer Science (EECS), Intelligent systems, Robotics, Perception and Learning, RPL.
The Hong Kong University of Science and Technology, Hong Kong, China.
Show others and affiliations
2024 (English)Conference paper, Published paper (Other academic)
Abstract [en]

The recent advance in neural rendering has enabled the ability to reconstruct high-quality 4D scenes using neural networks. Although 4D neural reconstruction is popular, registration for such representations remains a challenging task, especially for dynamic scene registration in surgical planning and simulation. In this paper, we propose a novel strategy for dynamic surgical neural scene registration. We first utilize 4D Gaussian Splatting to represent the surgical scene and capture both static and dynamic scenes effectively. Then, a spatial aware feature aggregation method, Spatially Weight Cluttering (SWC) is proposed to accurately align the feature between surgical scenes, enabling precise and realistic surgical simulations. Lastly, we present a novel strategy of deformable scene registration to register two dynamic scenes. By incorporating both spatial and temporal information for correspondence matching, our approach achieves superior performance compared to existing registration methods for implicit neural representation. The proposed method has the potential to improve surgical planning and training, ultimately leading to better patient outcomes.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE) , 2024. p. 1996-2001
National Category
Medical Imaging Neurosciences Computer graphics and computer vision
Identifiers
URN: urn:nbn:se:kth:diva-362215DOI: 10.1109/ROBIO64047.2024.10907406Scopus ID: 2-s2.0-105001480768OAI: oai:DiVA.org:kth-362215DiVA, id: diva2:1951009
Conference
2024 IEEE International Conference on Robotics and Biomimetics, ROBIO 2024, Bangkok, Thailand, December 10-14, 2024
Note

Part of ISBN 9781665481090

QC 20250410

Available from: 2025-04-09 Created: 2025-04-09 Last updated: 2025-04-10Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Ikemura, Kei

Search in DiVA

By author/editor
Ikemura, Kei
By organisation
Robotics, Perception and Learning, RPL
Medical ImagingNeurosciencesComputer graphics and computer vision

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 16 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf