BIM-JEPA: Self-supervised learning for BIM element classification using a joint embedding predictive architecture

Automation in Construction (Under Review)

  • Jack Wei Lun Shi1,
  • Wawan Solihin1,2,
  • Yufeng Weng1,
  • Yimin Zhao1,
  • Leong Hien Poh1,
  • Justin Ker-Wei Yeoh1
  • 1Department of Civil and Environmental Engineering, National University of Singapore
  • 2Research and Innovation, NovaCITYNETS Pte.Ltd.
teaser

Abstract

The development of scalable models for automated Building Information Modeling (BIM) element classification is hindered by the reliance on supervised learning, which requires expensive and laborious manual data annotation. This paper introduces BIM-JEPA, a foundation model that leverages a Joint Embedding Predictive Architecture for self-supervised pre-training on unlabeled 3D point cloud representations of individual BIM elements. By predicting the latent representations of masked regions of element geometry, BIM-JEPA learns semantically rich features that achieve competitive accuracy on a downstream classification task, outperforming existing supervised methods without heavy data augmentation, while excelling in data-scarce scenarios. This paper mitigates the data annotation bottleneck and establishes a path toward developing a foundation model for BIM geometry, enabling more scalable, data-efficient, and generalizable representation learning in the Architecture, Engineering, and Construction domain.