This repository contains model weights for LitePT: Lighter Yet Stronger Point Transformer, a lightweight, high-performance 3D point cloud architecture. LitePT proposes an improved 3D point cloud backbone that employs convolutions in early stages and switches to attention for deeper layers, introducing a novel, training-free 3D positional encoding, PointROPE. The resulting model achieves state-of-the-art performance while being significantly more efficient.
Paper & Resources
Models
We release the pretrained model weights for the benchmarks we reported in our paper.
Semantic segmentation
Instance segmentation
| Model |
Params |
Benchmark |
mAP25 |
mAP50 |
mAP |
Config |
Checkpoint |
| LitePT-S* |
16.0M |
ScanNet |
78.5 |
64.9 |
41.7 |
link |
Download |
| LitePT-S* |
16.0M |
ScanNet200 |
40.3 |
33.1 |
22.2 |
link |
Download |
Object detection
| Model |
Params |
Benchmark |
mAPH |
Config |
Checkpoint |
| LitePT |
9.0M |
Waymo |
70.7 |
link |
Download |
Citation
@article{yuelitept2025,
title={{LitePT: Lighter Yet Stronger Point Transformer}},
author={Yue, Yuanwen and Robert, Damien and Wang, Jianyuan and Hong, Sunghwan and Wegner, Jan Dirk and Rupprecht, Christian and Schindler, Konrad},
journal={arXiv preprint arXiv:2512.13689},
year={2025}
}