(a) The main structure of the self-supervised pretraining model, including three parts—a token embedding at the forefront, followed by a hierarchical encoder–decoder and a point reconstruction module.
Some results have been hidden because they may be inaccessible to you
Show inaccessible results