publications
publications by categories in reversed chronological order. generated by jekyll-scholar.
2025
- Mahalanobis k-NN: A Statistical Lens for Robust Point-Cloud RegistrationsTejas Anvekar, and Shivanand Venkanna SheshappanavarFeb 2025
In this paper, we discuss Mahalanobis k-NN: a statistical lens designed to address the challenges of feature matching in learning-based point cloud registration when confronted with an arbitrary density of point clouds, either in the source or target point cloud. We tackle this by adopting Mahalanobis k-NN’s inherent property to capture the distribution of the local neighborhood and surficial geometry. Our method can be seamlessly integrated into any local-graph-based point cloud analysis method. In this paper, we focus on two distinct methodologies: Deep Closest Point (DCP) and Deep Universal Manifold Embedding (DeepUME). Our extensive benchmarking on the ModelNet40 and Faust datasets highlights the efficacy of the proposed method in point cloud registration tasks. Moreover, we establish for the first time that the features acquired through point cloud registration inherently can possess discriminative capabilities. This is evident by a substantial improvement of about 20% in the average accuracy observed in the point cloud few-shot classification task benchmarked on ModelNet40 and ScanObjectNN. The code is publicly available at https://github.com/TejasAnvekar/Mahalanobis-k-NN
@misc{anvekar2025mahalanobisknnstatisticallens, title = {Mahalanobis k-NN: A Statistical Lens for Robust Point-Cloud Registrations}, author = {Anvekar, Tejas and Sheshappanavar, Shivanand Venkanna}, month = feb, year = {2025}, eprint = {2409.06267}, bibtex_show = true, }
2024
- Novel Class Discovery for Representation of Real-World Heritage Data as Neural Radiance Fields (Student Abstract)Shivanand Kundargi, Tejas Anvekar, Ramesh Ashok Tabib, and Uma MudenagudiIn Proceedings of the AAAI Conference on Artificial Intelligence, Mar 2024
Neural Radiance Fields (NeRF) have been extensively explored as a leading approach for modeling and representing 3D data across various domains. Their ability to capture arbitrary scale point clouds and generate novel views makes them particularly valuable for digitizing cultural heritage sites. However, despite their impressive rendering capabilities, prior methods have often overlooked a significant real-world challenge: handling open-world scenarios characterized by unstructured data containing multiple classes in a single set of unlabeled images. To address this challenge, we propose a novel method NCD-NeRF that leverages Novel-Class Discovery to effectively tackle the complexities inherent in real-world data with unlabeled classes while excelling in producing high-quality NeRF representation. To validate our approach, we conducted a benchmarking analysis using a custom-collected dataset featuring UNESCO World Heritage sites in India. We observe that our proposed NCD-NeRF can parallely discover novel classes and render high-quality 3D volumes.
@inproceedings{aaai, author = {Kundargi, Shivanand and Anvekar, Tejas and Tabib, Ramesh Ashok and Mudenagudi, Uma}, title = {Novel Class Discovery for Representation of Real-World Heritage Data as Neural Radiance Fields (Student Abstract)}, month = mar, year = {2024}, booktitle = {Proceedings of the AAAI Conference on Artificial Intelligence}, bibtex_show = true, }
- A Benchmark Grocery Dataset of Realworld Point Clouds from Single ViewShivanand Venkanna Sheshappanavar, Tejas Anvekar, Shivanand Kundargi, Yufan Wang, and Chandra KembhamettuIn 2024 International Conference on 3D Vision (3DV), Mar 2024
Fine-grained grocery object recognition is an important computer vision problem with broad applications in automatic checkout, in-store robotic navigation, and assistive technologies for the visually impaired. Existing datasets on groceries are mainly 2D images. Models trained on these datasets are limited to learning features from the regular 2D grids. While portable 3D sensors such as Kinect were commonly available for mobile phones, sensors such as LiDAR and TrueDepth, have recently been integrated into mobile phones. Despite the availability of mobile 3D sensors, there are currently no dedicated real-world large-scale benchmark 3D datasets for grocery. In addition, existing 3D datasets lack fine-grained grocery categories and have limited training samples. Furthermore, collecting data by going around the object versus the traditional photo capture makes data collection cumbersome. Thus, we introduce a large-scale grocery dataset called 3DGrocery100. It constitutes 100 classes, with a total of 87,898 3D point clouds created from 10,755 RGB-D single-view images. We benchmark our dataset on six recent stateof-the-art 3D point cloud classification models. Additionally, we also benchmark the dataset on few-shot and continual learning point cloud classification tasks. Project Page: https://bigdatavision.org/3DGrocery100/.
@inproceedings{3dv, author = {Sheshappanavar, Shivanand Venkanna and Anvekar, Tejas and Kundargi, Shivanand and Wang, Yufan and Kembhamettu, Chandra}, title = {A Benchmark Grocery Dataset of Realworld Point Clouds from Single View}, month = mar, year = {2024}, booktitle = {2024 International Conference on 3D Vision (3DV)}, bibtex_show = true, }
2023
- GPr-Net: Geometric Prototypical Network for Point Cloud Few-Shot LearningTejas Anvekar, and Dena BazazianIn Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, Jun 2023
In the realm of 3D-computer vision applications, point cloud few-shot learning plays a critical role. However, it poses an arduous challenge due to the sparsity, irregularity, and unordered nature of the data. Current methods rely on complex local geometric extraction techniques such as convolution, graph, and attention mechanisms, along with extensive data-driven pre-training tasks. These approaches contradict the fundamental goal of few-shot learning, which is to facilitate efficient learning. To address this issue, we propose GPr-Net (Geometric Prototypical Network), a lightweight and computationally efficient geometric prototypical network that captures the intrinsic topology of point clouds and achieves superior performance. Our proposed method, IGI++(Intrinsic Geometry Interpreter++) employs vector-based hand-crafted intrinsic geometry interpreters and Laplace vectors to extract and evaluate point cloud morphology, resulting in improved representations for FSL (Few-Shot Learning). Additionally, Laplace vectors enable the extraction of valuable features from point clouds with fewer points. To tackle the distribution drift challenge in few-shot metric learning, we leverage hyperbolic space and demonstrate that our approach handles intra and inter-class variance better than existing point cloud few-shot learning methods. Experimental results on the ModelNet40 dataset show that GPr-Net outperforms state-of-the-art methods in few-shot learning on point clouds, achieving utmost computational efficiency that is 170x better than all existing works. The code is publicly available at https://github. com/TejasAnvekar/GPr-Net.
@inproceedings{Anvekar_2023_CVPR, author = {Anvekar, Tejas and Bazazian, Dena}, title = {GPr-Net: Geometric Prototypical Network for Point Cloud Few-Shot Learning}, booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops}, month = jun, year = {2023}, pages = {4178-4187}, bibtex_show = true, selected = true, }
- IPD-Net: SO(3) Invariant Primitive Decompositional Network for 3D Point CloudsRamesh Ashok Tabib, Nitishkumar Upasi, Tejas Anvekar, Dikshit Hegde, and Uma MudenagudiIn Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, Jun 2023
In this paper, we propose IPD-Net: Invariant Primitive Decompositional Network, a SO (3) invariant framework for decomposition of a point cloud. The human cognitive system is able to identify and interpret familiar objects regardless of their orientation and abstraction. Recent research aims to bring this capability to machines for understanding the 3D world. In this work, we present a framework inspired by human cognition to decompose point clouds into four primitive 3D shapes (plane, cylinder, cone, and sphere) and enable machines to understand the objects in various orientations. We employ Implicit Invariant Features (IIF) to learn local geometric relations by implicitly representing the point cloud with enhanced geometric information invariant towards SO (3) rotations. We also use Spatial Rectification Unit (SRU) to extract invariant global signatures. We demonstrate the results of our proposed methodology for SO (3) invariant decomposition on TraceParts Dataset, and show the generalizability of proposed IPD-Net as plugin for downstream task on classification of point clouds. We compare the results of classification with state-of-the-art methods on benchmark dataset (ModelNet40).
@inproceedings{Tabib_2023_CVPR, author = {Tabib, Ramesh Ashok and Upasi, Nitishkumar and Anvekar, Tejas and Hegde, Dikshit and Mudenagudi, Uma}, title = {IPD-Net: SO(3) Invariant Primitive Decompositional Network for 3D Point Clouds}, booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops}, month = jun, year = {2023}, pages = {2735-2743}, bibtex_show = true, }
- PointCLIMB: An Exemplar-Free Point Cloud Class Incremental BenchmarkShivanand Kundargi, Tejas Anvekar, Ramesh Ashok Tabib, and Uma MudenagudiJun 2023
Point clouds offer comprehensive and precise data regarding the contour and configuration of objects. Employing such geometric and topological 3D information of objects in class incremental learning can aid endless application in 3D-computer vision. Well known 3D-point cloud class incremental learning methods for addressing catastrophic forgetting generally entail the usage of previously encountered data, which can present difficulties in situations where there are restrictions on memory or when there are concerns about the legality of the data. Towards this we pioneer to leverage exemplar free class incremental learning on Point Clouds. In this paper we propose PointCLIMB: An exemplar Free Class Incremental Learning Benchmark. We focus on a pragmatic perspective to consider novel classes for class incremental learning on 3D point clouds. We setup a benchmark for 3D Exemplar free class incremental learning. We investigate performance of various backbones on 3D-Exemplar Free Class Incremental Learning framework. We demonstrate our results on ModelNet40 dataset.
@misc{kundargi2023pointclimb, title = {PointCLIMB: An Exemplar-Free Point Cloud Class Incremental Benchmark}, author = {Kundargi, Shivanand and Anvekar, Tejas and Tabib, Ramesh Ashok and Mudenagudi, Uma}, year = {2023}, eprint = {2304.06775}, archiveprefix = {arXiv}, primaryclass = {cs.CV}, bibtex_show = true, }
- 1st Workshop on Maritime Computer Vision (MaCVi) 2023: Challenge ResultsBenjamin Kiefer, Matej Kristan, Janez Perš, Lojze Žust, Fabio Poiesi, and 67 more authorsIn Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV) Workshops, Jan 2023
The 1st Workshop on Maritime Computer Vision (MaCVi)| 2023 focused on maritime computer vision for Unmanned| Aerial Vehicles (UAV) and Unmanned Surface Vehicle (USV),| and organized several subchallenges in this domain:(i) UAV-based Maritime Object Detection,(ii) UAV-based Maritime Object Tracking,(iii) USV-based Maritime Obstacle Segmentation and (iv) USV-based Maritime Obstacle Detection. The subchallenges were based on the SeaDronesSee and MODS| benchmarks. This report summarizes the main findings of the| individual subchallenges and introduces a new benchmark,| called SeaDronesSee Object Detection v2, which extends the| previous benchmark by including more classes and footage.| We provide statistical and qualitative analyses, and assess| trends in the best-performing methodologies of over 130| submissions. The methods are summarized in the appendix.| The datasets, evaluation code and the leaderboard are| publicly available (https://seadronessee. cs. uni-tuebingen. de/macvi).
@inproceedings{Kiefer_2023_WACV, author = {Kiefer, Benjamin and Kristan, Matej and Per\v{s}, Janez and \v{Z}ust, Lojze and Poiesi, Fabio and Andrade, Fabio and Bernardino, Alexandre and Dawkins, Matthew and Raitoharju, Jenni and Quan, Yitong and Atmaca, Adem and H\"ofer, Timon and Zhang, Qiming and Xu, Yufei and Zhang, Jing and Tao, Dacheng and Sommer, Lars and Spraul, Raphael and Zhao, Hangyue and Zhang, Hongpu and Zhao, Yanyun and Augustin, Jan Lukas and Jeon, Eui-ik and Lee, Impyeong and Zedda, Luca and Loddo, Andrea and Di Ruberto, Cecilia and Verma, Sagar and Gupta, Siddharth and Muralidhara, Shishir and Hegde, Niharika and Xing, Daitao and Evangeliou, Nikolaos and Tzes, Anthony and Bartl, Vojt\v{e}ch and \v{S}pa\v{n}hel, Jakub and Herout, Adam and Bhowmik, Neelanjan and Breckon, Toby P. and Kundargi, Shivanand and Anvekar, Tejas and Tabib, Ramesh Ashok and Mudenagudi, Uma and Vats, Arpita and Song, Yang and Liu, Delong and Li, Yonglin and Li, Shuman and Tan, Chenhao and Lan, Long and Somers, Vladimir and De Vleeschouwer, Christophe and Alahi, Alexandre and Huang, Hsiang-Wei and Yang, Cheng-Yen and Hwang, Jenq-Neng and Kim, Pyong-Kun and Kim, Kwangju and Lee, Kyoungoh and Jiang, Shuai and Li, Haiwen and Ziqiang, Zheng and Vu, Tuan-Anh and Nguyen-Truong, Hai and Yeung, Sai-Kit and Jia, Zhuang and Yang, Sophia and Hsu, Chih-Chung and Hou, Xiu-Yu and Jhang, Yu-An and Yang, Simon and Yang, Mau-Tsuen}, title = {1st Workshop on Maritime Computer Vision (MaCVi) 2023: Challenge Results}, booktitle = {Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV) Workshops}, month = jan, year = {2023}, pages = {265-302}, bibtex_show = true, }
- ASUR3D: Arbitrary Scale Upsampling and Refinement of 3D Point Clouds Using Local Occupancy FieldsAkash Kumbar, Tejas Anvekar, Ramesh Ashok Tabib, and Uma MudenagudiIn Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV) Workshops, Oct 2023
In this paper, we introduce ASUR3D, a novel methodology for the arbitrary-scale upsampling of 3D point clouds employing Local Occupancy Representation. Our proposed implicit occupancy representation enables efficient point classification, effectively discerning points belonging to the surface from non-surface points. Learning an implicit representation of open surfaces, enables one to capture the better local neighbourhood representation, leading to finer refinement and reconstruction with enhanced preservation of intricate geometric details. Leveraging this capability, we can accurately sample an arbitrary number of points on the surface, facilitating precise and flexible upsampling. We demonstrate the effectiveness of ASUR3D on PUGAN and PU1K benchmark datasets. Our proposed method achieves state-of-the-art results on all benchmarks and for all evaluation metrics. Additionally, we demonstrate the efficacy of our methodology on self-proposed heritage data generated through photogrammetry, further confirming its effectiveness in diverse scenarios. The code is publicly available at https://github. com/Akash-Kumbar/ASUR3D
@inproceedings{Kumbar_2023_ICCV, author = {Kumbar, Akash and Anvekar, Tejas and Tabib, Ramesh Ashok and Mudenagudi, Uma}, title = {ASUR3D: Arbitrary Scale Upsampling and Refinement of 3D Point Clouds Using Local Occupancy Fields}, booktitle = {Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV) Workshops}, month = oct, year = {2023}, pages = {1652-1661}, selected = true, bibtex_show = true, }
- DeFi: Detection and Filling of Holes in Point Clouds Towards Restoration of Digitized Cultural Heritage ModelsRamesh Ashok Tabib, Dikshit Hegde, Tejas Anvekar, and Uma MudenagudiIn Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV) Workshops, Oct 2023
In this paper, we propose DeFi: a novel perspective for hole detection and filling of a given deteriorated 3D point cloud towards digital preservation of cultural heritage sites. Preservation of heritage demands digitization as cultural heritage sites deteriorate due to natural calamities and human activities. Digital preservation promotes acquisition of 3D data using 3D sensor or Multi-view reconstruction. Unfortunately, 3D data acquisition finds challenges due to the limitations in sensor technology and inappropriate capture conditions, leading to formation of missing regions or holes in the acquired point cloud. To address this, we propose a pipeline consisting of detection of hole boundaries, and understanding the geometry of the hole boundaries to fill the region of the point cloud. Recent research on hole detection and filling fails to generalize on complex structures such as heritage sites, as they find challenges in differentiating between the hole boundary and non-hole boundary points. To address this, we propose to detect boundary points of point cloud and learn to classify them into" hole boundary" and" non-hole boundary" points. We generate a synthetic dataset based on ModelNet40 to learn the detection of hole boundaries. We demonstrate the results of the proposed pipeline on (i) ModelNet40 dataset,(ii) Heritage 3D models generated via photogrammetry, and compare the results with state-of-the-art methods.
@inproceedings{Tabib_2023_ICCV, author = {Tabib, Ramesh Ashok and Hegde, Dikshit and Anvekar, Tejas and Mudenagudi, Uma}, title = {DeFi: Detection and Filling of Holes in Point Clouds Towards Restoration of Digitized Cultural Heritage Models}, booktitle = {Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV) Workshops}, month = oct, year = {2023}, pages = {1603-1612}, bibtex_show = true, }
- TP-NoDe: Topology-Aware Progressive Noising and Denoising of Point Clouds Towards UpsamplingAkash Kumbar, Tejas Anvekar, Tulasi Amitha Vikrama, Ramesh Ashok Tabib, and Uma MudenagudiIn Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV) Workshops, Oct 2023
In this paper, we propose TP-NoDe, a novel Topology-aware Progressive Noising and Denoising technique for 3D point cloud upsampling. TP-NoDe revisits the traditional method of upsampling of the point cloud by introducing a novel perspective of adding local topological noise by incorporating a novel algorithm Density-Aware k nearest neighbour (DA-kNN) followed by denoising to map noisy perturbations to the topology of the point cloud. Unlike previous methods, we progressively upsample the point cloud, starting at a 2 x upsampling ratio and advancing to a desired ratio. TP-NoDe generates intermediate upsampling resolutions for free, obviating the need to train different models for varying upsampling ratios. TP-NoDe mitigates the need for task-specific training of upsampling networks for a specific upsampling ratio by reusing a point cloud denoising framework. We demonstrate the supremacy of our method TP-NoDe on the PU-GAN dataset and compare it with state-of-the-art upsampling methods. The code is publicly available at https://github. com/Akash-Kumbar/TPNoDe.
@inproceedings{Kumbar_2023_ICCW, author = {Kumbar, Akash and Anvekar, Tejas and Vikrama, Tulasi Amitha and Tabib, Ramesh Ashok and Mudenagudi, Uma}, title = {TP-NoDe: Topology-Aware Progressive Noising and Denoising of Point Clouds Towards Upsampling}, booktitle = {Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV) Workshops}, month = oct, year = {2023}, pages = {2272-2282}, bibtex_show = true, }
2022
- DA-AE: Disparity-Alleviation Auto-Encoder Towards Categorization of Heritage Images for Aggrandized 3D Reconstruction.Dikshit Hegde, Tejas Anvekar, Ramesh Ashok Tabib, and Uma MudengudiIn Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, Jun 2022
In this paper, we propose DA-AE: Disparity Alleviation AutoEncoder for categorization of heritage images towards 3D reconstruction. Recent survey on preservation of heritage shows demand for the digitization and conservations of heritage sites owing to their susceptibility to natural disasters and human acts. Digital conservation can be facilitated via crowdsourcing of data useful for construction of 3D models. Data from multiple sites sourced may result in elimination of relevant images due to the limitations of the pipeline. Curation and categorization of the crowdsourced data enables better 3D reconstruction. 3D reconstruction pipelines demand correlation between the data and also tries to eliminate the irrelevant information. The reconstruction pipeline is sensitive to selection of initial pair for reconstruction. By categorising individual sites, crowdsourced data can be used to create better 3D reconstructed models. Categorization of crowdsourced data demands learning robust representations of data. Towards this, we propose DA-AE for improved representation and categorization of data in latent space, along with a disparity alleviation loss. We demonstrate categorization as an event, with clustering as a downstream task. We compare our results of clustering with state-of-the-art methods on benchmark datasets (MNIST, FashionMNIST, and USPS). We demonstrate the effects of our categorization using custom dataset IDH10 and compare the results with state-of-the-art methods. We show a systematic and qualitative influence of the proposed method on 3D reconstruction of data.
@inproceedings{Hegde_2022_CVPR, author = {Hegde, Dikshit and Anvekar, Tejas and Tabib, Ramesh Ashok and Mudengudi, Uma}, title = {DA-AE: Disparity-Alleviation Auto-Encoder Towards Categorization of Heritage Images for Aggrandized 3D Reconstruction.}, booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops}, month = jun, year = {2022}, pages = {5093-5100}, bibtex_show = true, }
- Metric-KNN is All You NeedTejas Anvekar, Ramesh Ashok Tabib, Dikshit Hegde, and Uma MudenagudiIn SIGGRAPH Asia 2022 Posters, Jun 2022
@inproceedings{10.1145/3550082.3564198, abstrac = {In this work, we propose a novel Metric-K Nearest Neighbor (M-KNN) to facilitate topology aware learning in point clouds. Topology aware learning is achieved by accumulation of local features in deep-learning model. Recent work rely on Ball queries or K-Nearest-Neighbor (KNN) for local feature extraction of point clouds and finds challenges in retaining topological information. -KNN employes a generalised Minkowski distance in the KNN search algorithm for topological representation of point clouds. -KNN enables state-of-the-art point cloud methods to perform topology aware downstream tasks. We demonstrate the performance of -KNN as plugin towards point cloud classification, part-segmentation, and denoising using benchmark dataset. }, author = {Anvekar, Tejas and Tabib, Ramesh Ashok and Hegde, Dikshit and Mudenagudi, Uma}, title = {Metric-KNN is All You Need}, year = {2022}, isbn = {9781450394628}, publisher = {Association for Computing Machinery}, address = {New York, NY, USA}, url = {https://doi.org/10.1145/3550082.3564198}, doi = {10.1145/3550082.3564198}, booktitle = {SIGGRAPH Asia 2022 Posters}, articleno = {31}, numpages = {2}, keywords = {Classification, Denoising, K Nearest Neighbour, Point cloud representation, Segmentation}, location = {Daegu, Republic of Korea}, series = {SA '22}, bibtex_show = true, }
- VG-VAE: A Venatus Geometry Point-Cloud Variational Auto-EncoderTejas Anvekar, Ramesh Ashok Tabib, Dikshit Hegde, and Uma MudengudiIn Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, Jun 2022
In this paper, we propose VG-VAE: Venatus Geometric Variational Auto-Encoder for capturing unsupervised hierarchical local and global geometric signatures in pointcloud. Recent research emphasises the significance of the underlying intrinsic geometry for pointcloud processing. Our contribution is to extract and analyse the morphology of the pointcloud using the proposed Geometric Proximity Correlator (GPC) and variational sampling of the latent. The extraction of local geometric signatures is facilitated by the GPC, whereas the extraction of global geometry is facilitated by variational sampling. Furthermore, we apply a naive mix of vector algebra and 3D geometry to extract the basic per-point geometric signature, which assists the unsupervised hypothesis. We provide statistical analyses of local and global geometric signatures. The impacts of our geometric features are demonstrated on pointcloud classification as downstream task using the classic pointcloud feature extractor PointNet. We demonstrate our analysis on ModelNet40 a benchmark dataset, and compare with state-of-the-art techniques.
@inproceedings{Anvekar_2022_CVPR, author = {Anvekar, Tejas and Tabib, Ramesh Ashok and Hegde, Dikshit and Mudengudi, Uma}, title = {VG-VAE: A Venatus Geometry Point-Cloud Variational Auto-Encoder}, booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops}, month = jun, year = {2022}, pages = {2978-2985}, selected = true, bibtex_show = true, }