
Kunhao-Liu/3D-OVS - GitHub
This repository contains a pytorch implementation for the paper: Weakly Supervised 3D Open-vocabulary Segmentation. Our method can segment 3D scenes using open-vocabulary texts without requiring any segmentation annotations. Install environment: Please download the datasets from this link and put the datasets in ./data.
[CVPR2024 Highlight] LangSplat: 3D Language Gaussian Splatting
A PyTorch-based optimizer to produce a LangSplat model from SfM datasets with language feature inputs to; A scene-wise language autoencode to alleviate substantial memory demands imposed by explicit modeling. A script to help you turn your own images into optimization-ready SfM data sets with language feature
Overcoming Domain Limitations in Open-vocabulary Segmentation
2024年10月15日 · Open-vocabulary segmentation (OVS) has gained attention for its ability to recognize a broader range of classes. However, OVS models show significant performance drops when applied to unseen domains beyond the previous training dataset.
[ICCV2023] The repository contains the implementation of ... - GitHub
To address this challenge, our objective is to develop a fast OVS model that can perform comparably or better without the extra computational burden of the CLIP image encoder during inference. To this end, we propose a core idea of preserving the generalizable representation when fine-tuning on known classes.
[2305.14093] Weakly Supervised 3D Open-vocabulary …
2023年5月23日 · Distilling knowledge from pre-trained 2D open-vocabulary segmentation models helps but it compromises the open-vocabulary feature as the 2D models are mostly finetuned with close-vocabulary datasets. We tackle the challenges in 3D open-vocabulary segmentation by exploiting pre-trained foundation models CLIP and DINO in a weakly supervised manner.
Global Knowledge Calibration for Fast Open-Vocabulary …
Abstract: Recent advancements in pre-trained vision-language models, such as CLIP, have enabled the segmentation of arbitrary concepts solely from textual inputs, a process commonly referred to as open-vocabulary semantic segmentation (OVS). However, existing OVS techniques confront a fundamental challenge: the trained classifier tends to over ...
Transferable and Principled Efficiency for Open-Vocabulary …
In this paper, we introduce a transferable subnetwork approach that markedly reduces the size and computational demands of popular OVS models. Additionally, we propose a principled layer-selective fine-tuning method, enabling efficient fine-tuning of …
Revisit the Open Nature of Open Vocabulary Semantic Segmentation
In Open Vocabulary Semantic Segmentation (OVS), we observe a consistent drop in model performance as the query vocabulary set expands, especially when it includes semantically similar and ambiguous vocabularies, such as ‘sofa’ and ‘couch’.
In Open Vocabulary Semantic Segmentation (OVS), we observe a consistent drop in model performance as the query vocabulary set expands, especially when it includes semantically similar and ambiguous vocabularies, such as ‘sofa’ and ‘couch’.
Weakly Supervised 3D Open-vocabulary Segmentation
Distilling knowledge from pre-trained 2D open-vocabulary segmentation models helps but it compromises the open-vocabulary feature as the 2D models are mostly finetuned with close-vocabulary datasets. We tackle the challenges in 3D open-vocabulary segmentation by exploiting pre-trained foundation models CLIP and DINO in a weakly supervised manner.