JamBot Logo
1-92 of about 92 matches for site:arxiv.org feature
https://arxiv.org/abs/2306.09012
2306.09012] Yes, we CANN: Constrained Approximate Nearest Neighbors for local feature-based visual localization Skip to main
https://arxiv.org/abs/2207.07605
2207.07605] Algorithms to estimate Shapley value feature attributions arXiv Is Hiring a DevOps Engineer Work on
https://arxiv.org/abs/2308.16891
2308.16891] GNFactor: Multi-Task Real Robot Learning with Generalizable Neural Feature Fields Skip to main content We gratefully
https://arxiv.org/abs/2405.12979
2405.12979] OmniGlue: Generalizable Feature Matching with Foundation Model Guidance Skip to main content We gratefully acknowledge support from
https://arxiv.org/abs/2410.04642
2410.04642] The Optimization Landscape of SGD Across the Feature Learning Strength arXiv Is Hiring
https://arxiv.org/abs/2312.16610
2312.16610] Efficient Deweather Mixture-of-Experts with Uncertainty-aware Feature-wise Linear Modulation Skip to main
https://arxiv.org/abs/2206.08655
2206.08655] Learning Implicit Feature Alignment Function for Semantic Segmentation Skip to main content We gratefully acknowledge
https://arxiv.org/abs/2303.13953
2303.13953] AssetField: Assets Mining and Reconfiguration in Ground Feature Plane Representation Skip to main
https://arxiv.org/abs/2404.01223
2404.01223] Feature Splatting: Language-Driven Physics-Based Scene Synthesis and Editing Skip to main
https://arxiv.org/abs/2404.12862
2404.12862] A Guide to Feature Importance Methods for Scientific Inference arXiv Is Hiring a
https://arxiv.org/abs/1612.06321
scale image retrieval, referred to as DELF (DEep Local Feature). The new feature is based on convolutional
https://arxiv.org/abs/2104.05279
and simple approach to long-tailed visual recognition is to learn feature representations and a
https://arxiv.org/abs/2205.01643
transfer knowledge between domains via pseudo labels. We further propose the comprehensive multi-level feature alignment to improve
https://arxiv.org/abs/2308.06954
due to the significant storage and computation cost incurred by local feature matching in the
https://arxiv.org/abs/2001.05027
and local features into a single deep model, enabling accurate retrieval with efficient feature extraction. We refer to
https://arxiv.org/abs/2304.00583
and Describe Keypoints, by Guilherme Potje and 4 other authors View PDF Abstract: Local feature extraction is a
https://arxiv.org/abs/2207.10662
lines of each reference view. Each patch is linearly projected into a 1D feature vector and a
https://arxiv.org/abs/2304.02602
crucially, incorporates geometry priors in the form of a 3D feature volume. This latent feature
https://arxiv.org/abs/1603.08511
show that colorization can be a powerful pretext task for self-supervised feature learning, acting as a
https://arxiv.org/abs/2205.02841
frequent conditions (eg: enlarged cardiomediastinum, lung lesion, pneumothorax). Furthermore, we demonstrate that the double feature model improves performance for
https://arxiv.org/abs/2309.17002
that the reason behind is noise in pre-training shapes the feature space differently. We then
https://arxiv.org/abs/2112.09106
pretrains our model to align these region-text pairs in the feature space. When transferring our
https://arxiv.org/abs/2503.16413
techniques with foundation models, M3 builds a multimodal memory capable of rendering feature representations across granularities, encompassing
https://arxiv.org/abs/2309.03453
at every step of the reverse process through a 3D-aware feature attention mechanism that correlates
https://arxiv.org/abs/2404.12333
rendered from the target viewpoint. During training, we fine-tune the 3D feature prediction modules to
https://arxiv.org/abs/2405.14871
points and traces them through the NeRF representation to render feature vectors which are decoded
https://arxiv.org/abs/2401.02416
view and 3D cross-view information fusion. Our model differentiates 2D and 3D feature operations through the
https://arxiv.org/abs/2307.13226
as a radiance field with sparsely distributed and compactly factorized local tensor feature grids. Our approach leverages
https://arxiv.org/abs/2312.09398
beyond a single neural architecture. We combine an MLP decoder with a feature grid. Shading consists of
https://arxiv.org/abs/1611.05365
Gaussian mixtures to construct a highly non-linear visual spatiotemporal basketball assessment feature. Finally, we use this
https://arxiv.org/abs/1612.00796
of this and it has been widely thought that catastrophic forgetting is an inevitable feature of connectionist
https://arxiv.org/abs/2310.11448
rasterization and enables unprecedented rendering speed. Our representation is built on a 4D feature grid so that the
https://arxiv.org/abs/2003.00393
To implement such acquisition function, we propose a low-complexity method for feature density matching using self
https://arxiv.org/abs/2209.11799
LLM and (ii) Aug-Tree, which augments a decision tree with LLM feature expansions. Across a
https://arxiv.org/abs/2112.11427
representation with a style-based 2D generator. Our 3D implicit network renders low-resolution feature maps, from which the
https://arxiv.org/abs/2403.03221
our design choices and demonstrates that our method adapts flexibly to various feature extractors and correspondence
https://arxiv.org/abs/2405.14847
NeRF) for rendering specular objects. NDE transfers the concept of feature-grid-based spatial encoding
https://arxiv.org/abs/1904.03542
art and new adaptive evolutionary attackers need up to 10 times larger $L_0$ feature distance and 21
https://arxiv.org/abs/2109.01349
quality and high-fidelity results. Our proposed method generalizes the standard patch-based feature matching with spatial alignment
https://arxiv.org/abs/2303.09665
test). To this end, we first find interaction areas and extract their feature embeddings. Then we learn
https://arxiv.org/abs/2309.09457
Processes (MDPs) is that the learner has access to a known feature map $\phi(x, a
http://arxiv.org/abs/1312.4400
with a multilayer perceptron, which is a potent function approximator. The feature maps are obtained by
https://arxiv.org/abs/2003.01607
spaces. Most deep learning-based methods rely on a late fusion technique whereby multiple feature types are encoded and
https://arxiv.org/abs/2403.04115
approach utilizing diffusion training to learn a vision and language feature that encapsulates the
http://arxiv.org/abs/1312.4996
Survey: Analysis of Potential Systematics in Fitting of Baryon Acoustic Feature Skip to main
https://arxiv.org/abs/2306.05410
unlike the baselines. Our results also indicate our model can be complementary to feature-based SfM pipelines as
https://arxiv.org/abs/2306.12423
in-depth analyses, such as the comparison across different types of point feature, the necessity
http://arxiv.org/abs/1211.5555
nucleation dynamics, which in turn determines many snow-crystal characteristics. A key feature in our
http://arxiv.org/abs/1111.2786
plate-like ice crystals from water vapor near -15 C, which is a dominant feature in the
https://arxiv.org/abs/2312.02222
shot inversion techniques fail to fully leverage multiple input images for detailed feature extraction. We propose a
https://arxiv.org/abs/2112.07945
consistent images in real time but also produces high-quality 3D geometry. By decoupling feature generation and neural
https://arxiv.org/abs/2210.10362
in a joint optimization framework. Particularly, CPL constructs counterfactual by identifying minimal non-spurious feature change between semantically-similar
https://arxiv.org/abs/2310.02541
the data distribution is not linearly separable. Our proofs rely on analyzing the feature learning process under GD
https://arxiv.org/abs/1501.01990
from a constructive reinforcement effect of the standing waves that surround any feature. There are several common
https://arxiv.org/abs/1907.05388
sqrt{d^3H^3T})$ regret, where $d$ is the ambient dimension of feature space, $H$ is the
http://arxiv.org/abs/gr-qc/9311010
space of gauge equivalent connections is developed. Loops, knots, links and graphs feature prominently in this
https://arxiv.org/abs/1903.07603
source or measurement, increasing the odds that it results from a cosmological feature beyond LambdaCDM. Comments: accepted
https://arxiv.org/abs/2010.01374
optimal action-value function lies in the span of a feature map that is available
https://arxiv.org/abs/2106.04067
a local attention map specifically for each position in the feature. By combining the
https://arxiv.org/abs/2408.07495
to determine if two sites are (according to the Related Website Sets feature) related to each
https://arxiv.org/abs/2304.11900
field with an additional visibility field--it not only resolves occlusion ambiguity in multiview feature aggregation, but can also
https://arxiv.org/abs/2210.16859
propose a statistical model -- a joint generative data model and random feature model -- that captures this
https://arxiv.org/abs/2311.11666
is to lift multi-view inconsistent 2D segmentations into a consistent 3D feature field through a
https://arxiv.org/abs/1711.02305
model. This enables memory-limited execution of several statistical analyses over streams, including online feature selection, streaming data explanation
https://arxiv.org/abs/2307.03997
Decision Processes -- where transition probabilities admit a low-rank factorization based on an unknown feature embedding -- offer a
https://arxiv.org/abs/2401.16437
capable of processing raw radar imagery without the need for manual feature extraction required for
https://arxiv.org/abs/2504.03011
lighting module is combined with the diffusion models through the spatio-temporal feature blending algorithms without extra
https://arxiv.org/abs/2003.04448
to-end approaches typically extract entities from inputs by directly interpreting the latent feature representations as a
https://arxiv.org/abs/2505.20283
a sparse set of spatially grounded blobs and an instance-aware feature volume to disentangle
https://arxiv.org/abs/2405.17401
to the given text prompt. We also introduce a cross-attention-based feature aggregation scheme that allows
https://arxiv.org/abs/1509.03700
for colour maps to have perceptual flat spots that can hide a feature as large as one
http://arxiv.org/abs/1101.1559
on Galactic structure and chemical evolution, measurements of the baryon oscillation feature in the
https://arxiv.org/abs/2304.00776
a hybrid masking strategy, which enable dynamically updated guidance at test time and improve feature representation of the
http://arxiv.org/abs/0903.3591
over a fixed number field, uniformly in all aspects. A novel feature of the
http://arxiv.org/abs/1101.1529
and Lya forest spectra of 150,000 quasars, using the BAO feature of large
https://arxiv.org/abs/2010.10325
contains a variant of the $\tau$ map, which is a feature conspicuously absent from the
http://arxiv.org/abs/1312.4877
and power spectrum, including density-field reconstruction of the baryon acoustic oscillation (BAO) feature. The acoustic
https://arxiv.org/abs/2301.04183
such a codec. We introduce a novel variational formulation that explicitly takes feature data relevant to
https://arxiv.org/a/ozyilkan_e_1.html
eess.IV) [10] arXiv:2207.08489 [ pdf , other ] Title: Neural Distributed Image Compression with Cross-Attention Feature Alignment Authors: Nitish Mital
http://arxiv.org/abs/1607.03155
method after applying reconstruction to reduce non-linear effects on the BAO feature. Using the anisotropic
http://arxiv.org/abs/1211.2616
Abstract: We report a detection of the baryon acoustic oscillation (BAO) feature in the
https://arxiv.org/a/harrow_a_1.html
Quantum Physics (quant-ph) [34] arXiv:1804.11326 [ pdf , other ] Title: Supervised learning with quantum enhanced feature spaces Authors: Vojtech Havlicek
https://arxiv.org/abs/2004.12452
subject. To handle portrait reenactment from unseen subjects, we also introduce a feature dictionary-based generative adversarial
https://arxiv.org/abs/2005.11742
achieve this by extending the Contextual Attention module to borrow high-resolution feature patches in the
https://arxiv.org/abs/2005.12399
of two-magnon bound states in the transverse susceptibility. This bound state feature generalizes the one
https://arxiv.org/abs/2112.09343
is to perform adversarial training so that point clouds in the feature space can align. However
https://arxiv.org/abs/1806.03589
all input pixels as valid ones, generalizes partial convolution by providing a learnable dynamic feature selection mechanism for
https://arxiv.org/abs/2011.14619
method to embed the UV-based representations into a continuous feature space, which enables garment
https://arxiv.org/abs/2202.01771
of the LM-based policy. We find that sequential input representations (vs. fixed-dimensional feature vectors) and LM
https://arxiv.org/abs/2206.01067
the user can specify an arbitrary collection of subsets of the feature space -- possibly intersecting -- and
https://arxiv.org/a/roberts_d_3.html
additional authors not shown) Subjects: Artificial Intelligence (cs.AI) [3] arXiv:2310.07765 [ pdf , other ] Title: Feature Learning and Generalization