1-100 of about 132 matches for site:arxiv.org site:arxiv.org site:arxiv.org site:arxiv.org site:arxiv.org explicit
https://arxiv.org/abs/2311.10709
2311.10709] Emu Video: Factorizing Text-to-Video Generation by Explicit Image Conditioning Happy Open Access Week from arXiv
https://arxiv.org/abs/1806.03393
1806.03393] Explicit Coleman Integration in Larger Characteristic Skip to main content We gratefully acknowledge
https://arxiv.org/abs/2002.08988
masks, or multiple views of the same scene. Our experiments show that using explicit 3D features to
https://arxiv.org/abs/2406.01592
as output. The resulting method produces an output within seconds and offers explicit user control over the
https://arxiv.org/abs/2303.01469
zero-shot data editing, such as image inpainting, colorization, and super-resolution, without requiring explicit training on these tasks
https://arxiv.org/abs/2407.01525
great progress has been made in 3D visual grounding, current models still rely on explicit textual descriptions for
[2304.13681] Ray Conditioning: Trading Photo-consistency for Photo-realism in Multi-view Image Gener
https://arxiv.org/abs/2304.13681
and then rendered into 2D images to ensure photo-consistency across viewpoints. However, such explicit bias for photo
https://arxiv.org/abs/2101.02697
to-end manner solely based on a photometric re-rendering loss without requiring explicit 3D this http URL
https://arxiv.org/abs/2401.09416
for texture information extraction, variational score distillation for detailed appearance synthesis, and explicit geometry guidance with ControlNet
[2111.11215] Direct Voxel Grid Optimization: Super-fast Convergence for Radiance Fields Reconstructi
https://arxiv.org/abs/2111.11215
grid with a shallow network for complex view-dependent appearance. Modeling with explicit and discretized
https://arxiv.org/abs/2401.09416
for texture information extraction, variational score distillation for detailed appearance synthesis, and explicit geometry guidance with ControlNet
https://arxiv.org/abs/2406.07520
condition, simply by conditioning an image generator on a target environment map, without an explicit scene decomposition. Our method
https://arxiv.org/abs/2502.10377
to a real-world scene represented by multiple views. The method combines explicit semantic correspondences with multi
https://arxiv.org/abs/2410.16512
dense vision applications (e.g. depth estimation, semantic segmentation), despite the lack of explicit supervisory signals. In
https://arxiv.org/abs/2412.21079
solution based on diffusion models, featuring a fundamental design principle of using explicit image correspondence to
https://arxiv.org/abs/2211.15521
task of predicting the location where an image was taken. Here we study explicit knowledge from human-written
https://arxiv.org/abs/2404.12333
coarse view control. In this work, we introduce a new task -- enabling explicit control of the
https://arxiv.org/abs/2502.10377
to a real-world scene represented by multiple views. The method combines explicit semantic correspondences with multi
https://arxiv.org/abs/2412.21079
solution based on diffusion models, featuring a fundamental design principle of using explicit image correspondence to
https://arxiv.org/abs/2404.12333
coarse view control. In this work, we introduce a new task -- enabling explicit control of the
[2304.13681] Ray Conditioning: Trading Photo-consistency for Photo-realism in Multi-view Image Gener
https://arxiv.org/abs/2304.13681
and then rendered into 2D images to ensure photo-consistency across viewpoints. However, such explicit bias for photo
https://arxiv.org/abs/2408.14652
the regularity lemma for the case of Ta-Shma's explicit codes near the
[1906.01618] Scene Representation Networks: Continuous 3D-Structure-Aware Neural Scene Representatio
https://arxiv.org/abs/1906.01618
learning has explored 3D-structure-aware representations of scene geometry, these models typically require explicit 3D supervision. Emerging neural
https://arxiv.org/abs/2210.02875
state-of-the-art results on WikiTableQuestions and TabFact datasets, with explicit output programs that benefit
https://arxiv.org/abs/2312.08885
of objects with globality of scenes by introducing a hybrid 3D representation: explicit for objects
[2210.15893] When Life Gives You Lemons, Make Cherryade: Converting Feedback from Bad Responses into
https://arxiv.org/abs/2210.15893
to integrate human feedback to continuously improve themselves. However, humans may not always provide explicit signals when the
https://arxiv.org/abs/1209.5739
respect to the bound $n$ without having to rely on an explicit analytic formula for
https://arxiv.org/abs/2312.11461
as a powerful 3D representation that harnesses the advantages of both explicit (mesh) and implicit
[2012.09995] Data Leverage: A Framework for Empowering the Public in its Relationship with Technolog
https://arxiv.org/abs/2012.09995
4 other authors View PDF Abstract: Many powerful computing technologies rely on implicit and explicit data contributions from the
https://arxiv.org/abs/2312.11461
as a powerful 3D representation that harnesses the advantages of both explicit (mesh) and implicit
[2012.09995] Data Leverage: A Framework for Empowering the Public in its Relationship with Technolog
https://arxiv.org/abs/2012.09995
4 other authors View PDF Abstract: Many powerful computing technologies rely on implicit and explicit data contributions from the
https://arxiv.org/abs/2310.07018
investigate the physical reasoning capabilities of several mainstream language models across foundational, explicit, and implicit
https://arxiv.org/abs/2310.04415
from ResNets on vision tasks to LLMs: weight decay is never useful as an explicit regularizer but instead changes
[2406.00578] ContextFlow++: Generalist-Specialist Flow-based Generative Models with Mixed-Variable C
https://arxiv.org/abs/2406.00578
model. We propose ContextFlow++ approach to overcome these limitations using an additive conditioning with explicit generalist-specialist knowledge decoupling
[2306.04637] Transformers as Statisticians: Provable In-Context Learning with In-Context Algorithm S
https://arxiv.org/abs/2306.04637
different base ICL algorithms -- or even perform qualitatively different tasks -- on different input sequences, without any explicit prompting of the
[2504.19165] IM-Portrait: Learning 3D-aware Video Diffusion for Photorealistic Talking Heads from Mo
https://arxiv.org/abs/2504.19165
generating photorealistic talking head videos directly from a single identity image and explicit control signals (e.g., expressions
https://arxiv.org/abs/2312.06657
To this end, we propose to learn the color field as an explicit 2D appearance aggregation, also
https://arxiv.org/abs/2409.05862
human participants, collecting 35K trials of behavioral data from over 500 participants. This includes explicit choice behaviors as well
https://arxiv.org/abs/2306.01693
fine-grained human feedback (e.g., which sentence is false, which sub-sentence is irrelevant) as an explicit training signal. We introduce
https://arxiv.org/abs/2203.09905
region in objects, which faces the challenge of establishing an explicit link with object parts
https://arxiv.org/abs/2208.13196
individuals leads to diverse interactions, which makes it difficult to establish an explicit link between object parts
https://arxiv.org/abs/2312.04561
a deformation field to optical flows, making it possible to apply explicit structural regularizations for
https://arxiv.org/abs/1908.07070
image to orientation parameters, we propose an end-to-end framework that incorporates explicit geometric reasoning. In
https://arxiv.org/abs/2312.06657
To this end, we propose to learn the color field as an explicit 2D appearance aggregation, also
https://arxiv.org/abs/2312.07504
the other hand, the recently proposed 3D Gaussian Splatting provides new opportunities given its explicit point cloud representations. This
https://arxiv.org/abs/2502.10834
experimental) Abstract: Social media empower distributed content creation by algorithmically harnessing "the social fabric" (explicit and implicit
https://arxiv.org/abs/2309.06554
problem from the communication complexity point of view and constructing explicit protocols for ExactlyN
http://arxiv.org/abs/gr-qc/9302027
constraint. Using the new canonical coordinates, it is then easy to obtain explicit expressions of Dirac
https://arxiv.org/abs/1906.00103
using the Hankel continued fraction, instead of the $J$-fraction. Consequently, an explicit formula for the
https://arxiv.org/abs/2209.05521
Raymond F. Vozzo View PDF Abstract: Motivated by the problem of constructing explicit geometric string structures, we
https://arxiv.org/abs/2112.07945
quality of 3D GANs without overly relying on these approximations. We introduce an expressive hybrid explicit-implicit network architecture that
https://arxiv.org/abs/2310.08579
skeleton to fine-grained spatial geometry. Therefore, capturing such correlations between the explicit appearance and latent
https://arxiv.org/abs/2301.09457
4}$, thus (asymptotically) matching the best lower bound on trifferent codes. We also give explicit constructions of affine
https://arxiv.org/abs/2301.09457
4}$, thus (asymptotically) matching the best lower bound on trifferent codes. We also give explicit constructions of affine
https://arxiv.org/abs/2507.07230
between learned color embeddings and clothing attributes, validating color as an effective proxy when explicit clothing labels are unavailable
https://arxiv.org/abs/2111.01785
for the two agents to develop a communication protocol without explicit or implicit supervision. We
https://arxiv.org/abs/2210.05519
be capable of inferring objects and their properties from visual scenes without explicit supervision. However, existing works
[1512.03044] Enumeration and investigation of acute 0/1-simplices modulo the action of the hyperocta
https://arxiv.org/abs/1512.03044
that computes the cycle index of Bn, which can in explicit form only be found
https://arxiv.org/abs/2312.04558
the virtual and real worlds. Recent advancements in head avatar techniques, including explicit 3D morphable meshes (3DMM
https://arxiv.org/abs/2312.02157
for reconstructing and rendering photo-realistic views, yet they possess limited editability. Conversely, explicit 3D representations, such as
https://arxiv.org/abs/2112.07945
quality of 3D GANs without overly relying on these approximations. We introduce an expressive hybrid explicit-implicit network architecture that
https://arxiv.org/abs/1302.2274
d$ are greater than zero and the remaining elements are zero. We provide explicit recurrence relations to
https://arxiv.org/abs/2302.04871
composition of these two radiance fields with carefully designed regularization. We demonstrate that our explicit decomposition alleviates the
https://arxiv.org/abs/1110.5027
TUY] and the abelian analog in [AU1]. We here provide an explicit isomorphism from the
https://arxiv.org/abs/2308.12693
terms of alternating permutations in 2012. In this paper, we provide explicit descriptions of the
http://arxiv.org/abs/hep-th/9309063
Invariant, by Louis Crane and 2 other authors View PDF Abstract: We provide an explicit formula for the
http://arxiv.org/abs/hep-th/9202053
formulation of the ``loop transform'' proposed by Rovelli and Smolin. Several explicit representations of the
https://arxiv.org/abs/2308.12419
end approach based on iterative attention that allows recognition from a raw video without explicit hand detection. I further
https://arxiv.org/abs/1902.10782
uncertainty is measured exactly, its value is an eigenvalue of A. * has no explicit collapse -- the latter
[gr-qc/9211028] Thawing the Frozen Formalism: The Difference Between Observables and What We Observe
http://arxiv.org/abs/gr-qc/9211028
how dynamics is possible when all observables are constants of the motion. An explicit model of a
https://arxiv.org/abs/1702.01514
model for the string 2-group using the free loop group, with explicit formulas for all
https://arxiv.org/abs/0904.0125
2-theories, by Jonathan Asher Cohen View PDF Abstract: General coherence theorems are constructed that yield explicit presentations of categorical
[2005.10804] Reinforcement Learning with General Value Function Approximation: Provably Efficient Ap
https://arxiv.org/abs/2005.10804
generalizes recent progress on RL with linear value function approximation and does not make explicit assumptions on the
http://arxiv.org/abs/hep-th/9305101
Kirby calculus, by Boguslaw Broda (U. Clausthal and U. Lodz) View PDF Abstract: An explicit derivation of the
https://arxiv.org/abs/2312.04558
the virtual and real worlds. Recent advancements in head avatar techniques, including explicit 3D morphable meshes (3DMM
[1703.00542] A note on the approximate admissibility of regularized estimators in the Gaussian seque
https://arxiv.org/abs/1703.00542
Moreover, we simplify and shorten the original proof considerably. We also provide explicit upper and lower
https://arxiv.org/abs/2502.15657
explain data and a question-answering inference machine. Both components operate with an explicit notion of uncertainty
https://arxiv.org/abs/2205.10712
In Housekeep, an embodied agent must tidy a house by rearranging misplaced objects without explicit instructions specifying which objects
https://arxiv.org/abs/2107.14795
achieves state-of-the-art performance on Sintel optical flow estimation with no explicit mechanisms for multiscale
https://arxiv.org/abs/2404.11483
to the modular nature and the intuitive design to simulate explicit human thought process, a
https://arxiv.org/abs/2210.15601
18 figures. Comments welcomed! v2: new corollaries 2.3 and 4.5 with more explicit discussions of computability
https://arxiv.org/abs/2407.12883
score of nDCG@10 of 18.3 on BRIGHT. We show that incorporating explicit reasoning about the
[2107.07511] A Gentle Introduction to Conformal Prediction and Distribution-Free Uncertainty Quantif
https://arxiv.org/abs/2107.07511
Critically, the sets are valid in a distribution-free sense: they possess explicit, non-asymptotic guarantees even
https://arxiv.org/abs/2107.14795
achieves state-of-the-art performance on Sintel optical flow estimation with no explicit mechanisms for multiscale
https://arxiv.org/abs/2210.15601
18 figures. Comments welcomed! v2: new corollaries 2.3 and 4.5 with more explicit discussions of computability
https://arxiv.org/abs/2107.04121
multi-linear) can be expressed in a compact way, without the explicit loops, using a
https://arxiv.org/abs/2307.10852
construct concrete examples that show that these variations of the conjecture are false. Explicit emphasis is put on
[2310.04243] Towards Global Solutions for Nonconvex Two-Stage Stochastic Programs: A Polynomial Lowe
http://arxiv.org/abs/2310.04243
for prospective solutions like stationary solutions computed from other methods, but it also yields an explicit polynomial approximation for
https://arxiv.org/abs/1508.02661
and characterize Archimedean circular orders, in analogy with linear Archimedean orders. We describe explicit examples of circular
https://arxiv.org/abs/2306.03881
In this paper, we show that correspondence emerges in image diffusion models without any explicit supervision. We propose a
https://arxiv.org/abs/2003.04448
first use synthetic image experiments to demonstrate how our approach effectively decomposes objects without explicit supervision. Then, we insert
https://arxiv.org/abs/2311.04235
systems in a reliable manner. Model developers may wish to set explicit rules for the
https://arxiv.org/abs/2001.04383
implies a negative answer to Tsirelson's problem: we show, by providing an explicit example, that the
https://arxiv.org/abs/2002.08909
QA). We compare against state-of-the-art models for both explicit and implicit
https://arxiv.org/abs/2403.09632
complex non-Lambertian lighting effects (e.g., specular highlights and cast shadows) without using any explicit physical lighting priors. We
https://arxiv.org/abs/2311.01378
trained VLMs for single-step vision-language comprehension, models sequential history information with an explicit policy head, and
[2106.04067] LocalTrans: A Multiscale Local Transformer Network for Cross-Resolution Homography Esti
https://arxiv.org/abs/2106.04067
Existing deep homography methods concatenate the input images or features, neglecting the explicit formulation of correspondences
https://arxiv.org/abs/2312.02157
for reconstructing and rendering photo-realistic views, yet they possess limited editability. Conversely, explicit 3D representations, such as