keyword
https://read.qxmd.com/read/38656855/neuralrecon-real-time-coherent-3d-scene-reconstruction-from-monocular-video
#1
JOURNAL ARTICLE
Xi Chen, Jiaming Sun, Yiming Xie, Hujun Bao, Xiaowei Zhou
We present a novel framework named NeuralRecon for real-time 3D scene reconstruction from a monocular video. Unlike previous methods that estimate single-view depth maps separately on each key-frame and fuse them later, we propose to directly reconstruct local surfaces represented as sparse TSDF volumes for each video fragment sequentially by a neural network. A learning-based TSDF fusion module based on gated recurrent units is used to guide the network to fuse features from previous fragments. This design allows the network to capture local smoothness prior and global shape prior of 3D surfaces when sequentially reconstructing the surfaces, resulting in accurate, coherent, and real-time surface reconstruction...
April 24, 2024: IEEE Transactions on Pattern Analysis and Machine Intelligence
https://read.qxmd.com/read/38652604/exploring-the-semantic-inconsistency-effect-in-scenes-using-a-continuous-measure-of-linguistic-semantic-similarity
#2
JOURNAL ARTICLE
Claudia Damiano, Maarten Leemans, Johan Wagemans
Viewers use contextual information to visually explore complex scenes. Object recognition is facilitated by exploiting object-scene relations (which objects are expected in a given scene) and object-object relations (which objects are expected because of the occurrence of other objects). Semantically inconsistent objects deviate from these expectations, so they tend to capture viewers' attention (the semantic-inconsistency effect ). Some objects fit the identity of a scene more or less than others, yet semantic inconsistencies have hitherto been operationalized as binary (consistent vs...
April 23, 2024: Psychological Science
https://read.qxmd.com/read/38648137/deepmesh-differentiable-iso-surface-extraction
#3
JOURNAL ARTICLE
Benoit Guillard, Edoardo Remelli, Artem Lukoianov, Pierre Yvernay, Stephan R Richter, Timur Bagautdinov, Pierre Baque, Pascal Fua
Geometric Deep Learning has recently made striking progress with the advent of continuous deep implicit fields. They allow for detailed modeling of watertight surfaces of arbitrary topology while not relying on a 3D Euclidean grid, resulting in a learnable parameterization that is unlimited in resolution. Unfortunately, these methods are often unsuitable for applications that require an explicit mesh-based surface representation because converting an implicit field to such a representation relies on the Marching Cubes algorithm, which cannot be differentiated with respect to the underlying implicit field...
April 22, 2024: IEEE Transactions on Pattern Analysis and Machine Intelligence
https://read.qxmd.com/read/38641188/a-multi-featured-expression-recognition-model-incorporating-attention-mechanism-and-object-detection-structure-for-psychological-problem-diagnosis
#4
JOURNAL ARTICLE
Xiufeng Zhang, Bingyi Li, Guobin Qi
Expression is the main method for judging the emotional state and psychological condition of the human body, and the prediction of changes in facial expressions can effectively determine the mental health of a person, thus avoiding serious psychological or psychiatric disorders due to early negligence. From a computer vision perspective, most researchers have focused on studying facial expression analysis, and in some cases, body posture is also considered. However their performance is more limited under unconstrained natural conditions, which requires more information to be used in human emotion analysis...
April 17, 2024: Physiology & Behavior
https://read.qxmd.com/read/38637993/macaque-claustrum-pulvinar-and-putative-dorsolateral-amygdala-support-the-cross-modal-association-of-social-audio-visual-stimuli-based-on-meaning
#5
JOURNAL ARTICLE
Mathilda Froesel, Maëva Gacoin, Simon Clavagnier, Marc Hauser, Quentin Goudard, Suliann Ben Hamed
Social communication draws on several cognitive functions such as perception, emotion recognition and attention. The association of audio-visual information is essential to the processing of species-specific communication signals. In this study, we use functional magnetic resonance imaging in order to identify the subcortical areas involved in the cross-modal association of visual and auditory information based on their common social meaning. We identified three subcortical regions involved in audio-visual processing of species-specific communicative signals: the dorsolateral amygdala, the claustrum and the pulvinar...
April 18, 2024: European Journal of Neuroscience
https://read.qxmd.com/read/38637870/beyond-visual-integration-sensitivity-of-the-temporal-parietal-junction-for-objects-places-and-faces
#6
JOURNAL ARTICLE
Johannes Rennig, Christina Langenberger, Hans-Otto Karnath
One important role of the TPJ is the contribution to perception of the global gist in hierarchically organized stimuli where individual elements create a global visual percept. However, the link between clinical findings in simultanagnosia and neuroimaging in healthy subjects is missing for real-world global stimuli, like visual scenes. It is well-known that hierarchical, global stimuli activate TPJ regions and that simultanagnosia patients show deficits during the recognition of hierarchical stimuli and real-world visual scenes...
April 18, 2024: Behavioral and Brain Functions: BBF
https://read.qxmd.com/read/38636427/mediating-sequential-turn-on-and-turn-off-fluorescence-signals-for-discriminative-detection-of-ag-and-hg-2-via-readily-available-cdse-quantum-dots
#7
JOURNAL ARTICLE
Rong Wang, Zi Yi Xu, Ting Li, Nian Bing Li, Hong Qun Luo
Realizing the accurate recognition and quantification of heavy metal ions is pivotal but challenging in the environmental, biological, and physiological science fields. In this work, orange fluorescence emitting quantum dots (OQDs) have been facilely synthesized by one-step method. The participation of silver ion (Ag+ ) can evoke the unique aggregation-induced emission (AIE) of OQDs, resulting in prominent fluorescence enhancement, which is scarcely reported previously. Moreover, the Ag+ -triggered turn-on fluorescence can be continuously shut down by mercury ion (Hg2+ )...
April 13, 2024: Spectrochimica Acta. Part A, Molecular and Biomolecular Spectroscopy
https://read.qxmd.com/read/38633658/multiscale-apple-recognition-method-based-on-improved-centernet
#8
JOURNAL ARTICLE
Han Zhou
Traditional apple-picking robots are unable to detect apples in real-time in complex environments. In order to improve detection efficiency, a fast CenterNet apple recognition method for multiple apple targets in dense scenes is proposed. This method can quickly and accurately identify multiple apple targets in dense scenes. The backbone network mainly consists of resnet-44 fully convolutional network, region of interest network (RPN), and region of interest (ROI). The experimental results show that the improved YoloV5 network model has a higher recognition accuracy of 94...
April 15, 2024: Heliyon
https://read.qxmd.com/read/38626617/weakly-supervised-temporal-action-localization-with-actionness-guided-false-positive-suppression
#9
JOURNAL ARTICLE
Zhilin Li, Zilei Wang, Qinying Liu
Weakly supervised temporal action localization aims to locate the temporal boundaries of action instances in untrimmed videos using video-level labels and assign them the corresponding action category. Generally, it is solved by a pipeline called "localization-by-classification", which finds the action instances by classifying video snippets. However, since this approach optimizes the video-level classification objective, the generated activation sequences often suffer interference from class-related scenes, resulting in a large number of false positives in the prediction results...
April 15, 2024: Neural Networks: the Official Journal of the International Neural Network Society
https://read.qxmd.com/read/38625775/fast-building-instance-proxy-reconstruction-for-large-urban-scenes
#10
JOURNAL ARTICLE
Jianwei Guo, Haobo Qin, Yinchang Zhou, Xin Chen, Liangliang Nan, Hui Huang
Digitalization of large-scale urban scenes (in particular buildings) has been a long-standing open problem, which attributes to the challenges in data acquisition, such as incomplete scene coverage, lack of semantics, low efficiency, and low reliability in path planning. In this paper, we address these challenges in urban building reconstruction from aerial images, and we propose an effective workflow and a few novel algorithms for efficient 3D building instance proxy reconstruction for large urban scenes. Specifically, we propose a novel learning-based approach to instance segmentation of urban buildings from aerial images followed by a voting-based algorithm to fuse the multi-view instance information to a sparse point cloud (reconstructed using a standard Structure from Motion pipeline)...
April 16, 2024: IEEE Transactions on Pattern Analysis and Machine Intelligence
https://read.qxmd.com/read/38625774/bridging-visual-and-textual-semantics-towards-consistency-for-unbiased-scene-graph-generation
#11
JOURNAL ARTICLE
Ruonan Zhang, Gaoyun An, Yiqing Hao, Dapeng Oliver Wu
Scene Graph Generation (SGG) aims to detect visual relationships in an image. However, due to long-tailed bias, SGG is far from practical. Most methods depend heavily on the assistance of statistics co-occurrence to generate a balanced dataset, so they are dataset-specific and easily affected by noises. The fundamental cause is that SGG is simplified as a classification task instead of a reasoning task, thus the ability capturing the fine-grained details is limited and the difficulty in handling ambiguity is increased...
April 16, 2024: IEEE Transactions on Pattern Analysis and Machine Intelligence
https://read.qxmd.com/read/38616235/figure-ground-segmentation-based-on-motion-in-the-archerfish
#12
JOURNAL ARTICLE
Svetlana Volotsky, Ronen Segev
Figure-ground segmentation is a fundamental process in visual perception that involves separating visual stimuli into distinct meaningful objects and their surrounding context, thus allowing the brain to interpret and understand complex visual scenes. Mammals exhibit varying figure-ground segmentation capabilities, ranging from primates that can perform well on figure-ground segmentation tasks to rodents that perform poorly. To explore figure-ground segmentation capabilities in teleost fish, we studied how the archerfish, an expert visual hunter, performs figure-ground segmentation...
April 15, 2024: Animal Cognition
https://read.qxmd.com/read/38615047/using-a-flashlight-contingent-window-paradigm-to-investigate-visual-search-and-object-memory-in-virtual-reality-and-on-computer-screens
#13
JOURNAL ARTICLE
Julia Beitner, Jason Helbing, Erwan Joël David, Melissa Lê-Hoa Võ
A popular technique to modulate visual input during search is to use gaze-contingent windows. However, these are often rather discomforting, providing the impression of visual impairment. To counteract this, we asked participants in this study to search through illuminated as well as dark three-dimensional scenes using a more naturalistic flashlight with which they could illuminate the rooms. In a surprise incidental memory task, we tested the identities and locations of objects encountered during search. Importantly, we tested this study design in both immersive virtual reality (VR; Experiment 1) and on a desktop-computer screen (Experiment 2)...
April 13, 2024: Scientific Reports
https://read.qxmd.com/read/38610574/wi-char-a-wifi-sensing-approach-with-focus-on-both-scenes-and-restricted-data
#14
JOURNAL ARTICLE
Zhanjun Hao, Kaikai Han, Zinan Zhang, Xiaochao Dang
Significant strides have been made in the field of WiFi-based human activity recognition, yet recent wireless sensing methodologies still grapple with the reliance on copious amounts of data. When assessed in unfamiliar domains, the majority of models experience a decline in accuracy. To address this challenge, this study introduces Wi-CHAR, a novel few-shot learning-based cross-domain activity recognition system. Wi-CHAR is meticulously designed to tackle both the intricacies of specific sensing environments and pertinent data-related issues...
April 8, 2024: Sensors
https://read.qxmd.com/read/38610414/an-appearance-semantic-descriptor-with-coarse-to-fine-matching-for-robust-vpr
#15
JOURNAL ARTICLE
Jie Chen, Wenbo Li, Pengshuai Hou, Zipeng Yang, Haoyu Zhao
In recent years, semantic segmentation has made significant progress in visual place recognition (VPR) by using semantic information that is relatively invariant to appearance and viewpoint, demonstrating great potential. However, in some extreme scenarios, there may be semantic occlusion and semantic sparsity, which can lead to confusion when relying solely on semantic information for localization. Therefore, this paper proposes a novel VPR framework that employs a coarse-to-fine image matching strategy, combining semantic and appearance information to improve algorithm performance...
March 29, 2024: Sensors
https://read.qxmd.com/read/38609405/cataract-1k-dataset-for-deep-learning-assisted-analysis-of-cataract-surgery-videos
#16
JOURNAL ARTICLE
Negin Ghamsarian, Yosuf El-Shabrawi, Sahar Nasirihaghighi, Doris Putzgruber-Adamitsch, Martin Zinkernagel, Sebastian Wolf, Klaus Schoeffmann, Raphael Sznitman
In recent years, the landscape of computer-assisted interventions and post-operative surgical video analysis has been dramatically reshaped by deep-learning techniques, resulting in significant advancements in surgeons' skills, operation room management, and overall surgical outcomes. However, the progression of deep-learning-powered surgical technologies is profoundly reliant on large-scale datasets and annotations. In particular, surgical scene understanding and phase recognition stand as pivotal pillars within the realm of computer-assisted surgery and post-operative assessment of cataract surgery videos...
April 12, 2024: Scientific Data
https://read.qxmd.com/read/38607711/vote2cap-detr-decoupling-localization-and-describing-for-end-to-end-3d-dense-captioning
#17
JOURNAL ARTICLE
Sijin Chen, Hongyuan Zhu, Mingsheng Li, Xin Chen, Peng Guo, Yinjie Lei, Gang Yu, Taihao Li, Tao Chen
3D dense captioning requires a model to translate its understanding of an input 3D scene into several captions associated with different object regions. Existing methods adopt a sophisticated "detect-then-describe" pipeline, which builds explicit relation modules upon a 3D detector with numerous hand-crafted components. While these methods have achieved initial success, the cascade pipeline tends to accumulate errors because of duplicated and inaccurate box estimations and messy 3D scenes. In this paper, we first propose Vote2Cap-DETR, a simple-yet-effective transformer framework that decouples the decoding process of caption generation and object localization through parallel decoding...
April 12, 2024: IEEE Transactions on Pattern Analysis and Machine Intelligence
https://read.qxmd.com/read/38602815/anhedonia-reflects-an-encoding-deficit-for-pleasant-stimuli-in-schizophrenia-evidence-from-the-emotion-induced-memory-trade-off-eye-tracking-paradigm
#18
JOURNAL ARTICLE
Kayla M Whearty, Ivan Ruiz, Anna R Knippenberg, Gregory P Strauss
OBJECTIVE: The present study explored the hypothesis that anhedonia reflects an emotional memory impairment for pleasant stimuli, rather than diminished hedonic capacity in individuals with schizophrenia (SZ). METHOD: Participants included 30 SZ and 30 healthy controls (HCs) subjects who completed an eye-tracking emotion-induced memory trade-off task where contextually relevant pleasant, unpleasant, or neutral items were inserted into the foreground of neutral background scenes...
April 11, 2024: Neuropsychology
https://read.qxmd.com/read/38598389/perf-panoramic-neural-radiance-field-from-a-single-panorama
#19
JOURNAL ARTICLE
Guangcong Wang, Peng Wang, Zhaoxi Chen, Wenping Wang, Chen Change Loy, Ziwei Liu
Neural Radiance Field (NeRF) has achieved substantial progress in novel view synthesis given multi-view images. Recently, some works have attempted to train a NeRF from a single image with 3D priors. They mainly focus on a limited field of view with a few occlusions, which greatly limits their scalability to real-world 360-degree panoramic scenarios with large-size occlusions. In this paper, we present PERF, a 360-degree novel view synthesis framework that trains a panoramic neural radiance field from a single panorama...
April 10, 2024: IEEE Transactions on Pattern Analysis and Machine Intelligence
https://read.qxmd.com/read/38598036/on-investigating-drivers-attention-allocation-during-partially-automated-driving
#20
JOURNAL ARTICLE
Reem Jalal Eddine, Claudio Mulatti, Francesco N Biondi
The use of partially-automated systems require drivers to supervise the system functioning and resume manual control whenever necessary. Yet literature on vehicle automation show that drivers may spend more time looking away from the road when the partially-automated system is operational. In this study we answer the question of whether this pattern is a manifestation of inattentional blindness or, more dangerously, it is also accompanied by a greater attentional processing of the driving scene. Participants drove a simulated vehicle in manual or partially-automated mode...
April 10, 2024: Cognitive Research: Principles and Implications
keyword
keyword
102340
1
2
Fetch more papers »
Fetching more papers... Fetching...
Remove bar
Read by QxMD icon Read
×

Save your favorite articles in one place with a free QxMD account.

×

Search Tips

Use Boolean operators: AND/OR

diabetic AND foot
diabetes OR diabetic

Exclude a word using the 'minus' sign

Virchow -triad

Use Parentheses

water AND (cup OR glass)

Add an asterisk (*) at end of a word to include word stems

Neuro* will search for Neurology, Neuroscientist, Neurological, and so on

Use quotes to search for an exact phrase

"primary prevention of cancer"
(heart or cardiac or cardio*) AND arrest -"American Heart Association"

We want to hear from doctors like you!

Take a second to answer a survey question.