Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Filters








1,186 Hits in 6.6 sec

Semantic Driven Multi-Camera Pedestrian Detection [article]

Alejandro López-Cifuentes, Marcos Escudero-Viñolo, Jesús Bescós, Pablo Carballeira
2022 arXiv   pre-print
Context information, obtained via semantic segmentation, is used 1) to automatically generate a common Area of Interest for the scene and all the cameras, avoiding the usual need of manually defining it  ...  This noteworthy attribute does not require ad hoc training with labelled data, expediting the deployment of the proposed method in real-world situations.  ...  Acknowledgements This study has been partially supported by the Spanish Government through its TEC2017-88169-R MobiNetVideo project.  ... 
arXiv:1812.10779v3 fatcat:r4rf4emci5bglg74k4v4djcvge

Mid-Air: A Multi-Modal Dataset for Extremely Low Altitude Drone Flights

Michael Fonder, Marc Van Droogenbroeck
2019 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)  
It contains synchronized data of multiple sensors for a total of 54 trajectories and more than 420k video frames simulated in various climate conditions.  ...  In this work, we motivate design choices, explain how the data was simulated, and present the content of the dataset.  ...  This allows to have a scenario where the visual features are moving towards the camera, and another where they are scrolling from left to right in the frame.  ... 
doi:10.1109/cvprw.2019.00081 dblp:conf/cvpr/FonderD19 fatcat:qorspe2ydnfetcrt65sgfb5wte

Exploring OpenStreetMap Availability for Driving Environment Understanding [article]

Yang Zheng, Izzat H. Izzat, John H.L. Hansen
2019 arXiv   pre-print
First, driving scenario attributes are retrieved from OSM elements, which are combined with vehicle dynamic signals for the driving event recognition.  ...  With the great achievement of artificial intelligence, vehicle technologies have advanced significantly from human centric driving towards fully automated driving.  ...  The KITTI dataset has been recorded from a moving platform (Volkswagen Passat station wagon) while driving in and around the mid-sized city of Karlsruhe, Germany.  ... 
arXiv:1903.04084v1 fatcat:tbkoaplpmvghjdx7r2ybf5xeca

Video Analytics for Business Intelligence [chapter]

Asaad Hakeem, Himaanshu Gupta, Atul Kanaujia, Tae Eun Choe, Kiran Gunda, Andrew Scanlon, Li Yu, Zhong Zhang, Peter Venetianer, Zeeshan Rasheed, Niels Haering
2012 Studies in Computational Intelligence  
The goal is to provide the reader with an overview of the state of the art approaches in the field of video analytics, and also describe the various applications where these technologies can be applied  ...  We have also presented the evaluation results of each of these technology components using in-house and other publicly available datasets.  ...  In scenarios where the image resolution is not sufficient to identify an individual, PTZ cameras can be automatically pointed at the target of interest to capture highresolution images.  ... 
doi:10.1007/978-3-642-28598-1_10 fatcat:shu47nw2unh73fsfpx46uvfdgi

A real-time robotic model of human reference resolution using visual constraints

Matthias Scheutz, Kathleen Eberhard, Virgil Andronache
2004 Connection science  
Results from experiments with the model confirm the viability of the algorithm to process semantic interpretations, in particular, reference incrementally, as demonstrated to be employed by humans.  ...  phase, which typically involves the construction of parse trees.  ...  its basic properties (e.g., color and shape) and automatically tracks the object (e.g., if they should move or the camera should move).  ... 
doi:10.1080/09540090412331314803 fatcat:e5swghydlfg5vfr5fr4stf2c5i

Automatic Association of Chats and Video Tracks for Activity Learning and Recognition in Aerial Video Surveillance

Riad Hammoud, Cem Sahin, Erik Blasch, Bradley Rhodes, Tao Wang
2014 Sensors  
/false alarm results due to the complexity of the scenario.  ...  tracking in challenging scenarios.  ...  The ideas and opinions expressed here are not official policies of the United States Air Force.  ... 
doi:10.3390/s141019843 pmid:25340453 pmcid:PMC4239870 fatcat:ony3ylej4nhzxbnap2zide3kwi

Multi-source Multi-modal Activity Recognition in Aerial Video Surveillance

Riad I. Hammoud, Cem S. Sahin, Erik P. Blasch, Bradley J. Rhodes
2014 2014 IEEE Conference on Computer Vision and Pattern Recognition Workshops  
In the context of this research, we deal with two unsynchronized data sources collected in real-world operating scenarios: full-motion videos (FMV) and analyst call-outs (ACO) in the form of chat messages  ...  Recognizing activities in wide aerial/overhead imagery remains a challenging problem due in part to low-resolution video and cluttered scenes with a large number of moving objects.  ...  The ideas and opinions expressed here are not official policies of the United States Air Force.  ... 
doi:10.1109/cvprw.2014.44 dblp:conf/cvpr/HammoudSBR14 fatcat:p7pqnfbzefdplpm5eztcgk2lge

Unifying Terrain Awareness for the Visually Impaired through Real-Time Semantic Segmentation

Kailun Yang, Kaiwei Wang, Luis Bergasa, Eduardo Romera, Weijian Hu, Dongming Sun, Junwei Sun, Ruiqi Cheng, Tianxue Chen, Elena López
2018 Sensors  
In this paper, we put forward seizing pixel-wise semantic segmentation to cover navigation-related perception needs in a unified way.  ...  The core of our unification proposal is a deep architecture, aimed at attaining efficient semantic understanding.  ...  Acknowledgments: We would like to thank the anonymous reviewers for the insightful comments and valuable suggestions. Conflicts of Interest: The authors declare no conflict of interest.  ... 
doi:10.3390/s18051506 pmid:29748508 pmcid:PMC5982125 fatcat:meowkrc67bhedigybnpl7snowq

Interpretation of complex situations in a semantic-based surveillance framework

Carles Fernández, Pau Baiget, Xavier Roca, Jordi Gonzàlez
2008 Signal processing. Image communication  
communication with casual end users in multiple languages.  ...  The integration of cognitive capabilities in computer vision systems requires both to enable high semantic expressiveness and to deal with high computational costs as large amounts of data are involved  ...  Jordi Gonzà lez also acknowledges the support of a Juan de la Cierva Postdoctoral fellowship from the Spanish MEC.  ... 
doi:10.1016/j.image.2008.04.015 fatcat:7hun6y3evndodcfcqvic2zszvy

A Fast and Accurate Spatial Target Snapping Method for 3D Scene Modeling and Mapping in Mobile Augmented Reality

Kejia Huang, Chenliang Wang, Runying Liu, Guoxiong Chen
2022 ISPRS International Journal of Geo-Information  
Finally, the algorithm is experimented with by an AR modeling system, including the evaluation of snapping efficiency and snapping accuracy.  ...  The snapping efficiency of the algorithm proposed in this paper is 1.6 times higher than that of the traditional algorithm on average, while the data acquisition accuracy based on the algorithm in this  ...  When the user moves the object near the constraint, the object is automatically aligned precisely with the virtual constraint according to specific constraint rules that include extension lines, horizontal  ... 
doi:10.3390/ijgi11010069 fatcat:g4pyjgk5lbemrbgz6j73f7djxm

Geometric Edge Description and Classification in Point Cloud Data with Application to 3D Object Recognition

Troels Bo Jørgensen, Anders Glent Buch, Dirk Kraft
2015 Proceedings of the 10th International Conference on Computer Vision Theory and Applications  
We present an integrated system for the recognition, pose estimation and simultaneous tracking of multiple objects in 3D scenes.  ...  We evaluate the proposed method on real scenarios by comparing tracked outputs to ground truth trajectories and we compare the results to Iterative Closest Point and Particle Filter based trackers.  ...  Bottom left: Two of the cameras recording the scene. Figure 6 : 6 Results for the simple sequences. The object is moved along a circular arc during manipulation.  ... 
doi:10.5220/0005196703330340 dblp:conf/visapp/JorgensenBK15 fatcat:drh737k3yndq5htfwmledhzmy4

An Online Vision System for Understanding Complex Assembly Tasks

Thiusius Rajeeth Savarimuthu, Jeremie Papon, Anders Glent Buch, Eren Erdal Aksoy, Wail Mustafa, Florentin Wörgötter, Norbert Krüger
2015 Proceedings of the 10th International Conference on Computer Vision Theory and Applications  
We present an integrated system for the recognition, pose estimation and simultaneous tracking of multiple objects in 3D scenes.  ...  We evaluate the proposed method on real scenarios by comparing tracked outputs to ground truth trajectories and we compare the results to Iterative Closest Point and Particle Filter based trackers.  ...  Bottom left: Two of the cameras recording the scene. Figure 6 : 6 Results for the simple sequences. The object is moved along a circular arc during manipulation.  ... 
doi:10.5220/0005260804540461 dblp:conf/visapp/SavarimuthuPBAM15 fatcat:di4mawh3zzav5japg7jo2n7br4

Multi-Modal Detection and Mapping of Static and Dynamic Obstacles in Agriculture for Process Evaluation

Timo Korthals, Mikkel Kragh, Peter Christiansen, Henrik Karstoft, Rasmus N. Jørgensen, Ulrich Rückert
2018 Frontiers in Robotics and AI  
Detection information is mapped globally into semantical occupancy grid maps and fused across all sensors with late fusion, resulting in accurate traversability assessment and semantical mapping of process-relevant  ...  in the planned path.  ...  This research is sponsored by the Innovation Fund Denmark as part of the project "SAFE-Safer Autonomous Farming Equipment" (Project No. 16-2014-0).  ... 
doi:10.3389/frobt.2018.00028 pmid:33500915 pmcid:PMC7806069 fatcat:hp4krcsw3fazdeugmzypcnnwry

TartanAir: A Dataset to Push the Limits of Visual SLAM [article]

Wenshan Wang, Delong Zhu, Xiangwei Wang, Yaoyu Hu, Yuheng Qiu, Chen Wang, Yafei Hu, Ashish Kapoor, Sebastian Scherer
2020 arXiv   pre-print
The data is collected in photo-realistic simulation environments with the presence of moving objects, changing light and various weather conditions.  ...  By collecting data in simulations, we are able to obtain multi-modal sensor data and precise ground truth labels such as the stereo RGB image, depth image, segmentation, optical flow, camera poses, and  ...  We thank Microsoft for the technical support of AirSim and Azure services.  ... 
arXiv:2003.14338v2 fatcat:g42d4koxk5cijmr3325ufwtouq

Representations and Benchmarking of Modern Visual SLAM Systems

Yuchen Cao, Lan Hu, Laurent Kneip
2020 Sensors  
We discuss the structure of the representations and optimisation problems involved in Spatial AI, and propose new synthetic datasets that, for the first time, include accurate ground truth information  ...  and semantic annotations of the recovered 3D model.  ...  The ground truth semantic segmentations are again generated in camera views that coincide with the left stereo view.  ... 
doi:10.3390/s20092572 pmid:32366018 pmcid:PMC7248763 fatcat:myrm7wcqtfghxgptlnbphhk3p4
« Previous Showing results 1 — 15 out of 1,186 results