A copy of this work was available on the public web and has been preserved in the Wayback Machine. The capture dates from 2022; you can also visit the original URL.
The file type is application/pdf
.
Filters
Semantic Driven Multi-Camera Pedestrian Detection
[article]
2022
arXiv
pre-print
Context information, obtained via semantic segmentation, is used 1) to automatically generate a common Area of Interest for the scene and all the cameras, avoiding the usual need of manually defining it ...
This noteworthy attribute does not require ad hoc training with labelled data, expediting the deployment of the proposed method in real-world situations. ...
Acknowledgements This study has been partially supported by the Spanish Government through its TEC2017-88169-R MobiNetVideo project. ...
arXiv:1812.10779v3
fatcat:r4rf4emci5bglg74k4v4djcvge
Mid-Air: A Multi-Modal Dataset for Extremely Low Altitude Drone Flights
2019
2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)
It contains synchronized data of multiple sensors for a total of 54 trajectories and more than 420k video frames simulated in various climate conditions. ...
In this work, we motivate design choices, explain how the data was simulated, and present the content of the dataset. ...
This allows to have a scenario where the visual features are moving towards the camera, and another where they are scrolling from left to right in the frame. ...
doi:10.1109/cvprw.2019.00081
dblp:conf/cvpr/FonderD19
fatcat:qorspe2ydnfetcrt65sgfb5wte
Exploring OpenStreetMap Availability for Driving Environment Understanding
[article]
2019
arXiv
pre-print
First, driving scenario attributes are retrieved from OSM elements, which are combined with vehicle dynamic signals for the driving event recognition. ...
With the great achievement of artificial intelligence, vehicle technologies have advanced significantly from human centric driving towards fully automated driving. ...
The KITTI dataset has been recorded from a moving platform (Volkswagen Passat station wagon) while driving in and around the mid-sized city of Karlsruhe, Germany. ...
arXiv:1903.04084v1
fatcat:tbkoaplpmvghjdx7r2ybf5xeca
Video Analytics for Business Intelligence
[chapter]
2012
Studies in Computational Intelligence
The goal is to provide the reader with an overview of the state of the art approaches in the field of video analytics, and also describe the various applications where these technologies can be applied ...
We have also presented the evaluation results of each of these technology components using in-house and other publicly available datasets. ...
In scenarios where the image resolution is not sufficient to identify an individual, PTZ cameras can be automatically pointed at the target of interest to capture highresolution images. ...
doi:10.1007/978-3-642-28598-1_10
fatcat:shu47nw2unh73fsfpx46uvfdgi
A real-time robotic model of human reference resolution using visual constraints
2004
Connection science
Results from experiments with the model confirm the viability of the algorithm to process semantic interpretations, in particular, reference incrementally, as demonstrated to be employed by humans. ...
phase, which typically involves the construction of parse trees. ...
its basic properties (e.g., color and shape) and automatically tracks the object (e.g., if they should move or the camera should move). ...
doi:10.1080/09540090412331314803
fatcat:e5swghydlfg5vfr5fr4stf2c5i
Automatic Association of Chats and Video Tracks for Activity Learning and Recognition in Aerial Video Surveillance
2014
Sensors
/false alarm results due to the complexity of the scenario. ...
tracking in challenging scenarios. ...
The ideas and opinions expressed here are not official policies of the United States Air Force. ...
doi:10.3390/s141019843
pmid:25340453
pmcid:PMC4239870
fatcat:ony3ylej4nhzxbnap2zide3kwi
Multi-source Multi-modal Activity Recognition in Aerial Video Surveillance
2014
2014 IEEE Conference on Computer Vision and Pattern Recognition Workshops
In the context of this research, we deal with two unsynchronized data sources collected in real-world operating scenarios: full-motion videos (FMV) and analyst call-outs (ACO) in the form of chat messages ...
Recognizing activities in wide aerial/overhead imagery remains a challenging problem due in part to low-resolution video and cluttered scenes with a large number of moving objects. ...
The ideas and opinions expressed here are not official policies of the United States Air Force. ...
doi:10.1109/cvprw.2014.44
dblp:conf/cvpr/HammoudSBR14
fatcat:p7pqnfbzefdplpm5eztcgk2lge
Unifying Terrain Awareness for the Visually Impaired through Real-Time Semantic Segmentation
2018
Sensors
In this paper, we put forward seizing pixel-wise semantic segmentation to cover navigation-related perception needs in a unified way. ...
The core of our unification proposal is a deep architecture, aimed at attaining efficient semantic understanding. ...
Acknowledgments: We would like to thank the anonymous reviewers for the insightful comments and valuable suggestions.
Conflicts of Interest: The authors declare no conflict of interest. ...
doi:10.3390/s18051506
pmid:29748508
pmcid:PMC5982125
fatcat:meowkrc67bhedigybnpl7snowq
Interpretation of complex situations in a semantic-based surveillance framework
2008
Signal processing. Image communication
communication with casual end users in multiple languages. ...
The integration of cognitive capabilities in computer vision systems requires both to enable high semantic expressiveness and to deal with high computational costs as large amounts of data are involved ...
Jordi Gonzà lez also acknowledges the support of a Juan de la Cierva Postdoctoral fellowship from the Spanish MEC. ...
doi:10.1016/j.image.2008.04.015
fatcat:7hun6y3evndodcfcqvic2zszvy
A Fast and Accurate Spatial Target Snapping Method for 3D Scene Modeling and Mapping in Mobile Augmented Reality
2022
ISPRS International Journal of Geo-Information
Finally, the algorithm is experimented with by an AR modeling system, including the evaluation of snapping efficiency and snapping accuracy. ...
The snapping efficiency of the algorithm proposed in this paper is 1.6 times higher than that of the traditional algorithm on average, while the data acquisition accuracy based on the algorithm in this ...
When the user moves the object near the constraint, the object is automatically aligned precisely with the virtual constraint according to specific constraint rules that include extension lines, horizontal ...
doi:10.3390/ijgi11010069
fatcat:g4pyjgk5lbemrbgz6j73f7djxm
Geometric Edge Description and Classification in Point Cloud Data with Application to 3D Object Recognition
2015
Proceedings of the 10th International Conference on Computer Vision Theory and Applications
We present an integrated system for the recognition, pose estimation and simultaneous tracking of multiple objects in 3D scenes. ...
We evaluate the proposed method on real scenarios by comparing tracked outputs to ground truth trajectories and we compare the results to Iterative Closest Point and Particle Filter based trackers. ...
Bottom left: Two of the cameras recording the scene.
Figure 6 : 6 Results for the simple sequences. The object is moved along a circular arc during manipulation. ...
doi:10.5220/0005196703330340
dblp:conf/visapp/JorgensenBK15
fatcat:drh737k3yndq5htfwmledhzmy4
An Online Vision System for Understanding Complex Assembly Tasks
2015
Proceedings of the 10th International Conference on Computer Vision Theory and Applications
We present an integrated system for the recognition, pose estimation and simultaneous tracking of multiple objects in 3D scenes. ...
We evaluate the proposed method on real scenarios by comparing tracked outputs to ground truth trajectories and we compare the results to Iterative Closest Point and Particle Filter based trackers. ...
Bottom left: Two of the cameras recording the scene.
Figure 6 : 6 Results for the simple sequences. The object is moved along a circular arc during manipulation. ...
doi:10.5220/0005260804540461
dblp:conf/visapp/SavarimuthuPBAM15
fatcat:di4mawh3zzav5japg7jo2n7br4
Multi-Modal Detection and Mapping of Static and Dynamic Obstacles in Agriculture for Process Evaluation
2018
Frontiers in Robotics and AI
Detection information is mapped globally into semantical occupancy grid maps and fused across all sensors with late fusion, resulting in accurate traversability assessment and semantical mapping of process-relevant ...
in the planned path. ...
This research is sponsored by the Innovation Fund Denmark as part of the project "SAFE-Safer Autonomous Farming Equipment" (Project No. 16-2014-0). ...
doi:10.3389/frobt.2018.00028
pmid:33500915
pmcid:PMC7806069
fatcat:hp4krcsw3fazdeugmzypcnnwry
TartanAir: A Dataset to Push the Limits of Visual SLAM
[article]
2020
arXiv
pre-print
The data is collected in photo-realistic simulation environments with the presence of moving objects, changing light and various weather conditions. ...
By collecting data in simulations, we are able to obtain multi-modal sensor data and precise ground truth labels such as the stereo RGB image, depth image, segmentation, optical flow, camera poses, and ...
We thank Microsoft for the technical support of AirSim and Azure services. ...
arXiv:2003.14338v2
fatcat:g42d4koxk5cijmr3325ufwtouq
Representations and Benchmarking of Modern Visual SLAM Systems
2020
Sensors
We discuss the structure of the representations and optimisation problems involved in Spatial AI, and propose new synthetic datasets that, for the first time, include accurate ground truth information ...
and semantic annotations of the recovered 3D model. ...
The ground truth semantic segmentations are again generated in camera views that coincide with
the left stereo view. ...
doi:10.3390/s20092572
pmid:32366018
pmcid:PMC7248763
fatcat:myrm7wcqtfghxgptlnbphhk3p4
« Previous
Showing results 1 — 15 out of 1,186 results