Jana Košecká

ORCID: 0000-0003-4619-3277
Publications
Citations
Views
---
Saved
---
About
Contact & Profiles
Research Areas
  • Robotics and Sensor-Based Localization
  • Advanced Image and Video Retrieval Techniques
  • Advanced Vision and Imaging
  • Advanced Neural Network Applications
  • Human Pose and Action Recognition
  • Multimodal Machine Learning Applications
  • Hand Gesture Recognition Systems
  • Robotic Path Planning Algorithms
  • 3D Surveying and Cultural Heritage
  • Optical measurement and interference techniques
  • Video Surveillance and Tracking Methods
  • Domain Adaptation and Few-Shot Learning
  • Image and Object Detection Techniques
  • Formal Methods in Verification
  • Hearing Impairment and Communication
  • Robot Manipulation and Learning
  • Natural Language Processing Techniques
  • Petri Nets in System Modeling
  • Image Enhancement Techniques
  • Indoor and Outdoor Localization Technologies
  • Adversarial Robustness in Machine Learning
  • Anomaly Detection Techniques and Applications
  • Advanced Image Processing Techniques
  • Image Retrieval and Classification Techniques
  • Muscle activation and electromyography studies

George Mason University
2015-2024

Yale University
2023

École Polytechnique Fédérale de Lausanne
2023

University of Twente
2023

Osaka University
2023

Czech Technical University in Prague
2022

Laboratoire d'Informatique de Paris-Nord
2020

Austrian Institute of Technology
2012

Sorbonne Université
2011

Institut Systèmes Intelligents et de Robotique
2011

We present a method for 3D object detection and pose estimation from single image. In contrast to current techniques that only regress the orientation of an object, our first regresses relatively stable properties using deep convolutional neural network then combines these estimates with geometric constraints provided by 2D bounding box produce complete box. The output novel hybrid discrete-continuous loss, which significantly outperforms L2 loss. second dimensions, have little variance...

10.1109/cvpr.2017.597 article EN 2017-07-01

Skillful mobile operation in three-dimensional environments is a primary topic of study Artificial Intelligence. The past two years have seen surge creative work on navigation. This output has produced plethora sometimes incompatible task definitions and evaluation protocols. To coordinate ongoing future research this area, we convened working group to empirical methodology navigation research. present document summarizes the consensus recommendations group. We discuss different problem...

10.48550/arxiv.1807.06757 preprint EN cc-by arXiv (Cornell University) 2018-01-01

In this paper we present a prototype system for image based localization in urban environments. Given database of views city street scenes tagged by GPS locations, the computes location novel query view. We first use wide-baseline matching technique on SIFT features to select closest database. Often due large change viewpoint and presence repetitive structures, percentage matches (> 50%) are not correct correspondences. The subsequent motion estimation between view reference view, is then...

10.1109/3dpvt.2006.80 article EN 2006-06-01

What is a good visual representation for navigation? We study this question in the context of semantic navigation, which problem robot finding its way through previously unseen environment to target object, e.g. go refrigerator. Instead acquiring metric map an and using planning our approach learns navigation policies on top representations that capture spatial layout contextual cues. propose use segmentation detection masks as observations obtained by state-of-the-art computer vision...

10.1109/icra.2019.8793493 article EN 2022 International Conference on Robotics and Automation (ICRA) 2019-05-01

We present a new public dataset with focus on simulating robotic vision tasks in everyday indoor environments using real imagery. The includes 20,000+ RGB-D images and 50,000+ 2D bounding boxes of object instances densely captured 9 unique scenes. train fast category detector for instance detection our data. Using the we show that, although increasingly accurate fast, state art is still severely impacted by scale, occlusion, viewing direction all which matter robotics applications. next...

10.1109/icra.2017.7989164 article EN 2017-05-01

Theoretical and analytical aspects of the visual servoing problem have not received much attention. Furthermore, estimation from vision measurements has been considered separately design control strategies. Instead addressing pose problems separately, we attempt to characterize types tasks which can be achieved using only quantities directly measurable in image, bypassing phase. We consider task navigation for a nonholonomic ground mobile base tracking an arbitrarily shaped continuous curve....

10.1109/70.768184 article EN IEEE Transactions on Robotics and Automation 1999-06-01

Detection of objects in cluttered indoor environments is one the key enabling functionalities for service robots. The best performing object detection approaches computer vision exploit deep Convolutional Neural Networks (CNN) to simultaneously detect and categorize interest scenes. Training such models typically requires large amounts annotated training data which time consuming costly obtain. In this work we explore ability using synthetically generated composite images state-of-the-art...

10.15607/rss.2017.xiii.043 article EN 2017-07-12

Multi-view stereo methods frequently fail to properly reconstruct 3D scene geometry if visible texture is sparse or the exhibits difficult self-occlusions. Time-of-Flight (ToF) depth sensors can provide information regardless of but with only limited resolution and accuracy. To find an optimal reconstruction, we propose integrated multi-view sensor fusion approach that combines from multiple color cameras ToF sensors. First, measurements are combined obtain a coarse complete model. Then,...

10.1109/iccvw.2009.5457430 article EN 2009-09-01

Surface electromyography (sEMG) has been the predominant method for sensing electrical activity a number of applications involving muscle-computer interfaces, including myoelectric control prostheses and rehabilitation robots. Ultrasound imaging mechanical deformation functional muscle compartments can overcome several limitations sEMG, inability to differentiate between deep contiguous compartments, low signal-to-noise ratio, lack robust graded signal. The objective this study was evaluate...

10.1109/tbme.2015.2498124 article EN publisher-specific-oa IEEE Transactions on Biomedical Engineering 2015-11-05

City environments often lack textured areas, contain repetitive structures, strong lighting changes and therefore are very difficult for standard 3D modeling pipelines. We present a novel unified framework creating city models which overcomes these difficulties by exploiting image segmentation cues as well presence of dominant scene orientations piecewise planar structures. Given panoramic street view sequences, we first demonstrate how to robustly estimate camera poses without need bundle...

10.1109/cvpr.2009.5206535 article EN 2009 IEEE Conference on Computer Vision and Pattern Recognition 2009-06-01

Multi-scale deep CNNs have been used successfully for problems mapping each pixel to a label, such as depth estimation and semantic segmentation. It has also shown that architectures are reusable can be multiple tasks. These networks typically trained independently task by varying the output layer(s) training objective. In this work we present new model simultaneous segmentation from single RGB image. Our approach demonstrates feasibility of parts then fine tuning full, combined on both...

10.1109/3dv.2016.69 preprint EN 2016-10-01

For applications in navigation and robotics, estimating the 3D pose of objects is as important detection. Many approaches to estimation rely on detecting or tracking parts keypoints [11, 21]. In this paper we build a recent state-of-the-art convolutional network for sliding-window detection [10] provide rough single shot, without intermediate stages initial bounding boxes. While not first system treat categorization problem, attempt combine at same level using deep learning approach. The key...

10.1109/3dv.2016.78 article EN 2016-10-01

With the increasing speeds of modern microprocessors, it has become ever more common for computer-vision algorithms to find application in real-time control tasks. In this paper, we present an analysis problem steering autonomous vehicle along a highway based on images obtained from CCD camera mounted vehicle. We explore effects changing various important system parameters like velocity, look-ahead range vision sensor, and processing delay associated with perception systems. also results...

10.1177/027836499901800502 article EN The International Journal of Robotics Research 1999-05-01

10.1023/a:1012276232049 article EN International Journal of Computer Vision 2001-01-01

This paper presents a nonparametric approach to semantic parsing using small patches and simple gradient, color location features. We learn the relevance of individual feature channels at test time locally adaptive distance metric. To further improve accuracy approach, we examine importance retrieval set used compute nearest neighbours novel descriptor retrieve better candidates. The is validated by experiments on several datasets for demonstrating superiority method compared state art approaches.

10.1109/cvpr.2013.405 article EN 2009 IEEE Conference on Computer Vision and Pattern Recognition 2013-06-01

10.1007/s11263-010-0327-9 article EN International Journal of Computer Vision 2010-03-16

Vision-based topological localization and mapping for autonomous robotic systems have received increased research interest in recent years. The need to map larger environments requires models at different levels of abstraction additional abilities deal with large amounts data efficiently. Most successful approaches appearance-based datasets typically represent locations using local image features. We study the feasibility performing these tasks urban global descriptors instead taking...

10.1109/tro.2012.2220211 article EN IEEE Transactions on Robotics 2012-11-16

Multi-scale deep CNNs have been used successfully for problems mapping each pixel to a label, such as depth estimation and semantic segmentation. It has also shown that architectures are reusable can be multiple tasks. These networks typically trained independently task by varying the output layer(s) training objective. In this work we present new model simultaneous segmentation from single RGB image. Our approach demonstrates feasibility of parts then fine tuning full, combined on both...

10.13016/m2uswx-e8bj article EN International Conference on 3D Vision 2016-04-25

Finding correspondences between images or 3D scans is at the heart of many computer vision and image retrieval applications often enabled by matching local keypoint descriptors. Various learning approaches have been applied in past to different stages pipeline, considering detection, description, metric objectives. These objectives were typically addressed separately most previous work has focused on data. This paper proposes an end-to-end framework for detection its representation...

10.1109/cvpr.2018.00210 article EN 2018-06-01

This paper presents a new multi-view RGB-D dataset of nine kitchen scenes, each containing several objects in realistic cluttered environments including subset from the BigBird dataset. The viewpoints scenes are densely sampled and annotated with bounding boxes 3D point cloud. Also, an approach for detection recognition is presented, which comprised two parts: (i) proposal generation method (ii) development baselines using AlexNet to score our proposals, trained either on crops or...

10.1109/3dv.2016.52 article EN 2016-10-01

Gestures in American Sign Language (ASL) are characterized by fast, highly articulate motion of upper body, including arm movements with complex hand shapes and facial expressions. In this work, we propose a new method for word-level sign recognition from using video. Our uses both shape cues while being robust to variations execution. We exploit the knowledge body pose, estimated an off-the-shelf pose estimator. Using as guide, pool spatio-temporal feature maps different layers 3D...

10.1109/wacv48630.2021.00347 article EN 2021-01-01
Coming Soon ...