Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Adam Schmidt is active.

Publication


Featured researches published by Adam Schmidt.


international conference on computer vision | 2010

An evaluation of image feature detectors and descriptors for robot navigation

Adam Schmidt; Marek Kraft; Andrzej J. Kasinski

The detection and matching of feature points is crucial in many computer vision systems. Successful establishing of points correspondences between concurrent frames is important in such tasks as visual odometry, structure from motion or simultaneous localization and mapping. This paper compares of the performance of the well established, single scale detectors and descriptors and the increasingly popular, multiscale approaches.


Pattern Analysis and Applications | 2010

The architecture and performance of the face and eyes detection system based on the Haar cascade classifiers

Andrzej J. Kasinski; Adam Schmidt

The precise face and eyes detection is essential in many human–machine interface systems. Therefore, it is necessary to develop a reliable and efficient object detection method. In this paper we present the architecture of a hierarchical face and eyes detection system using the Haar cascade classifiers (HCC) augmented with some simple knowledge-based rules. The influence of the training procedure on the performance of the particular HCCs has been investigated. Additionally, we compared the efficiency of other authors’ face and eyes HCCs with the efficiency of those trained by us. By applying the proposed system to the set of 10,000 test images we were able to properly detect and precisely localize 94% of the eyes.


international conference on computer vision | 2010

The visual SLAM system for a hexapod robot

Adam Schmidt; Andrzej J. Kasinski

The precise localization of the mobile robot plays a vital role in the autonomous operation of the mobile robot. The vision based simultaneous localization and mapping (SLAM) is a widely known technique for tracking the movement of the camera in the unknown environment. This paper presents a robots movement model which is based on the reference trajectory of the robot. The proposed model was compared with the state-of-the-art model used in the successful MonoSLAM system[8] and provided good results.


advanced concepts for intelligent vision systems | 2013

An Indoor RGB-D Dataset for the Evaluation of Robot Navigation Algorithms

Adam Schmidt; Michał Fularz; Marek Kraft; Andrzej J. Kasinski; Michał Nowicki

The paper presents a RGB-D dataset for development and evaluation of mobile robot navigation systems. The dataset was registered using a WiFiBot robot equipped with a Kinect sensor. Unlike the presently available datasets, the environment was specifically designed for the registration with the Kinect sensor. Moreover, it was ensured that the registered data is synchronized with the ground truth position of the robot. The presented dataset will be made publicly available for research purposes.


computer recognition systems | 2007

The Performance of the Haar Cascade Classifiers Applied to the Face and Eyes Detection

Adam Schmidt; Andrzej J. Kasinski

Recently we have presented the hierarchical face and eye detection system based on Haar Cascade Classifiers. In this paper we focus on the optimization of detectors training. Moreover, we compare the performance of Lienhart’s face detectors [1] and Castrillon-Santana’s eyes detectors [2] with those which have been trained by us.


International Journal of Advanced Robotic Systems | 2014

Calibration of the Multi-camera Registration System for Visual Navigation Benchmarking

Adam Schmidt; Andrzej J. Kasinski; Marek Kraft; Michał Fularz; Zuzanna Domagała

This paper presents the complete calibration procedure of a multi-camera system for mobile robot motion registration. Optimization-based, purely visual methods for the estimation of the relative poses of the motion registration system cameras, as well as the relative poses of the cameras and markers placed on the mobile robot were proposed. The introduced methods were applied to the calibration of the system and the quality of the obtained results was evaluated. The obtained results compare favourably with the state of the art solutions, allowing the use of the considered motion registration system for the accurate reconstruction of the mobile robot trajectory and to register new datasets suitable for the benchmarking of indoor, visual-based navigation algorithms.


International Journal of Advanced Robotic Systems | 2015

A High-performance FPGA-based Image Feature Detector and Matcher Based on the FAST and BRIEF Algorithms

Michał Fularz; Marek Kraft; Adam Schmidt; Andrzej J. Kasinski

Image feature detection and matching is a fundamental operation in image processing. As the detected and matched features are used as input data for high-level computer vision algorithms, the matching accuracy directly influences the quality of the results of the whole computer vision system. Moreover, as the algorithms are frequently used as a part of a real-time processing pipeline, the speed at which the input image data are handled is also a concern. The paper proposes an embedded system architecture for feature detection and matching. The architecture implements the FAST feature detector and the BRIEF feature descriptor and is capable of establishing key point correspondences in the input image data stream coming from either an external sensor or memory at a speed of hundreds of frames per second, so that it can cope with most demanding applications. Moreover, the proposed design is highly flexible and configurable, and facilitates the trade-off between the processing speed and programmable logic resource utilization. All the designed hardware blocks are designed to use standard, widely adopted hardware interfaces based on the AMBA AXI4 interface protocol and are connected using an underlying direct memory access (DMA) architecture, enabling bottleneck-free inter-component data transfers.


computer recognition systems | 2013

The Classification of the Terrain by a Hexapod Robot

Adam Schmidt; Krzysztof Walas

This paper presents a new approach to the terrain classification by a hexapod robot using the tactile information. The data was acquired using the force/torque sensor mounted on the walking robot foot. Two types of classifiers were used and compared: the Normal Bayes Classifier (NBC) and the Classification And Regression Tree (CART). The article comprises the description of the experimental setup followed by the presentation of feature selection process and the comparison of the two classifiers’ accuracy. The classification system presented in the article allows the walking robot to recognize the type of the terrain on which it is currently walking on with over 90% accuracy.


machine vision applications | 2017

Toward evaluation of visual navigation algorithms on RGB-D data from the first- and second-generation Kinect

Marek Kraft; Michał Nowicki; Adam Schmidt; Michał Fularz; Piotr Skrzypczyński

Although the introduction of commercial RGB-D sensors has enabled significant progress in the visual navigation methods for mobile robots, the structured-light-based sensors, like Microsoft Kinect and Asus Xtion Pro Live, have some important limitations with respect to their range, field of view, and depth measurements accuracy. The recent introduction of the second- generation Kinect, which is based on the time-of-flight measurement principle, brought to the robotics and computer vision researchers a sensor that overcomes some of these limitations. However, as the new Kinect is, just like the older one, intended for computer games and human motion capture rather than for navigation, it is unclear how much the navigation methods, such as visual odometry and SLAM, can benefit from the improved parameters. While there are many publicly available RGB-D data sets, only few of them provide ground truth information necessary for evaluating navigation methods, and to the best of our knowledge, none of them contains sequences registered with the new version of Kinect. Therefore, this paper describes a new RGB-D data set, which is a first attempt to systematically evaluate the indoor navigation algorithms on data from two different sensors in the same environment and along the same trajectories. This data set contains synchronized RGB-D frames from both sensors and the appropriate ground truth from an external motion capture system based on distributed cameras. We describe in details the data registration procedure and then evaluate our RGB-D visual odometry algorithm on the obtained sequences, investigating how the specific properties and limitations of both sensors influence the performance of this navigation method.


Progress in Automation, Robotics and Measuring Techniques | 2015

The Architecture of an Embedded Smart Camera for Intelligent Inspection and Surveillance

Michał Fularz; Marek Kraft; Adam Schmidt; Andrzej J. Kasinski

Real time video surveillance and inspection is complex task, requiring processing large amount of image data. Performing this task in each node of a multi-camera system requires high performance and power efficient architecture of the smart camera. Such solution, based on a Xilinx Zynq heterogeneous FPGA (Field Programmable Logic Array) is presented in this paper. The proposed architecture is a general foundation, which allows easy and flexible prototyping and implementation of a range of image and video processing algorithms. Two example algorithm implementations using the described architecture are presented for illustration – moving object detection and feature points detection, description and matching.

Collaboration


Dive into the Adam Schmidt's collaboration.

Top Co-Authors

Avatar

Marek Kraft

Poznań University of Technology

View shared research outputs
Top Co-Authors

Avatar

Andrzej J. Kasinski

Poznań University of Technology

View shared research outputs
Top Co-Authors

Avatar

Michał Fularz

Poznań University of Technology

View shared research outputs
Top Co-Authors

Avatar

Zuzanna Domagała

Poznań University of Technology

View shared research outputs
Top Co-Authors

Avatar

Michał Nowicki

Poznań University of Technology

View shared research outputs
Top Co-Authors

Avatar

Piotr Skrzypczyński

Poznań University of Technology

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Krzysztof Walas

Poznań University of Technology

View shared research outputs
Researchain Logo
Decentralizing Knowledge