Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Christian Eggert is active.

Publication


Featured researches published by Christian Eggert.


international conference on image processing | 2014

Improving VLAD: Hierarchical coding and a refined local coordinate system

Christian Eggert; Stefan Romberg; Rainer Lienhart

The enormous growth of image databases calls for new techniques for fast and effective image search that scales with millions of images. Most importantly, the setting requires a compact but also descriptive image signature. Recently, the vector of aggregated local descriptors (VLAD) [1] has received much attention in large-scale image retrieval. In this paper we present two modifications for VLAD which improve the retrieval performance of the signature.


acm multimedia | 2015

On the Benefit of Synthetic Data for Company Logo Detection

Christian Eggert; Anton Winschel; Rainer Lienhart

In this paper we explore the benefits of synthetically generated data for the task of company logo detection with deep-learned features in the absence of a large training set. We use pre-trained deep convolutional neural networks for feature extraction and use a set of support vector machines for classifying those features. In order to generate sufficient training examples we synthesize artificial training images. Using a bootstrapping process, we iteratively add new synthesized examples from an unlabeled dataset to the training set. Using this setup we are able to obtain a performance which is close to the performance of the full training set.


international conference on multimedia and expo | 2017

A closer look: Small object detection in faster R-CNN

Christian Eggert; Stephan Brehm; Anton Winschel; Dan Zecha; Rainer Lienhart

Faster R-CNN is a well-known approach for object detection which combines the generation of region proposals and their classification into a single pipeline. In this paper we apply Faster R-CNN to the task of company logo detection. Motivated by the weak performance of Faster R-CNN on small object instances, we perform a detailed examination of both the proposal and the classification stage, examining their behavior for a wide range of object sizes. Additionally, we look at the influence of feature map resolution on the performance of those stages. We introduce an improved scheme for generating anchor proposals and propose a modification to Faster R-CNN which leverages higher-resolution feature maps for small objects. We evaluate our approach on the Flicker data set improving the detection performance on small object instances.


electronic imaging | 2017

Pose Estimation for Deriving Kinematic Parameters of Competitive Swimmers

Dan Zecha; Christian Eggert; Rainer Lienhart

In the field of competitive swimming a quantitative evaluation of kinematic parameters is a valuable tool for coaches but also a labor intensive task. We present a system which is able to automate the extraction of many kinematic parameters such as stroke frequency, kick rates and stroke-specific intra-cyclic parameters from video footage of an athlete. While this task can in principle be solved by human pose estimation, the problem is exacerbated by permanently changing self-occlusion and severe noise caused by air bubbles, splashes, light reflection and light refraction. Current approaches for pose estimation are unable to provide the necessary localization precision under these conditions in order to enable accurate estimates of all desired kinematic parameters. In this paper we reduce the problem of kinematic parameter derivation to detecting key frames with a deep neural network human pose estimator. We show that we can correctly detect key frames with a precision which is on par with the human annotation performance. From the correctly located key frames, aforementioned parameters can be successfully inferred. Motivation In this paper, we consider a real-world computer vision application that assists a competitive athlete in assessing and improving his or her performance by taking advantage of the proposed pose estimation system. In the field of competitive swimming a quantitative evaluation of kinematic parameters is both a valuable tool for trainers as well as a labor intensive task. The training scenario is limited to a competitive swimmer swimming in a special pool: a swimming channel (see Figure 1). The water in this pool can be artificially accelerated to constantly flow from one end of the pool to the other. The swimmer performs regular swimming motions while being filmed through a glass wall by a single camera. After the training session, the video footage is evaluated by trainers and athletes. A qualitative evaluation is usually supplemented by a quantitative analysis, where the video footage is assessed and annotated frame by frame to extract points of interest like joint positions, body part angles relative to the image plane, and other variables of interest. From theses annotations, kinematic parameters can easily be derived and used for rating the efficiency of an athlete’s swimming style and giving recommendations for pose adjustments, which finally can lead to a significant performance improvement. The whole task of manually performing a quantitative analyses is extremely time consuming and therefore performed only for very few athletes nowadays. We present a system which is able to automate the extraction of many kinematic parameters such as stroke frequency, kick rates and stroke-specific intra-cyclic parameters from video footage of an athlete. It reduces the problem of kinematic parameter derivaFigure 1. A female swimmer in a swimming channel. The proposed system continuously and reliably detects poses from which kinematic parameters are


international conference on pattern recognition | 2016

Saliency-guided selective magnification for company logo detection

Christian Eggert; Anton Winschel; Dan Zecha; Rainer Lienhart

Fast R-CNN is a well-known approach to object detection which is generally reported to be robust to scale changes. In this paper we examine the influence of scale within the detection pipeline in the case of company logo detection. We demonstrate that Fast R-CNN encounters problems when handling objects which are significantly smaller than the receptive field of the utilized network. In order to overcome these difficulties, we propose a saliency-guided multiscale approach that does not rely on building a full image pyramid. We use the feature representation computed by Fast R-CNN to directly classify large objects while at the same time predicting salient regions which contain small objects with high probability. Only selected regions are magnified and a new feature representation for these enlarged regions is calculated. Feature representations from both scales are used for classification, improving the detection quality of small objects while keeping the computational overhead low. Compared to a naive magnification strategy we are able to retain 79% of the performance gain while only spending 36% of the computation time.


international conference on multimedia retrieval | 2018

Visual Question Answering With a Hybrid Convolution Recurrent Model

Philipp Harzig; Christian Eggert; Rainer Lienhart

Visual Question Answering (VQA) is a relatively new task, which tries to infer answer sentences for an input image coupled with a corresponding question. Instead of dynamically generating answers, they are usually inferred by finding the most probable answer from a fixed set of possible answers. Previous work did not address the problem of finding all possible answers, but only modeled the answering part of VQA as a classification task. To tackle this problem, we infer answer sentences by using a Long Short-Term Memory (LSTM) network that allows us to dynamically generate answers for (image, question) pairs. In a series of experiments, we discover an end-to-end Deep Neural Network structure, which allows us to dynamically answer questions referring to a given input image by using an LSTM decoder network. With this approach, we are able to generate both less common answers, which are not considered by classification models, and more complex answers with the appearance of datasets containing answers that consist of more than three words.


Proceedings of the 1st International Workshop on Multimedia Content Analysis in Sports - MMSports'18 | 2018

A Convolutional Sequence to Sequence Model for Multimodal Dynamics Prediction in Ski Jumps

Dan Zecha; Christian Eggert; Moritz Einfalt; Stephan Brehm; Rainer Lienhart

A convolutional sequence to sequence model for predicting the jump forces of ski jumpers directly from pose estimates is presented. We collect the footage of multiple, unregistered cameras together with the output of force measurement plates and present a spatiotemporal calibration procedure for all modalities which is merely based on the athletes pose estimates. The synchronized data is used to train a fully convolutional sequence to sequence network for predicting jump forces directly from the human pose. We demonstrate that the best performing networks produce a mean squared error of 0.062 on normalized force time series while being able to identify the moment of maximal force occurrence in the original video at 55% recall within +- 2 frames around the ground truth.


international conference on multimedia retrieval | 2017

Improving Small Object Proposals for Company Logo Detection

Christian Eggert; Dan Zecha; Stephan Brehm; Rainer Lienhart

Many modern approaches for object detection are two-staged pipelines. The first stage identifies regions of interest which are then classified in the second stage. Faster R-CNN is such an approach for object detection which combines both stages into a single pipeline. In this paper we apply Faster R-CNN to the task of company logo detection. Motivated by its weak performance on small object instances, we examine in detail both the proposal and the classification stage with respect to a wide range of object sizes. We investigate the influence of feature map resolution on the performance of those stages. Based on theoretical considerations, we introduce an improved scheme for generating anchor proposals and propose a modification to Faster R-CNN which leverages higher-resolution feature maps for small objects. We evaluate our approach on the FlickrLogos dataset improving the RPN performance from 0.52 to 0.71 (MABO) and the detection performance from 0.52 to


international conference on document analysis and recognition | 2015

Fisher vector encoding of micro color features for (real world) jigsaw puzzles

Fabian Richter; Christian Eggert; Rainer Lienhart

0.67


computer vision and pattern recognition | 2018

Kinematic Pose Rectification for Performance Analysis and Retrieval in Sports

Dan Zecha; Moritz Einfalt; Christian Eggert; Rainer Lienhart

(mAP).

Collaboration


Dive into the Christian Eggert's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar

Dan Zecha

University of Augsburg

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge