Willem-Jan Vriend
Kapteyn Astronomical Institute
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Willem-Jan Vriend.
The Astrophysical Journal | 2004
F. Kemper; Willem-Jan Vriend; Aggm Tielens
Infrared spectroscopy provides a direct handle on the composition and structure of interstellar dust. We have studied the dust along the line of sight toward the Galactic center using Short Wavelength Spectrometer data obtained with the Infrared Space Observatory (ISO). We focused on the wavelength region from 8 to 13 μm, which is dominated by the strong silicate absorption feature. Using the absorption profiles observed toward Galactic center sources 3 and 4, which are C-rich Wolf-Rayet Stars, as reference objects, we are able to disentangle the interstellar silicate absorption and the silicate emission intrinsic to the source, toward Sgr A* and derive a very accurate profile for the intrinsic 9.7 μm band. The interstellar absorption band is smooth and featureless and is well reproduced using a mixture of 15.1% amorphous pyroxene and 84.9% of amorphous olivine by mass, all in spherical submicron-sized grains. There is no direct evidence for substructure due to interstellar crystalline silicates. By minimizing χ2 of spectral fits to the absorption feature, we are able to determine an upper limit to the degree of crystallinity of silicates in the diffuse interstellar medium (ISM) and conclude that the crystalline fraction of the interstellar silicates is 0.2% ± 0.2% by mass. This is much lower than the degree of crystallinity observed in silicates in the circumstellar environment of evolved stars, the main contributors of dust to the ISM. There are two possible explanations for this discrepancy. First, an amorphization process occurs in the ISM on a timescale significantly shorter than the destruction timescale, possibly caused by particle bombardment by heavyweight ions. Second, we consider the possibility that the crystalline silicates in stellar ejecta are diluted by an additional source of amorphous silicates, in particular supernovae. We also compare our results with a study on silicate presolar grains found in interplanetary dust particles.
Journal of Grid Computing | 2010
Kornelis Begeman; Andrey Belikov; Danny Boxhoorn; Fokke Dijkstra; E Valentijn; Willem-Jan Vriend; Z. Zhao
This paper reports the integration of the astronomical Grid solution realised in the Astro-WISE information system with the EGEE Grid and the porting of Astro-WISE applications on EGEE. We review the architecture of the Astro-WISE Grid, define the problems for the integration of the Grid infrastructures and our solution to these problems. We give examples of applications running on Astro-WISE and EGEE and review future development of the merged system.
grid computing | 2011
Kornelis Begeman; A. Belikov; Danny Boxhoorn; Fokke Dijkstra; H. A. Holties; Z. Meyer-Zhao; G.~A. Renting; E Valentijn; Willem-Jan Vriend
In this paper, we present a newly designed and implemented scientific information system for the LOFAR Long Term Archive. It is a distributed multi-tier storage and data processing system that allows a number of users to process Petabytes of data. The LOFAR Information System is designed on the base of the Astro-WISE Information System. The system allows the use of the computing and storage resources of the BiGGrid combined with the metadata database along with the data access and data processing interfaces of Astro-WISE. The architecture of the system, problems solved during the implementation and scientific use cases for the system are also described.
Big Data from Space (BiDS'14) | 2014
E Valentijn; Kornelis Begeman; Andrey Belikov; Danny Boxhoorn; Gijs Verdoes Kleijn; John McFarland; Willem-Jan Vriend; Owen Williams
This paper presents a Near-Real-Time multi-GPU accelerated solution of the ωk Algorithm for Synthetic Aperture Radar (SAR) data focusing, obtained in Stripmap SAR mode. Starting from an input raw data, the algorithm subdivides it in a grid of a configurable number of bursts along track. A multithreading CPU-side support is made available in order to handle each graphic device in parallel. Then each burst is assigned to a separate GPU and processed including Range Compression, Stolt Mapping via ChirpZ and Azimuth Compression steps. We prove the efficiency of our algorithm by using Sentinel-1 raw data (approx. 3.3 GB) on a commodity graphics card; the single-GPU solution is approximately 4x faster than the industrial multi-core CPU implementation (General ACS SAR Processor, GASP), without significant loss of quality. Using a multi-GPU system, the algorithm is approximately 6x faster with respect to the CPU processor.For decades, field help in case of disasters on the Earth’s surface - like floods, fires or earthquakes - is supported by the analysis of remotely sensed data. In recent years, the monitoring of vehicles, buildings or areas fraught with risk has become another major task for satellite-based crisis intervention. Since these scenarios are unforeseen and time-critical, they require a fast and well coordinated reaction. If useful information is extracted out of image data in realtime directly on board a spacecraft, the timespan between image acquisition and an appropriate reaction can be shortened significantly. Furthermore, on board image analysis allows data of minor interest, e.g. cloud-contaminated scenes, to be discarded and/or treated with lower priority, which leads to an optimized usage of storage and downlink capacity. This paper describes the modular application framework of VIMOS, an on board image processing experiment for remote sensing applications. Special focus will be on resource management, safety and modular commandability.Gaia is an ESA cornerstone mission, which was successfully launched December 2013 and commenced operations in July 2014. Within the Gaia Data Processing and Analysis consortium, Coordination Unit 7 (CU7) is responsible for the variability analysis of over a billion celestial sources and nearly 4 billion associated time series (photometric, spectrophotometric, and spectroscopic), encoding information in over 800 billion observations during the 5 years of the mission, resulting in a petabyte scale analytical problem. In this article, we briefly describe the solutions we developed to address the challenges of time series variability analysis: from the structure for a distributed data-oriented scientific collaboration to architectural choices and specific components used. Our approach is based on Open Source components with a distributed, partitioned database as the core to handle incrementally: ingestion, distributed processing, analysis, results and export in a constrained time window.The seamless mosaicing of massive very high resolution imagery addresses several aspects related to big data from space. Data volume is directly proportional to the size the input data, i.e., order of several TeraPixels for a continent. Data velocity derives from the fact that the input data is delivered over several years to meet maximum cloud contamination constraints with the considered satellites. Data variety results from the need to collect and integrate various ancillary data for cloud detection, land/sea mask delineation, and adaptive colour balancing. This paper details how these 3 aspects of big data are handled and illustrates them for the creation of a seamless pan-European mosaic from 2.5m imagery (Land Monitoring/Urban Atlas Copernicus CORE 03 data set).The current development of satellite imagery means that a great volume of images acquired globally has to be understood in a fast and precise manner. Processing this large quantity of information comes at the cost of finding unsupervised algorithms to fulfill these tasks. Change detection is one of the main issues when talking about the analysis of satellite image time series (SITS). In this paper, we propose a method to analyze changes in SITS based on binary descriptors and on the Hamming distance, regarded as a similarity metric. In order to render an automatic and completely unsupervised technique towards solving this problem, the obtained distances are quantized into change levels using the Lloyd-Max’s algorithm. The experiments are carried on 11 Landsat images at 30 meters spatial resolution, covering an area of approximately 59 × 51 km2 over the surroundings of Bucharest, Romania, and containing information from six subbands of frequency.The Euclid Archive System prototype is a functional information system which is used to address the numerous challenges in the development of fully functional data processing system for Euclid. The prototype must support the highly distributed nature of the Euclid Science Ground System, with Science Data Centres in at least eight countries. There are strict requirements both on data quality control and traceability of the data processing. Data volumes will be greater than 10 Pbyte, with the actual volume being dependent on the amount of reprocessing required.In the space domain, all scientific and technological developments are accompanied by a growth of the number of data sources. More specifically, the world of observation knows this very strong acceleration and the demand for information processing follows the same pace. To meet this demand, the problems associated with non-interoperability of data must be efficiently resolved upstream and without loss of information. We advocate the use of linked data technologies to integrate heterogeneous and schema-less data that we aim to publish in the 5 stars scale in order to foster their re-use. By proposing the 5 stars data model, Tim Berners-Lee drew the perfect roadmap for the production of high quality linked data. In this paper, we present a technological framework that allows to go from raw, scattered and heterogeneous data to structured data with a well-defined and agreed upon semantics, interlinked with other dataset for their common objects.Reference data sets, necessary to the advancement of the field of object recognition by providing a point of comparison for different algorithms, are prevalent in the field of multimedia. Although sharing the same basic object recognition problem, in the field of remote sensing there is a need for specialized reference data sets. This paper would like to open the topic for discussion, by taking a first attempt at creating a reference data set for a satellite image. In doing so, important differences between annotating photographic and satellite images are highlighted, along with their impact on the creation of a reference data set. The results are discussed with a view toward creating a future methodology for the manual annotation of satellite images.The future atmospheric composition Sentinel missions will generate two orders of magnitude more data than the current missions and the operational processing of these big data is a big challenge. The trace gas retrieval from remote sensing data usually requires high-performance radiative transfer model (RTM) simulations and the RTM are usually the bottleneck for the operational processing of the satellite data. To date, multi-core CPUs and also Graphical Processing Units (GPUs) have been used for highly intensive parallel computations. In this paper, we are comparing multi-core and GPU implementations of an RTM based on the discrete ordinate solution method. With GPUs, we have achieved a 20x-40x speed-up for the multi-stream RTM, and 50x speed-up for the two-stream RTM with respect to the original single-threaded CPU codes. Based on these performance tests, an optimal workload distribution scheme between GPU and CPU is proposed. Finally, we discuss the performance obtained with the multi-core-CPU and GPU implementations of the RTM.The effective use of Big Data in current and future scientific missions requires intelligent data handling systems which are able to interface the user to complicated distributed data collections. We review the WISE Concept of Scientific Information Systems and the WISE solutions for the storage and processing as applied to Big Data.Interactive visual data mining, where the user plays a key role in learning process, has gained high attention in data mining and human-machine communication. However, this approach needs Dimensionality Reduction (DR) techniques to visualize image collections. Although the main focus of DR techniques lays on preserving the structure of the data, the occlusion of images and inefficient usage of display space are their two main drawbacks. In this work, we propose to use Non-negative Matrix Factorization (NMF) to reduce the dimensionality of images for immersive visualization. The proposed method aims to preserve the structure of data and at the same time reduce the occlusion between images by defining regularization terms for NMF. Experimental validations performed on two sets of image collections show the efficiency of the proposed method in respect to controlling the trade-off between structure preserving and less occluded visualization.This article provides a short overview about the TanDEM-X mission, its objectives and the payload ground segment (PGS) based on data management, processing systems and long term archive. Due to the large data volume of the acquired and processed products a main challenge in the operation of the PGS is to handle the required data throughput, which is a new dimension for the DLR PGS. To achieve this requirement, several solutions were developed and coordinated. Some of them were more technical nature whereas others optimized the workflows.Clustering of Earth Observation (EO) images has gained a high amount of attention in remote sensing and data mining. Here, each image is represented by a high-dimensional feature vector which could be computed as the results of coding algorithms of extracted local descriptors or raw pixel values. In this work, we propose to learn the features using discriminative Nonnegative Matrix factorization (DNMF) to represent each image. Here, we use the label of some images to produce new representation of images with more discriminative property. To validate our algorithm, we apply the proposed algorithm on a dataset of Synthetic Aperture Radar (SAR) and compare the results with the results of state-of-the-art techniques for image representation. The results confirm the capability of the proposed method in learning discriminative features leading to higher accuracy in clustering.
The Astrophysical Journal | 2005
F. Kemper; Willem-Jan Vriend; A. G. G. M. Tielens
arXiv: Astrophysics | 2007
E Valentijn; John Mc Farland; Jan Snigula; Kornelis Begeman; Danny Boxhoorn; Roeland Rengelink; Ewout Helmich; P. Heraudeau; G. Verdoes Kleijn; R. Vermeij; Willem-Jan Vriend; M. J. Tempelaar; E. Deul; K. Kuijken; M. Capaccioli; R. Silvotti; Ralf Bender; M. Neeser; R. P. Saglia; E. Bertin; Y. Mellier
Astronomy and Astrophysics | 2017
Jelte T. A. de Jong; Gijs Verdoes Kleijn; Thomas Erben; Hendrik Hildebrandt; Konrad Kuijken; Gert Sikkema; Massimo Brescia; Maciej Bilicki; N. R. Napolitano; Valeria Amaro; Kor G. Begeman; Danny Boxhoorn; Hugo Buddelmeijer; Stefano Cavuoti; F. Getman; A. Grado; Ewout Helmich; Z. Huang; Nancy Irisarri; Francesco La Barbera; Guiseppe Longo; John McFarland; Reiko Nakajima; M. Paolillo; E. Puddu; M. Radovich; A. Rifatto; C. Tortora; E Valentijn; Civita Vellucci
Astronomical Society of the Pacific | 2012
H. A. Holties; G. van Diepen; D. van Dok; Fokke Dijkstra; M. Loose; G.~A. Renting; C. Schrijvers; Willem-Jan Vriend
Experimental Astronomy | 2013
Andrey Belikov; Willem-Jan Vriend; Gert Sikkema
Astronomical Data Analysis Software and Systems XVI ASP | 2007
E Valentijn; John Mc Farland; Jan Snigula; Kornelis Begeman; Danny Boxhoorn; Roeland Rengelink; Ewout Helmich; P. Heraudeau; Gijsbert Verdoes Kleijn; R. Vermeij; Willem-Jan Vriend; M. J. Tempelaar; E. Deul; K. Kuijken; M. Capaccioli; R. Silvotti; Ralf Bender; M. Neeser; R. P. Saglia; E. Bertin; Y. Mellier