Marcello Trovati
Edge Hill University
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Marcello Trovati.
soft computing | 2016
Marcello Trovati; Nik Bessis
The extraction of meaningful, accurate, and relevant information is at the core of Big Data research. Furthermore, the ability to obtain an insight is essential in any decision-making process, even though the diverse and complex nature of big data sets raises a multitude of challenges. In this paper, we propose a novel method to address the automated assessment of influence among concepts in big data sets. This is carried out by investigating their mutual co-occurrence, which is determined via topologically reducing the corresponding network. The main motivation is to provide a toolbox to classify and analyse influence properties, which can be used to investigate their dynamical and statistical behaviour, potentially leading to a better understanding and prediction of the properties of the system(s) they model. An evaluation was carried out on two real-world data sets, which were analysed to test the capabilities of our system. The results show the potential of our approach, indicating both accuracy and efficiency.
International Journal of Distributed Systems and Technologies | 2015
Marcello Trovati
The topological and dynamical properties of real-world networks have attracted extensive research from a variety of multi-disciplinary fields. They, in fact, model typically big datasets which pose interesting challenges, due to their intrinsic size and complex interactions, as well as the dependencies between their different sub-parts. Therefore, defining networks based on such properties, is unlikely to produce usable information due to their complexity and the data inconsistencies which they typically contain. In this paper, the authors discuss the evaluation of a method as part of ongoing research which aims to mine data to assess whether their associated networks exhibit properties comparable to well-known structures, namely scale-free, small world and random networks. For this, they will use a large dataset containing information on the seismologic activity recorded by the European-Mediterranean Seismological Centre. The authors will show that it provides an accurate, agile, and scalable tool to extract useful information. This further motivates their effort to produce a big data analytics tool which will focus on obtaining in-depth intelligence from both structured and unstructured big datasets. This will ultimately lead to a better understanding and prediction of the properties of the system(s) they model.
EPJ Data Science | 2016
Evangelos Pournaras; Jovan Nikolic; Pablo Velásquez; Marcello Trovati; Nik Bessis; Dirk Helbing
Participation in social sensing applications is challenged by privacy threats. Large-scale access to citizens’ data allow surveillance and discriminatory actions that may result in segregation phenomena in society. On the contrary are the benefits of accurate computing analytics required for more informed decision-making, more effective policies and regulation of techno-socio-economic systems supported by ‘Internet-of Things’ technologies. In contrast to earlier work that either focuses on privacy protection or Big Data analytics, this paper proposes a self-regulatory information sharing system that bridges this gap. This is achieved by modeling information sharing as a supply-demand system run by computational markets. On the supply side lie the citizens that make incentivized but self-determined decisions about the level of information they share. On the demand side stand data aggregators that provide rewards to citizens to receive the required data for accurate analytics. The system is empirically evaluated with two real-world datasets from two application domains: (i) Smart Grids and (ii) mobile phone sensing. Experimental results quantify trade-offs between privacy-preservation, accuracy of analytics and costs from the provided rewards under different experimental settings. Findings show a higher privacy-preservation that depends on the number of participating citizens and the type of data summarized. Moreover, analytics with summarization data tolerate high local errors without a significant influence on the global accuracy. In other words, local errors cancel out. Rewards can be optimized to be fair so that citizens with more significant sharing of information receive higher rewards. All these findings motivate a new paradigm of truly decentralized and ethical data analytics.
complex, intelligent and software intensive systems | 2014
Marcello Trovati; Nik Bessis; Anna Huber; Asta Zelenkauskaite; Eleana Asimakopoulou
Networks are widely used to model a variety of complex, often multi-disciplinary, systems in which the relationships between their sub-parts play a significant role. In particular, there is extensive research on the topological properties associated with their structure as this allows the analysis of the overall behaviour of such networks. However, extracting networks from structured and unstructured data sets raises several challenges, including addressing any inconsistency present in the data, as well as the difficulty in investigating their properties especially when the topological structure is not fully determined or not explicitly defined. In this paper, we propose a novel method to address the automated identification, assessment and ranking of the most likely structure associated with networks extracted from a variety of data sets. More specifically, our approach allows to mine data to assess whether their associated networks exhibit properties comparable to well-known structures, namely scale-free, small world and random networks. The main motivation is to provide a toolbox to classify and analyse real-world networks otherwise difficult to fully assess due to their potential lack of structure. This can be used to investigate their dynamical and statistical behaviour which would potentially lead to a better understanding and prediction of the properties of the system (s) they model. Our initial validation shows the potential of our method providing relevant and accurate results.
Concurrency and Computation: Practice and Experience | 2016
Omar Behadada; Marcello Trovati; Mohammed Amine Chikh; Nik Bessis
In this paper, we introduce a novel method to define semi‐automatically fuzzy partition rules to provide a powerful and accurate insight into cardiac arrhythmia. In particular, we define a text mining approach applied to a large dataset consisting of the freely available scientific papers provided by PubMed. The information extracted is then integrated with expert knowledge, as well as experimental data, to provide a robust, scalable and accurate system, which can successfully address the challenges posed by the management and assessment of big data in the medical sector. The evaluation we carried out shows an accuracy rate of 93% and interpretability of 0.646, which clearly shows that our method provides an excellent balance between accuracy and system transparency. Furthermore, this contributes substantially to the knowledge discovery and offers a powerful tool to facilitate the decision‐making process. Copyright
intelligent networking and collaborative systems | 2014
Marcello Trovati; Eleana Asimakopoulou; Nik Bessis
The topological and dynamical properties of real-world networks have attracted extensive research from a variety of multi-disciplinary fields. They, in fact, model typically big datasets which pose interesting challenges, due to their intrinsic size and complex interactions, as well as the dependencies between their different sub-parts. Therefore, defining networks based on such properties, is unlikely to produce usable information due to their complexity and the data inconsistencies which they typically contain. In this paper, we discuss the evaluation of a method as part of ongoing research which aims to mine data to assess whether their associated networks exhibit properties comparable to well-known structures, namely scale-free, small world and random networks. For this, we will use a large dataset containing information on the seismologic activity recorded by the European-Mediterranean Seismological Centre. We will show that it provides an accurate, agile, and scalable tool to extract useful information. This further motivates our effort to produce a big data analytics tool which will focus on obtaining in-depth intelligence from both structured and unstructured big datasets. This will ultimately lead to a better understanding and prediction of the properties of the system(s) they model.
soft computing | 2018
Xiaolong Xu; Hanzhong Rong; Marcello Trovati; Mark Liptrott; Nik Bessis
Combinatorial optimization problems are typically NP-hard, due to their intrinsic complexity. In this paper, we propose a novel chaotic particle swarm optimization algorithm (CS-PSO), which combines the chaos search method with the particle swarm optimization algorithm (PSO) for solving combinatorial optimization problems. In particular, in the initialization phase, the priori knowledge of the combination optimization problem is used to optimize the initial particles. According to the properties of the combination optimization problem, suitable classification algorithms are implemented to group similar items into categories, thus reducing the number of combinations. This enables a more efficient enumeration of all combination schemes and optimize the overall approach. On the other hand, in the chaos perturbing phase, a brand-new set of rules is presented to perturb the velocities and positions of particles to satisfy the ideal global search capability and adaptability, effectively avoiding the premature convergence problem found frequently in traditional PSO algorithm. In the above two stages, we control the number of selected items in each category to ensure the diversity of the final combination scheme. The fitness function of CS-PSO introduces the concept of the personalized constraints and general constrains to get a personalized interface, which is used to solve a personalized combination optimization problem. As part of our evaluation, we define a personalized dietary recommendation system, called Friend, where CS-PSO is applied to address a healthy diet combination optimization problem. Based on Friend, we implemented a series of experiments to test the performance of CS-PSO. The experimental results show that, compared with the typical HLR-PSO, CS-PSO can recommend dietary schemes more efficiently, while obtaining the global optimum with fewer iterations, and have the better global ergodicity.
soft computing | 2017
Yeqin Shao; Marcello Trovati; Quan Shi; Olga Angelopoulou; Eleana Asimakopoulou; Nik Bessis
The identification of non-genuine or malicious messages poses a variety of challenges due to the continuous changes in the techniques utilised by cyber-criminals. In this article, we propose a hybrid detection method based on a combination of image and text spam recognition techniques. In particular, the former is based on sparse representation-based classification, which focuses on the global and local image features, and a dictionary learning technique to achieve a spam and a ham sub-dictionary. On the other hand, the textual analysis is based on semantic properties of documents to assess the level of maliciousness. More specifically, we are able to distinguish between meta-spam and real spam. Experimental results show the accuracy and potential of our approach.
Chaos | 2007
Marcello Trovati; Peter Ashwin
Piecewise isometries (PWIs) are known to have dynamical properties that generate interesting geometric planar packings. We analyze a particular PWI introduced by Goetz that generates a packing by periodically coded cells, each of which is a pentagon. Our main result is that the tangency graph associated with this packing is a forest (i.e., has no nontrivial cycles). We show, however, that this is not a general property of PWIs by giving an example that has an infinite number of cycles in the tangency graph of its periodically coded cells.
Applied Soft Computing | 2017
Marcello Trovati; Jer Hayes; Francesco Palmieri; Nik Bessis
Abstract Mining large amounts of unstructured data for extracting meaningful, accurate, and actionable information, is at the core of a variety of research disciplines including computer science, mathematical and statistical modelling, as well as knowledge engineering. In particular, the ability to model complex scenarios based on unstructured datasets is an important step towards an integrated and accurate knowledge extraction approach. This would provide a significant insight in any decision making process driven by Big Data analysis activities. However, there are multiple challenges that need to be fully addressed in order to achieve this, especially when large and unstructured data sets are considered. In this article we propose and analyse a novel method to extract and build fragments of Bayesian networks (BNs) from unstructured large data sources. The results of our analysis show the potential of our approach, and highlight its accuracy and efficiency. More specifically, when compared with existing approaches, our method addresses specific challenges posed by the automated extraction of BNs with extensive applications to unstructured and highly dynamic data sources. The aim of this work is to advance the current state-of-the-art approaches to the automated extraction of BNs from unstructured datasets, which provide a versatile and powerful modelling framework to facilitate knowledge discovery in complex decision scenarios.