Adam Craig Pocock
University of Manchester
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Adam Craig Pocock.
Electronic Notes in Theoretical Computer Science | 2010
Jeremy Singer; Gavin Brown; Mikel Luján; Adam Craig Pocock; Paraskevas Yiapanis
Fundamental nano-patterns are simple, static, binary properties of Java methods, such as ObjectCreator and Recursive. We present a provisional catalogue of 17 such nano-patterns. We report statistical and information theoretic metrics to show the frequency of nano-pattern occurrence in a large corpus of open-source Java projects. We proceed to give two example case studies that demonstrate potential applications for nano-patterns. The first study involves a quantitative comparison of two popular Java benchmarking suites, in terms of their relative object-orientedness and diversity. The second study involves applying machine learning techniques to program comprehension, using method nano-patterns as learning features. In both studies, nano-patterns provide concise summaries of Java methods to enable efficient and effective analysis.
ieee international symposium on workload characterization | 2010
Nikolas Ioannou; Jeremy Singer; Salman Khan; Polychronis Xekalakis; Paraskevas Yiapanis; Adam Craig Pocock; Gavin Brown; Mikel Luján; Ian Watson; Marcelo Cintra
Thread-Level Speculation (TLS) facilitates the extraction of parallel threads from sequential applications. Most prior work has focused on developing the compiler and architecture for this execution paradigm. Such studies often narrowly concentrated on a specific design point. On the other hand, other studies have attempted to assess how well TLS performs if some architectural/ compiler constraint is relaxed. Unfortunately, such previous studies have failed to truly assess TLS performance potential, because they have been bound to some specific TLS architecture and have ignored one or another important TLS design choice, such as support for out-of-order task spawn or support for intermediate checkpointing.
international conference on multiple classifier systems | 2010
Adam Craig Pocock; Paraskevas Yiapanis; Jeremy Singer; Mikel Luján; Gavin Brown
Ozas Online Boosting algorithm provides a version of AdaBoost which can be trained in an online way for stationary problems. One perspective is that this enables the power of the boosting framework to be applied to datasets which are too large to fit into memory. The online boosting algorithm assumes the data distribution to be independent and identically distributed (i.i.d.) and therefore has no provision for concept drift. We present an algorithm called Online Non-Stationary Boosting (ONSBoost) that, like Online Boosting, uses a static ensemble size without generating new members each time new examples are presented, and also adapts to a changing data distribution. We evaluate the new algorithm against Online Boosting, using the STAGGER dataset and three challenging datasets derived from a learning problem inside a parallelising virtual machine. We find that the new algorithm provides equivalent performance on the STAGGER dataset and an improvement of up to 3% on the parallelisation datasets.
international conference on big data | 2015
Anthony Kleerekoper; Michael Pappas; Adam Craig Pocock; Gavin Brown; Mikel Luján
With the growth of high dimensional data, feature selection is a vital component of machine learning as well as an important stand alone data analytics tool. Without it, the computation cost of big data analytics can become unmanageable and spurious correlations and noise can reduce the accuracy of any results. Feature selection removes irrelevant and redundant information leading to faster, more reliable data analysis. Feature selection techniques based on information theory are among the fastest known and the Manchester AnalyticS Toolkit (MAST) provides an efficient, parallel and scalable implementation of these methods. This paper considers a number of data structures for storing the frequency counters that underpin MAST. We show that preprocessing the data to reduce the number of zero-valued counters in an array structure results in an order of magnitude reduction in both memory usage and execution time compared to state of the art structures that use explicit mappings to avoid zero-valued counters. We also describe a number of parallel processing techniques that enable MAST to scale linearly with the number of processors even on NUMA architectures. MAST targets scale-up servers rather than scale-out clusters and we show that it performs orders of magnitude faster than existing tools. Moreover, we show that MAST is 3.5 times faster than a scale-out solution built for Spark running on the same server. As an example of the performance of MAST, we were able to process a dataset of 100 million examples and 100,000 features in under 10 minutes on a four socket server which each socket containing an 8-core Intel Xeon E5-4620 processor.
Journal of Machine Learning Research | 2012
Gavin Brown; Adam Craig Pocock; Ming-Jie Zhao; Mikel Luján
Journal of Machine Learning Research | 2013
Ming-Jie Zhao; Narayanan Unny Edakunni; Adam Craig Pocock; Gavin Brown
neural information processing systems | 2014
Jean-Baptiste Tristan; Daniel Huang; Joseph Tassarotti; Adam Craig Pocock; Stephen J. Green; Guy L. Steele
Journal of Machine Learning Research | 2012
Gavin Brown; Adam Craig Pocock; Ming-Jie Zhao; Mikel Luján
[Thesis]. Manchester, UK: The University of Manchester; 2012. | 2012
Adam Craig Pocock
international conference on artificial intelligence and statistics | 2012
Adam Craig Pocock; Mikel Luján; Gavin Brown