Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Geoffrey Stewart Morrison is active.

Publication


Featured researches published by Geoffrey Stewart Morrison.


Studies in Second Language Acquisition | 2005

AN APPROPRIATE METRIC FOR CUE WEIGHTING IN L2 SPEECH PERCEPTION: Response to Escudero and Boersma (2004)

Geoffrey Stewart Morrison

Flege, Bohn, and Jang (1997) and Escudero and Boersma (2004) analyzed first language-Spanish second language-English listeners’ perception of English /i/–/I/ continua that varied in spectral and duration properties. They compared individuals and groups on the basis of spectral reliance and duration reliance measures. These reliance measures indicate the change in identification rates from one extreme of the stimulus space to the other; they make use of only a portion of the data collected and suffer from a ceiling effect. The current paper presents a reanalysis of Escudero and Boersma’s data using firstorder logistic regression modeling. All of the available data contribute to the calculation of logistic regression coefficients, and they do not suffer from the same ceiling effect as the reliance measures. It is argued that—as a metric of cue weighting—logistic regression coefficients offer methodological and substantive advantages over the reliance measures.


Science & Justice | 2016

A demonstration of the application of the new paradigm for the evaluation of forensic evidence under conditions reflecting those of a real forensic-voice-comparison case.

Ewald Enzinger; Geoffrey Stewart Morrison; Felipe Ochoa

The new paradigm for the evaluation of the strength of forensic evidence includes: The use of the likelihood-ratio framework. The use of relevant data, quantitative measurements, and statistical models. Empirical testing of validity and reliability under conditions reflecting those of the case under investigation. Transparency as to decisions made and procedures employed. The present paper illustrates the use of the new paradigm to evaluate strength of evidence under conditions reflecting those of a real forensic-voice-comparison case. The offender recording was from a landline telephone system, had background office noise, and was saved in a compressed format. The suspect recording included substantial reverberation and ventilation system noise, and was saved in a different compressed format. The present paper includes descriptions of the selection of the relevant hypotheses, sampling of data from the relevant population, simulation of suspect and offender recording conditions, and acoustic measurement and statistical modelling procedures. The present paper also explores the use of different techniques to compensate for the mismatch in recording conditions. It also examines how system performance would have differed had the suspect recording been of better quality.


Science & Justice | 2016

Special issue on measuring and reporting the precision of forensic likelihood ratios: Introduction to the debate ☆

Geoffrey Stewart Morrison

The present paper introduces the Science & Justice virtual special issue on measuring and reporting the precision of forensic likelihood ratios - whether this should be done, and if so how. The focus is on precision (aka reliability) as opposed to accuracy (aka validity). The topic is controversial and different authors are expected to express a range of nuanced opinions. The present paper frames the debate, explaining the underlying problem and referencing classes of solutions proposed in the existing literature. The special issue will consist of a number of position papers, responses to those position papers, and replies to the responses.


Speech Communication | 2015

Mismatched distances from speakers to telephone in a forensic-voice-comparison case

Ewald Enzinger; Geoffrey Stewart Morrison

Illustration of methodology for implementing FVC based on conditions of a real case.Use of relevant data, quantitative measurements, statistical models to calculate LRs.Procedure for testing of validity and reliability under the conditions of the case.Investigation of bias due to mismatched distances of speakers to the microphone.Demonstration of three methods for mismatch compensation. In a forensic-voice-comparison case, one speaker (A) was standing a short distance away from another speaker (B) who was talking on a mobile telephone. Later, speaker A moved closer to the telephone. Shortly thereafter, there was a section of speech where the identity of the speaker was in question - the prosecution claiming that it was speaker A and the defense claiming it was speaker B. All material for training a forensic-voice-comparison system could be extracted from this single recording, but there was a near-far mismatch: Training data for speaker A were mostly far, training data for speaker B were near, and the disputed speech was near. Based on the conditions of this case we demonstrate a methodology for handling forensic casework using relevant data, quantitative measurements, and statistical models to calculate likelihood ratios. A procedure is described for addressing the degree of validity and reliability of a forensic-voice-comparison system under such conditions. Using a set of development speakers we investigate the effect of mismatched distances to the microphone and demonstrate and assess three methods for compensation.


Speech Communication | 2016

Multi-laboratory evaluation of forensic voice comparison systems under conditions reflecting those of a real forensic case (forensic_eval_01) Introduction

Geoffrey Stewart Morrison; Ewald Enzinger

This paper introduces an evaluation of forensic voice comparison systems.It includes the rules for the evaluation.The training and test data reflect the conditions of a real case.Operational and research laboratories are invited to participate.Results will be published in a Virtual Special Issue of Speech Communication. There is increasing pressure on forensic laboratories to validate the performance of forensic analysis systems before they are used to assess strength of evidence for presentation in court. Different forensic voice comparison systems may use different approaches, and even among systems using the same general approach there can be substantial differences in operational details. From case to case, the relevant population, speaking styles, and recording conditions can be highly variable, but it is common to have relatively poor recording conditions and mismatches in speaking style and recording conditions between the known- and questioned-speaker recordings. In order to validate a system intended for use in casework, a forensic laboratory needs to evaluate the degree of validity and reliability of the system under forensically realistic conditions. The present paper is an introduction to a Virtual Special Issue consisting of papers reporting on the results of testing forensic voice comparison systems under conditions reflecting those of an actual forensic voice comparison case. A set of training and test data representative of the relevant population and reflecting the conditions of this particular case has been released, and operational and research laboratories are invited to use these data to train and test their systems. The present paper includes the rules for the evaluation and a description of the evaluation metrics and graphics to be used. The name of the evaluation is: forensic_eval_01. Display Omitted


Forensic Science International | 2016

Use of relevant data, quantitative measurements, and statistical models to calculate a likelihood ratio for a Chinese forensic voice comparison case involving two sisters

Cuiling Zhang; Geoffrey Stewart Morrison; Ewald Enzinger

Currently, the standard approach to forensic voice comparison in China is the aural-spectrographic approach. Internationally, this approach has been the subject of much criticism. The present paper describes what we believe is the first forensic voice comparison analysis presented to a court in China in which a numeric likelihood ratio was calculated using relevant data, quantitative measurements, and statistical models, and in which the validity and reliability of the analytical procedures were empirically tested under conditions reflecting those of the case under investigation. The hypotheses addressed were whether the female speaker on a recording of a mobile telephone conversation was a particular individual, or whether it was that individuals younger sister. Known speaker recordings of both these individuals were recorded using the same mobile telephone as had been used to record the questioned-speaker recording, and customised software was written to perform the acoustic and statistical analyses.


Science & Justice | 2016

Refining the relevant population in forensic voice comparison - A response to Hicks et alii (2015) The importance of distinguishing information from evidence/observations when formulating propositions.

Geoffrey Stewart Morrison; Ewald Enzinger; Cuiling Zhang

Hicks et alii [Sci. Just. 55 (2015) 520-525. http://dx.doi.org/10.1016/j.scijus.2015.06.008] propose that forensic speech scientists not use the accent of the speaker of questioned identity to refine the relevant population. This proposal is based on a lack of understanding of the realities of forensic voice comparison. If it were implemented, it would make data-based forensic voice comparison analysis within the likelihood ratio framework virtually impossible. We argue that it would also lead forensic speech scientists to present invalid unreliable strength of evidence statements, and not allow them to conduct the tests that would make them aware of this problem.


Science & Justice | 2017

Reply paperWhat should a forensic practitioner's likelihood ratio be? II☆

Geoffrey Stewart Morrison

In the debate as to whether forensic practitioners should assess and report the precision of the strength of evidence statements that they report to the courts, I remain unconvinced by proponents of the position that only a subjectivist concept of probability is legitimate. I consider this position counterproductive for the goal of having forensic practitioners implement, and courts not only accept but demand, logically correct and scientifically valid evaluation of forensic evidence. In considering what would be the best approach for evaluating strength of evidence, I suggest that the desiderata be (1) to maximise empirically demonstrable performance; (2) to maximise objectivity in the sense of maximising transparency and replicability, and minimising the potential for cognitive bias; and (3) to constrain and make overt the forensic practitioners subjective-judgement based decisions so that the appropriateness of those decisions can be debated before the judge in an admissibility hearing and/or before the trier of fact at trial. All approaches require the forensic practitioner to use subjective judgement, but constraining subjective judgement to decisions relating to selection of hypotheses, properties to measure, training and test data to use, and statistical modelling procedures to use - decisions which are remote from the output stage of the analysis - will substantially reduce the potential for cognitive bias. Adopting procedures based on relevant data, quantitative measurements, and statistical models, and directly reporting the output of the statistical models will also maximise transparency and replicability. A procedure which calculates a Bayes factor on the basis of relevant sample data and reference priors is no less objective than a frequentist calculation of a likelihood ratio on the same data. In general, a Bayes factor calculated using uninformative or reference priors will be closer to a value of 1 than a frequentist best estimate likelihood ratio. The bound closest to 1 based on a frequentist best estimate likelihood ratio and an assessment of its precision will also, by definition, be closer to a value of 1 than the frequentist best estimate likelihood ratio. From a practical perspective, both procedures shrink the strength of evidence value towards the neutral value of 1. A single-value Bayes factor or likelihood ratio may be easier for the courts to handle than a distribution. I therefore propose as a potential practical solution, the use of procedures which account for imprecision by shrinking the calculated Bayes factor or likelihood ratio towards 1, the choice of the particular procedure being based on empirical demonstration of performance.


Archive | 2018

Electronic Evidence: Challenges and Opportunities for Law Enforcement

Sabine Berghs; Geoffrey Stewart Morrison; Caroline Goemans-Dorny

This chapter presents a cross-section of law enforcement issues addressed as part of the research conducted by INTERPOL in the framework of the European Informatics Data Exchange Framework for Courts and Evidence (EVIDENCE) project. The aim of the research was to formulate recommendations and best practice based on the status quo and challenges identified regarding the handling of electronic evidence by law enforcement agencies (LEAs).


Forensic Science International | 2018

A response to Marquis et al. (2017) What is the error margin of your signature analysis

Geoffrey Stewart Morrison; Kaye Ballentyne; Patrick H. Geoghegan

Marquis et al (2017) [What is the error margin of your signature analysis? Forensic Science International, 281, e1–e8] ostensibly presents a model of how to respond to a request from a court to state an “error margin” for a conclusion from a forensic analysis. We interpret the court’s request as an explicit request for meaningful empirical validation to be conducted and the results reported. Marquis et al (2017), however, recommends a method based entirely on subjective judgement and does not subject it to any empirical validation. We believe that much resistance to the adoption of the likelihood ratio framework is not to the idea of assessing the relative probabilities (or likelihoods) of the evidence under prosecution and defence hypotheses per se, but to what is perceived to be unwarranted subjective assignment of those probabilities. In order to maximize transparency, replicability, and resistance to cognitive bias, we recommend the use of methods based on relevant data, quantitative measurements, and statistical models. If the method is based on subjective judgement, the output should be empirically calibrated. Irrespective of the basis of the method, its implementation should be empirically validated under conditions reflecting those of the case at hand.

Collaboration


Dive into the Geoffrey Stewart Morrison's collaboration.

Top Co-Authors

Avatar

Ewald Enzinger

Austrian Academy of Sciences

View shared research outputs
Top Co-Authors

Avatar

Cuiling Zhang

Australian National University

View shared research outputs
Top Co-Authors

Avatar

Felipe Ochoa

University of New South Wales

View shared research outputs
Top Co-Authors

Avatar

Ewald Enzinger

Austrian Academy of Sciences

View shared research outputs
Top Co-Authors

Avatar

Cuiling Zhang

Australian National University

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Kristy A. Martire

University of New South Wales

View shared research outputs
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge