Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Hirotaka Nakasone is active.

Publication


Featured researches published by Hirotaka Nakasone.


Journal of the Acoustical Society of America | 2011

Variations in recorded acoustic gunshot waveforms generated by small firearms

Steven D. Beck; Hirotaka Nakasone; Kenneth W. Marr

Analysis of recorded acoustic gunshot signals to determine firearm waveform characteristics requires an understanding of the impulsive signal events, how the waveforms vary among different sources, and how the waveforms are affected by the environment and the recording system. This paper presents empirical results from waveforms produced by different small firearms and an analysis of their variations under different and controlled conditions. Acoustic signals were generated using multiple firearm makes and models firing different ammunition types. Simultaneous recordings from the microphones located at different distances from the source and at different azimuth angles (from the line-of-fire) were used to study source characteristics and sound propagation effects. The results indicate that recorded gunshot waveforms generally consist of multiple acoustic events, and these are observable depending on the received distance and azimuth angle. The source blast size, microphone distance, and microphone azimuth angle are the primary factors affecting the recorded muzzle blast characteristics. Ground or object reflections and ballistic shockwaves and their reflections can interfere with the muzzle blast waveform and its measurements. This experiment confirmed and quantified the wide range of correlation results between waveforms recorded from different source, microphone distance, and microphone angle configurations.


Journal of the Acoustical Society of America | 2011

An introduction to forensic gunshot acoustics

Steven D. Beck; Hirotaka Nakasone; Kenneth W. Marr

Due to the proliferation of audio recording devices in the military, law enforcement, and the civilian community, there has been an increase in the number of recorded gunshot sounds submitted for forensic analysis. A gunshot sound is composed of one or more discrete acoustic transient events. The two primary acoustic events are the muzzle blast (bang) and the ballistic shockwave (crack). The acoustic event characteristics depend on their source generating mechanisms and vary according to the firearm make, model, barrel length, and the specific ammunition characteristics. Forensic gunshot analysis deals with a single recorded shot lasting for a fraction of a second. These acoustic events are usually high intensity, often up to 160 dB SPL, are highly directional, and are often recorded in high distortion environments. Forensic gunshot analysis must take into account variations in the source generation characteristics and the sources of distortion for these recorded acoustic events in order to answer these f...


2006 IEEE Odyssey - The Speaker and Language Recognition Workshop | 2006

Speaker Recognition Score-Normalization to Compensate for SNR and Duration

Jørgen E. Harmse; Steven D. Beck; Hirotaka Nakasone

The decision criterion for automatic speaker verification tests is based on minimization of a weighted sum of the miss and false alarm probabilities. These probabilities are derived from an evaluation of claimant and impostor scores using a representative population of recorded speech samples. However, in applications such as forensic speaker verification, the signal quality and the recording conditions of the speech samples are usually unknown and generally not matched to the evaluation conditions for the defined error probabilities. For example, test samples are often of short duration, have significant noise, and are from uncertain channels. It is therefore necessary to normalize the speaker test scores or to adjust detection thresholds in accordance with the recorded signal conditions. Instead of accounting for all possibilities, evaluations were conducted for a few specific joint combinations of signal-to-noise ratio (SNR) and speech duration for both the training and test sets. A composite regression model was developed to predict the necessary adjustments for any measured value of these conditions. In addition, a method is discussed to interpret the normalized scores relative to a set of desired Type I and Type II error probabilities


Journal of the Acoustical Society of America | 2011

Progress toward a forensic voice data format standard

James L. Wayman; Joseph P. Campbell; Pedro A. Torres-Carrasquillo; Peter T. Higgins; Alvin F. Martin; Hirotaka Nakasone; Craig S. Greenberg; Mark Pryzbocki

The de facto international standard for the forensic exchange of data for biometric recognition is ANSI/NIST ITL-1/2, “Data Format for the Interchange of Fingerprint, Facial, and Other Biometric Information.” This format is used by law enforcement, intelligence, military, and homeland security organizations thoughout the world to exchange fingerprint, face, scar/mark/tatoo, iris, and palmprint data. To date, however, there is no provision within the standard for the exchange of audio data for the purpose of forensic speaker recognition. During the recent 5-year update process for ANSI/NIST ITL-1/2, a consensus decision was made to advance a voice data format type under the name “Type 11 record.” Creating such an exchange format type, however, is far from straight forward—the problem being not the encoding of the autio data, for which many accepted standards exist, but rather in reaching a consensus on the metadata needed to support the varied mission requirements across the stakeholder communities. In thi...


Odyssey | 2001

Forensic automatic speaker recognition

Hirotaka Nakasone; Steven D. Beck


language resources and evaluation | 2004

The Mixer Corpus of Multilingual, Multichannel Speaker Recognition Data.

Christopher Cieri; Joseph P. Campbell; Hirotaka Nakasone; David Miller; Kevin Walker


language resources and evaluation | 2006

The Mixer and Transcript Reading Corpora: Resources for Multilingual, Crosschannel Speaker Recognition Research

Christopher Cieri; Walter D. Andrews; Joseph P. Campbell; George R. Doddington; John J. Godfrey; Shudong Huang; Mark Liberman; Alvin F. Martin; Hirotaka Nakasone; Mark A. Przybocki; Kevin Walker


language resources and evaluation | 2004

Conversational Telephone Speech Corpus Collection for the NIST Speaker Recognition Evaluation 2004

Alvin F. Martin; David Miller; Mark A. Przybocki; Joseph P. Campbell; Hirotaka Nakasone


The Speaker and Language Recognition Workshop | 2004

The MMSR Bilingual and Crosschannel Corpora for Speaker Recognition Research and Evaluation

Joseph P. Campbell; Hirotaka Nakasone; Christopher Cieri; David Miller; Kevin Walker; Alvin F. Martin; Mark A. Przybocki


Journal of The Audio Engineering Society | 1998

Signal Convolution of Recorded Free-Field Gunshot Sounds

Bruce E. Koenig; Shawn M. Hoffman; Hirotaka Nakasone; Steven D. Beck

Collaboration


Dive into the Hirotaka Nakasone's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar

Joseph P. Campbell

Massachusetts Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Alvin F. Martin

National Institute of Standards and Technology

View shared research outputs
Top Co-Authors

Avatar

Christopher Cieri

University of Pennsylvania

View shared research outputs
Top Co-Authors

Avatar

David Miller

University of Pennsylvania

View shared research outputs
Top Co-Authors

Avatar

Kevin Walker

University of Pennsylvania

View shared research outputs
Top Co-Authors

Avatar

Mark A. Przybocki

National Institute of Standards and Technology

View shared research outputs
Top Co-Authors

Avatar

Reva Schwartz

United States Secret Service

View shared research outputs
Top Co-Authors

Avatar

Craig S. Greenberg

National Institute of Standards and Technology

View shared research outputs
Top Co-Authors

Avatar

James L. Wayman

The Aerospace Corporation

View shared research outputs
Researchain Logo
Decentralizing Knowledge