Grin logo
de en es fr
Shop
GRIN Website
Publish your texts - enjoy our full service for authors
Go to shop › Engineering - Computer Engineering

Speaker Recognition

Title: Speaker Recognition

Bachelor Thesis , 2019 , 26 Pages , Grade: A

Autor:in: Bandar Hezam (Author)

Engineering - Computer Engineering
Excerpt & Details   Look inside the ebook
Summary Excerpt Details

Voice recognition is a computer software program or hardware device with the ability to decode the human voice. Voice recognition is a system that allows for a secure method of authenticating speakers, the system work in such a way that it general speaker model during the enrollment phase which based on the speaker characteristics. The system testing phase typically involves making a claim on the identity of an unknown speaker using the given speech characteristics and the trained models.

However, speaker identification is known to be one among the two categories of speaker recognition system because speaker recognition can be categorized also as speaker verification whereas, the main difference between both speaker identification and speaker verification ensure to known if the person speaking and claim to be is fully verified while speaker identification make multiple decision by comparing of the person speaking with the one trained or store in database as an attempt to identify the speaker. The interest of the assignment is speaker identification; therefore, speaker identification is the main focus for this study.

Excerpt


Table of Contents

1 Introduction.

1.1 Theoretical Concepts

1.1.1 Speaker Recognition

1.1.2 Classification of Automatic Speaker Recognition

1.1.3 Speech Feature Extraction

2 Objectives.

3 Design implementation.

3.1 Vocal Activity Detection (VAD).

3.2 Speaker Identification.

3.2.1 Frame Blocking:

3.2.2 Widowing:

3.2.3 Mel-frequency Wrapping:

3.2.4 Cepstrum and Feature Extraction:

3.2.5 Distance Calculation:

3.2.6 GUI.

4 Design innovativeness.

5 Simulation results.

5.1 Train/ Enrollment Result.

5.2 Recognition.

5.3 GUI Result.

5.4 Euclidean distance between voices.

6 Discussion.

7 Conclusion.

8 References.

Objectives and Topics

The primary aim of this work is to design and implement a robust text-independent speaker identification system in a MATLAB environment. The research focuses on developing an automated process to identify individual speakers from a recorded audio track, utilizing voice activity detection (VAD) and Mel Frequency Cepstral Coefficients (MFCC) to isolate and characterize voice biometrics for a set of ten known speakers while identifying unknown voices.

  • Implementation of Voice Activity Detection (VAD) for noise-robust speech extraction.
  • Application of Mel Frequency Cepstral Coefficients (MFCC) for acoustic feature representation.
  • Development of a vector quantization algorithm to classify and match speakers.
  • Design of a Graphical User Interface (GUI) for system interaction and speaker identification.
  • Evaluation of system accuracy via Euclidean distance calculations between trained and tested audio samples.

Excerpt from the Book

1 Introduction.

The ability to recognize people by their voice is an important social behavior. Identifying a person based on speech alone, also, is known that speech is like a speaker-dependent feature that enables us to identify or recognize friends over the phone. Human voice comprises of numerous discriminative features with the ability to identify speakers, our voice comprises of significant energy of frequency range from zero to around 5kHz whereas the primary target of this assignment speaker identification is to extract, characterize and recognize individual speaker using an audio track of minimum of minute recording in order to identify information about the speaker identity Though temporal period of speech signals like correlation, zero crossing and lots more are assume constant over short period (Vibha Tiwari, 2010) that means, in the case of using hamming window, voice signal is divided into number of blocks of short duration in order to be able to do further processing such as normal Fourier series transform.

However, speaker voice identification for audio track with group people can be challenging if individual voice activities are not properly analyzed or detected, to approach this issue voice activities detection (VAD) algorithm can approach, this is a very useful technique that helps to improve the performance of voice recognition or speaker identification system working in these scenarios. Voice activity detection (VAD) is used in most voice recognition systems within the features extraction process for speech enhancement. The techniques usually involve when noise statistics such as spectrum are estimated during the non-voice period so that the speech enhancement algorithm can be applied such as wiener filer or spectral subtraction. VAD is also useful for non-voice frame-dropping in voice recognition in order to reduce the number of insertion errors caused by noise.

Chapter Summary

1 Introduction.: This chapter defines the significance of voice recognition technology, introduces the core challenges of speaker identification, and outlines the utilization of VAD and MFCC algorithms.

1.1 Theoretical Concepts: This section provides the foundational knowledge behind voice recognition, differentiating between speaker verification and speaker identification.

1.1.1 Speaker Recognition: Describes the two essential sessions of the identification process: the training/enrollment phase and the testing/operation phase.

1.1.2 Classification of Automatic Speaker Recognition: Discusses the importance of speech signal analysis for authentication systems and the difference between user identification and verification.

1.1.3 Speech Feature Extraction: Explains the necessity of extracting distinct voice features, listing methods like MFCC and LPC, and details the mathematical implementation of Linear Predictive Analysis.

2 Objectives.: Sets out the specific goals for the study, including algorithm research, implementation, and validating the system with a ten-speaker data set.

3 Design implementation.: Details the structural framework of the project, including the system architecture, code-level procedures, and the integration of flow charts and a user interface.

3.1 Vocal Activity Detection (VAD).: Outlines the algorithm used to isolate human speech from noise by analyzing band ratios, energy estimates, and frame smoothing.

3.2 Speaker Identification.: Explains the MFCC structure and technical specifications used to process input audio files at a 44.1kHz sampling rate.

3.2.1 Frame Blocking:: Describes the initial signal processing stage where continuous audio is partitioned into N-sized samples for analysis.

3.2.2 Widowing:: Details the application of window tapering to minimize signal discontinuities and spectral distortion at frame boundaries.

3.2.3 Mel-frequency Wrapping:: Covers the conversion of frequencies to the Mel scale to better reflect human auditory perception characteristics.

3.2.4 Cepstrum and Feature Extraction:: Explains the final calculations performed on LogFBEs to derive cepstral coefficients for speaker model training.

3.2.5 Distance Calculation:: Describes the use of Euclidean distance metrics to compare test samples against the database and select the best-matching speaker.

3.2.6 GUI.: Presents the design and functionality of the software interface created to handle user-led testing and viewing of identification results.

4 Design innovativeness.: Reflections on the custom combination of VAD, MFCC-based extraction, and vector quantization to optimize the identification process.

5 Simulation results.: Presents the visual and quantitative output of the system, including graphs of signal processing and result logs.

5.1 Train/ Enrollment Result.: Shows the graphical output of the training session where speakers were enrolled into the database.

5.2 Recognition.: Displays the simulation outcomes when the system processes audio containing unknown speakers.

5.3 GUI Result.: Shows the practical application of the designed interface in displaying speaker IDs and photos after a test run.

5.4 Euclidean distance between voices.: Provides numerical proof of the identification accuracy by presenting computed distance arrays and identification selection.

6 Discussion.: Evaluates why MFCC was chosen over LPC, citing its alignment with human auditory perception and its simplicity in implementation.

7 Conclusion.: Summarizes that the developed MATLAB system is successful and suggests potential real-world applications in security and forensics.

8 References.: Lists the academic sources used to build the algorithmic foundation of the research.

Keywords

Speaker Identification, Voice Activity Detection, VAD, Mel Frequency Cepstral Coefficients, MFCC, MATLAB, Speech Feature Extraction, Linear Predictive Analysis, LPC, Euclidean Distance, Biometrics, Audio Signal Processing, Vector Quantization, Authentication Systems, Forensic Identification.

Frequently Asked Questions

What is the core focus of this research paper?

The paper focuses on developing an automated text-independent speaker identification system capable of distinguishing between various speakers within a single audio track, even in the presence of noise.

Which primary themes are addressed in this study?

The study centers on digital signal processing, acoustic feature extraction (specifically using MFCC), vocal segment isolation via VAD, and the practical implementation of identification workflows in MATLAB.

What is the ultimate objective of the proposed system?

The primary goal is to build a reliable solution that can identify one of ten known speakers from an unseen audio record, while also signaling when an unknown voice is detected.

Which scientific methods are utilized for feature extraction?

The author primarily utilizes Mel Frequency Cepstral Coefficients (MFCC), though it also discusses Linear Predictive Analysis (LPC) as an alternative for specific encoding scenarios.

What does the main implementation section cover?

The main section describes the step-by-step algorithms, including frame blocking, windowing, Mel-frequency wrapping, and the final classification logic based on calculated Euclidean distances.

How is the accuracy of the identification determined?

Accuracy is determined by measuring the Euclidean distance between the extracted features of the test audio and the stored centroids in the speaker database; the smallest distance indicates the closest match.

Why is Voice Activity Detection (VAD) included in the system?

VAD is included to automatically filter out silent periods and noise from the audio stream, ensuring that the feature extraction process only runs on valid speech segments, which significantly improves identification accuracy.

What role does the Graphical User Interface (GUI) play?

The GUI provides a user-friendly way to initiate the training and testing phases, visualize the results, and see the identity result — including the speaker's photo — immediately after the simulation finishes.

Is the system designed for text-dependent or text-independent recognition?

The system is designed for text-independent recognition, meaning it does not rely on the speaker uttering specific words to be identified.

What are the real-world implications of this work?

The developed system has potential applications in access control systems, forensic voice logging, telephonic banking, and biometric security systems where automated identity validation is critical.

Excerpt out of 26 pages  - scroll top

Details

Title
Speaker Recognition
College
National University of Malaysia  (Apu)
Course
Mechatronics
Grade
A
Author
Bandar Hezam (Author)
Publication Year
2019
Pages
26
Catalog Number
V1420967
ISBN (PDF)
9783346980229
ISBN (Book)
9783346980236
Language
English
Tags
Designing a speaker,
Product Safety
GRIN Publishing GmbH
Quote paper
Bandar Hezam (Author), 2019, Speaker Recognition, Munich, GRIN Verlag, https://www.grin.com/document/1420967
Look inside the ebook
  • Depending on your browser, you might see this message in place of the failed image.
  • Depending on your browser, you might see this message in place of the failed image.
  • Depending on your browser, you might see this message in place of the failed image.
  • Depending on your browser, you might see this message in place of the failed image.
  • Depending on your browser, you might see this message in place of the failed image.
  • Depending on your browser, you might see this message in place of the failed image.
  • Depending on your browser, you might see this message in place of the failed image.
  • Depending on your browser, you might see this message in place of the failed image.
  • Depending on your browser, you might see this message in place of the failed image.
Excerpt from  26  pages
Grin logo
  • Grin.com
  • Shipping
  • Contact
  • Privacy
  • Terms
  • Imprint