Abstract
The well-known and extensively studied Linear Discriminant Analysis (LDA) can have its performance lowered in
scenarios where data is not homoscedastic or not Gaussian. That is, the classical assumptions when LDA models
are built are not applicable, and consequently LDA projections would not be able to extract the needed features to
explain the intrinsic structure of data and for classes to be separated. As with many real word data sets, data
obtained using miniature spectrometers can suffer from such drawbacks which would limit the deployment of
such technology needed for food analysis. The solution presented in the paper is to divide classes into subclasses
and to use means of sub classes, classes, and data in the suggested between classes scatter metric. Further,
samples belonging to the same subclass are used to build a measure of within subclass scatterness. Such a solution solves the shortcoming of the classical LDA. The obtained results when using the proposed solution on food data and on general machine learning datasets show that the work in this paper compares well to and is very
competitive with similar sub-class LDA algorithms in the literature. An extension to a Hilbert space is also
presented; and the kernel version of the presented solution can be fused with its linear counter parts to yield
improved classification rates.
scenarios where data is not homoscedastic or not Gaussian. That is, the classical assumptions when LDA models
are built are not applicable, and consequently LDA projections would not be able to extract the needed features to
explain the intrinsic structure of data and for classes to be separated. As with many real word data sets, data
obtained using miniature spectrometers can suffer from such drawbacks which would limit the deployment of
such technology needed for food analysis. The solution presented in the paper is to divide classes into subclasses
and to use means of sub classes, classes, and data in the suggested between classes scatter metric. Further,
samples belonging to the same subclass are used to build a measure of within subclass scatterness. Such a solution solves the shortcoming of the classical LDA. The obtained results when using the proposed solution on food data and on general machine learning datasets show that the work in this paper compares well to and is very
competitive with similar sub-class LDA algorithms in the literature. An extension to a Hilbert space is also
presented; and the kernel version of the presented solution can be fused with its linear counter parts to yield
improved classification rates.
Original language | English |
---|---|
Article number | 105136 |
Pages (from-to) | 1-13 |
Number of pages | 13 |
Journal | Chemometrics and Intelligent Laboratory Systems |
Volume | 250 |
Early online date | 4 May 2024 |
DOIs | |
Publication status | Published (in print/issue) - 15 Jul 2024 |
Bibliographical note
Publisher Copyright:© 2024 The Authors
Data Access Statement
Data will be made available on request.Keywords
- Dimensionality reduction
- Feature extraction
- Linear discriminant analysis
- Subclass discriminant analysis
- Classification
- Food analysis
- Miniature spectrometers