Saturday, September 15, 2012

Day 9 - Team B (Manas Mani)


Lecture No. 17:

Continued with the example of music preference with the SPSS file - GSS93 subset.
Divided work into 2 parts.
  1. music preference according to age
  2. music preference according to race
Data found in SPSS by choosing the option: Analyze>Table>Custom Table. Below is the option/window we get:

Then arrange the data, i.e. Age or Race, in the Range option. Mark components Traditional, Soft, Country & Rock. Put the data in the excel sheet and change the options in SPSS to display the same. (Excel Sheet Snapshot below)

The Data achieved from Age (or Race) is then transferred into Permap by coping the data into first notepad and then opening the option in Permap. The rule to be remembered and followed is that once the output is achieved, the option closer to the arrow head, is the option not liked/followed by that particular age group.


Lecture No. 18:

A review of what has been done:
  1. 1st level - frequency and X tab
  2. Cluster - K-means and Hierarchical
  3. MDS (perceptual mapping) - Overall Similarity and Attribute Based
  4. Factor Analysis
New topic:
     5.  Discriminant Analysis

What is Discriminant Analysis?

- Discriminant Analysis may be used for two objectives: either   we want to assess the adequacy of classification, given the group memberships of the objects under study; or we wish to assign objects to one of a number of (known) groups of objects. Discriminant Analysis may thus have a descriptive or a predictive objective.

Discriminant Analysis: Objectives and Properties
  • Assess the adequacy of a classification, given the group memberships.
  • Assign objects to one of a number of (known) groups of objects.
  • Note: supervised classification (= discriminant analysis) vs. unsupervised classification (= cluster analysis). Sometimes, along these lines, classification is distinguised from clustering.
  • Remark: discriminant analysis is “discrete prediction”, whereas regression analysis is “continuous prediction”.
  • So a multilayer perceptron neural network used for classification purposes is discriminant analysis; but used for the purposes of continuous mapping it is nonlinear regression.
(Note: For further information/reading, refer to http://www.classification-society.org/csna/mda-sw/M2/expose-discr-new.pdf)

Methods implemented in this area are Multiple Discriminant Analysis, Fisher's Linear Discriminant Analysis, and K-Nearest Neighbours Discriminant Analysis. To explain the above mentioned:

Multiple Discriminant Analysis
(MDA) is also termed Discriminant       Factor Analysis and Canonical Discriminant Analysis. It adopts a similar perspective to PCA: the rows of the data matrix to be examined constitute points in a multidimensional space, as also do the group mean vectors. Discriminating axes are determined in this space, in such a way that optimal separation of the predefined groups is attained. As with PCA, the problem becomes mathematically the eigenreduction of a real, symmetric matrix. The eigenvalues represent the discriminating power of the associated eigenvectors. The nYgroups lie in a space of dimension at most nY - 1. This will be the number of discriminant axes or factors obtainable in the most common practical case when n > m > nY (where n is the number of rows, and m the number of columns of the input data matrix).

Linear Discriminant Analysis
is the 2-group case of MDA.   It optimally separates two groups, using the Mahalanobis metric or generalized distance    It also gives the same linear separating decision surface as Bayesian maximum likelihood discrimination in the case of equal class covariance matrices.

K-NNs Discriminant Analysis
Non-parametric (distribution-free) methods dispense with the need for assumptions regarding the probability density function. They have become very popular especially in the image processing area. The K-NNs method assigns an object of unknown affiliation to the group to which the majority of its K nearest neighbours belongs.

Advantages and disadvantages of the methods studied are as follows:
  • Analytical simplicity or computational reasons may lead to initial consideration of linear discriminant analysis or the NN-rule.
  • Linear discrimination is the most widely used in practice. Often the 2-group method is used repeatedly for the analysis of pairs of multigroup data (yielding $ { {k(k-1)} \over 2 }$ decision surfaces for k groups).
  • To estimate the parameters required in quadratic discrimination more computation and data is required than in the case of linear discrimination. If there is not a great difference in the group covariance matrices, then the latter will perform as well as quadratic discrimination.
  • The k-NN rule is simply defined and implemented, especially if there is insufficient data to adequately define sample means and covariance matrices.
  • MDA is most appropriately used for feature selection. As in   the case of PCA, we may want to focus on the variables used in order to investigate the differences between groups; to create synthetic variables which improve the grouping ability of the data; to arrive at a similar objective by discarding irrelevant variables; or to determine the most parsimonious variables for graphical representational purposes.

No comments:

Post a Comment