I made several LDAs using the scikit-learn LDA function, and in my graphs I noticed that there is a non-zero correlation between LDs.
from sklearn.lda import LDA sklearn_lda = LDA(n_components=2) transf_lda = sklearn_lda.fit_transform(X, y)
This is very important, so I came back and used the Iris dataset as a reference. I also found in the scikit documentation the same non-zero LDA correlation graph that I could reproduce.
Anyway, to give you an overview of what it looks like
- The plot is in the upper left corner: there is clearly something wrong here.
- The plot in the lower left corner: this is the source data, not the right approach, but one attempt to replicate scikit resuls
- Charts in the upper and lower right directions: this is how it should look.

I put the code in IPython notebook if you want to take a look at it and try it for yourself.
The scikit documentation, which is consistent with (incorrect), leads to the top left: http://scikit-learn.org/stable/auto_examples/decomposition/plot_pca_vs_lda.html
LDA in R, which is shown in the lower right: http://tgmstat.wordpress.com/2014/01/15/computing-and-visualizing-lda-in-r/
python scikit-learn r lda
Sebastian
source share