On Computational Issues of Semi-Supervised Local Fisher Discriminant Analysis
スポンサーリンク
概要
- 論文の詳細を見る
Dimensionality reduction is one of the important preprocessing steps in practical pattern recognition. SEmi-supervised Local Fisher discriminant analysis (SELF) — which is a semi-supervised and local extension of Fisher discriminant analysis — was shown to work excellently in experiments. However, when data dimensionality is very high, a naive use of SELF is prohibitive due to high computational costs and large memory requirement. In this paper, we introduce computational tricks for making SELF applicable to large-scale problems.
- 2009-05-01
著者
関連論文
- Statistical active learning for efficient value function approximation in reinforcement learning (ニューロコンピューティング)
- Improving the Accuracy of Least-Squares Probabilistic Classifiers
- Improving the Accuracy of Least-Squares Probabilistic Classifiers
- A New Meta-Criterion for Regularized Subspace Information Criterion
- Approximating the Best Linear Unbiased Estimator of Non-Gaussian Signals with Gaussian Noise
- A new algorithm of non-Gaussian component analysis with radial kernel functions (Special issue: Information geometry and its applications)
- Methods of cross-domain object matching (情報論的学習理論と機械学習)
- Multi-task learning with least-squares probabilistic classifiers (パターン認識・メディア理解)
- Multi-task learning with least-squares probabilistic classifiers (情報論的学習理論と機械学習)
- Adaptive importance sampling with automatic model selection in value function approximation (ニューロコンピューティング)
- Analytic Optimization of Adaptive Ridge Parameters Based on Regularized Subspace Information Criterion(Neural Networks and Bioengineering)
- Adaptive Ridge Learning in Kernel Eigenspace and Its Model Selection
- On Computational Issues of Semi-Supervised Local Fisher Discriminant Analysis
- Recent Advances and Trends in Large-Scale Kernel Methods
- Syntheses of New Artificial Zinc Finger Proteins Containing Trisbipyridine-ruthenium Amino Acid at The N-or C-terminus as Fluorescent Probes
- Analytic Optimization of Shrinkage Parameters Based on Regularized Subspace Information Criterion(Neural Networks and Bioengineering)
- Constructing Kernel Functions for Binary Regression(Pattern Recognition)
- Optimal design of regularization term and regularization parameter by subspace information criterion
- Information-maximization clustering: analytic solution and model selection (情報論的学習理論と機械学習)
- New feature selection method for reinforcement learning: conditional mutual information reveals implicit state-reward dependency (情報論的学習理論と機械学習)
- Least Absolute Policy Iteration-A Robust Approach to Value Function Approximation
- Independent component analysis by direct density-ratio estimation (ニューロコンピューティング)
- A New Meta-Criterion for Regularized Subspace Information Criterion(Pattern Recognition)
- Spectral Methods for Thesaurus Construction
- Adaptive importance sampling with automatic model selection in reward weighted regression (ニューロコンピューティング)
- SERAPH: semi-supervised metric learning paradigm with hyper sparsity (情報論的学習理論と機械学習)
- Analysis and improvement of policy gradient estimation (情報論的学習理論と機械学習)
- Direct density-ratio estimation with dimensionality reduction via hetero-distributional subspace analysis (情報論的学習理論と機械学習)
- Output divergence criterion for active learning in collaborative settings (数理モデル化と問題解決・バイオ情報学)
- Estimation of squared-loss mutual information from paired and unpaired samples (情報論的学習理論と機械学習)
- Dependence minimizing regression with model selection for non-linear causal inference under non-Gaussian noise (情報論的学習理論と機械学習)
- Canonical dependency analysis based on squared-loss mutual information (情報論的学習理論と機械学習)
- Artist agent A[2]: stroke painterly rendering based on reinforcement learning (パターン認識・メディア理解)
- Artist agent A[2]: stroke painterly rendering based on reinforcement learning (情報論的学習理論と機械学習)
- Generalization Error Estimation for Non-linear Learning Methods(Neural Networks and Bioengineering)
- Improving Precision of the Subspace Information Criterion(Neural Networks and Bioengineering)