✨ Dimensionality Reduction Techniques in Machine Learning
When dealing with high-dimensional datasets, models can become slow, prone to overfitting, and difficult to interpret. This is where dimensionality reduction comes in — the process of reducing the number of features while retaining as much useful information as possible.
📌 The Question
Which of the following are techniques for dimensionality reduction?
Options:
-
✅ PCA (Principal Component Analysis)
-
✅ t-SNE (t-distributed Stochastic Neighbor Embedding)
🌲 Explanation of Each Option
1. PCA (Principal Component Analysis) ✅
-
PCA is one of the most popular dimensionality reduction techniques.
-
It transforms correlated features into a set of uncorrelated principal components, ranked by the variance they capture.
-
Often used for visualization and noise reduction.
2. StandardScaler ❌
-
StandardScaler is not a dimensionality reduction technique.
-
It only normalizes the scale of features (mean = 0, variance = 1).
-
While scaling is important before applying PCA or regression, it doesn’t reduce dimensions.
3. Lasso Regression ✅
-
Lasso (Least Absolute Shrinkage and Selection Operator) adds L1 regularization.
-
It forces some coefficients to become exactly zero, effectively removing irrelevant features.
-
This makes it a feature selection method, which is a form of dimensionality reduction.
4. t-SNE (t-distributed Stochastic Neighbor Embedding) ✅
-
t-SNE is a nonlinear technique that projects high-dimensional data into 2D or 3D for visualization.
-
It preserves local similarities (points that are close in high dimensions remain close in low dimensions).
-
Extremely useful for visualizing clusters in high-dimensional data.
🚀 Key Takeaways
-
Dimensionality Reduction Techniques: PCA, Lasso Regression, t-SNE
-
Not Dimensionality Reduction: StandardScaler (it’s preprocessing, not feature reduction).
👉 Would you like me to merge this with the Decision Tree max leaf nodes blog into one combined blog post (like a "Machine Learning Interview Q&A" style article), or keep them as separate blog posts?
Comments
Post a Comment