interview

Notes
3 key question in data visualization
Accuracy
Activation Function
Active Learning
Adaboost
AdaBoost vs. Gradient Boosting vs. XGBoost
AdaDelta
AdaGrad
Adam
Adjusted R-squared Value
Alternative Hypothesis
Amazon Leadership Principles
Area Under Precision Recall Curve (AUPRC)
AUC Score
Autoencoder for Denoising Images
Averaging in Ensemble Learning
Backward Feature Elimination
Bag of Words
Bagging
Batch Normalization
Bayes Theorem
Bayesian Optimization Hyperparameter Finding
Beam Search
Behavioral Interview
BERT
BERT Embeddings
Bias & Variance
Bidirectional RNN or LSTM
Binary Cross Entropy
Binning or Bucketing
Binomial Distribution
BLEU Score
Boosting
Byte Level BPE
Byte Pair Encoding (BPE)
Causal Language Modeling
Causality vs. Correlation
Central Limit Theorem
Chain Rule
Challenges of NLP
Character Tokenizer
CNN
Co-occurrence based Word Embeddings
Co-Variance
Collinearity
Conditional Probability
Confusion Matrix
Contextualized Word Embeddings
Continuous Bag of Words
Continuous Random Variable
Contrastive Learning
Contrastive Loss
Convex vs Nonconvex Function
Cosine Similarity
Count based Word Embeddings
Cross Entropy
Cross Validation
Crossed Feature
Curse of Dimensionality
Data Augmentation
Data Imputation
data visualization
DBScan Clustering
Debugging Deep Learning
Decision Tree
Decision Tree (Classification)
Decision Tree (Regression)
Decoder Only Transformer
Decoding Strategies
Density Sparse Data
Dependent Variable
Derivative
Differentiation
Differentiation of Product
Digit Dp
Dimensionality Reduction
Discrete Random Variable
Discriminative vs. Generative Models
Dropout
DS & Algo Interview
Dying ReLU
Dynamic Programming (DP) in python
Elastic Net Regression
ELMo Embeddings
Encoder Only Transformer
Ensemble Learning
Entropy
Entropy and Information Gain
Essential Visualizations
Estimated Mean
Estimated Standard Deviation
Estimated Variance
Euclidian Norm
Expected Value
Expected Value for Continuous Events
Expected Value for Discrete Events
Exploding Gradient
Exponential Distribution
Extrinsic Evaluation
F-Beta Score
F-Beta@K
F1 Score
False Negative Error
False Positive Rate
FastText Embedding
Feature Engineering
Feature Extraction
Feature Hashing
Feature Preprocessing
Feature Selection
Finding Co-relation between two data or distribution
Forward Feature Selection
Foundation Model
Gaussian Distribution
GBM
Genetic Algorithm Hyperparameter Finding
Gini Impurity
GloVe Embedding
Gradient
Gradient Boost (Classification)
Gradient Boost (Regression)
Gradient Boosting
Gradient Clipping
Gradient Descent
Graph Convolutional Network (GCN)
Greedy Decoding
Grid Search Hyperparameter Finding
Group Normalization
GRU
Gumbel Softmax
Handling Imbalanced Dataset
Handling Missing Data
Handling Outliers
Heapq (nlargest or nsmalles)
Hierarchical Clustering
Hinge Loss
Histogram
Homonym or Polysemy
How to Choose Kernel in SVM
How to combine in Ensemble Learning
How to prepare for Behavioral Interview
Huber Loss
Hyperparameters
Hypothesis Testing
Independent Variable
InfoNCE Loss
Internal Covariate Shift
interview
Interview Scheduling
Intrinsic Evaluation
Jaccard Distance
Jaccard Similarity
K Fold Cross Validation
K-means Clustering
K-means vs. Hierarchical
K-nearest Neighbor (KNN)
Kernel in SVM
Kernel Regression
Kernel Trick
KL Divergence
L1 or Lasso Regression
L1 vs. L2 Regression
L2 or Ridge Regression
Label Encoding
Layer Normalization
Leaky ReLU
Learning Rate Scheduler
LightGBM
Likelihood
Line Equation
Linear Regression
Log (Odds Ratio)
Log (Odds)
Log Scale
Log-cosh Loss
Logistic Regression
Logistic Regression vs. Neural Network
Loss vs. Cost
LSTM
Machine Learning Algorithm Selection
Machine Learning vs. Deep Learning
Majority vote in Ensemble Learning
Margin in SVM
Marginal Probability
matplotlib legend
Maximal Margin Classifier
Maximum Likelihood
Mean
Mean Absolute Error (MAE)
Mean Absolute Percentage Error (MAPE)
Mean Squared Error (MSE)
Mean Squared Logarithmic Error (MSLE)
Median
Merge K-sorted List
Merge Overlapping Intervals
Meteor Score
Min Max Normalization
Mini Batch SGD
ML Case Study or ML Design
ML Interview
ML System Design
Mode
Model Based vs. Instance Based Learning
MRR
Multi Class Cross Entropy
Multi Label Cross Entropy
Multi Layer Perceptron
Multicollinearity
Multivariable Linear Regression
Multivariate Linear Regression
Multivariate Normal Distribution
Mutual Information
N-gram Method
Naive Bayes
Negative Log Likelihood
Negative Sampling
Nesterov Accelerated Gradient (NAG)
Neural Network
Neural Network Normalization
Normal Distribution
Null Hypothesis
Odds
Odds Ratio
One Class Classification
One Class Gaussian
One Hot Vector
One vs One Multi Class Classification
One vs Rest or One vs All Multi Class Classification
Optimizers
Overcomplete Autoencoder
Overfitting
Oversampling
p-value
Padding in CNN
Parameter vs. Hyperparameter
PCA vs. Autoencoder
Pearson Correlation
Perceptron
Perplexity
Plots Compared
Polynomial Regression
Pooling
Population
Posterior Probability
Precision
Precision Recall Curve (PRC)
Precision@K
Principal Component Analysis (PCA)
Prior Probability
Probability Density Function
Probability Distribution
Probability Mass Function
Probability vs. Likelihood
Problem Solving Algorithm Selection
Pruning in Decision Tree
PyTorch Refresher
Questions to ask in a Interview?
Quintile or Percentile
Quotient Rule or Differentiation of Division
R-squared Value
Random Forest
Random Variable
Recall
Recall@K
Recommender System (RecSys)
Regularization
Reinforcement Learning
Reinforcement Learning from Human Feedback (RLHF)
Relational GCN
ReLU
Retrieval Metrics
RMSProp
RNN
ROC Curve
Root Mean Squared Error (RMSE)
Root Mean Squared Logarithmic Error (RMSLE)
ROUGE-L Score
ROUGE-LSUM Score
ROUGE-N Score
Saddle Points
Self-Supervised Learning
Semi-supervised Learning
Sensitivity
SentencePiece Tokenization
Sigmoid Function
Simple Linear Regression
Singular Value Decomposition (SVD)
Skip Gram Model
Soft Margin in SVM
Softmax
Softplus
Softsign
Some Common Behavioral Questions
Specificity
Splitting tree in Decision Tree
Stacking or Meta Model in Ensemble Learning
Standard deviation
Standardization
Standardization or Normalization
Statistical Significance
Stochastic Gradient Descent (SGD)
Stochastic Gradient Descent with Momentum
Stop Words
Stride in CNN
Stump
Sub-sampling in Word2Vec
Sub-word Tokenizer
Supervised Learning
Support Vector
Support Vector Machine (SVM)
Surprise
SVC
Swallow vs. Deep Learning
Tanh
Text Preprocessing
TF-IDF
Three Way Partioning
Time Complexity of ML Algos
Time Complexity of ML Models
Tokenizer
Training a Deep Neural Network
Triplet Loss
True Negative Rate
True Positive Rate
Type 1 Error vs. Type 2 Error
Undercomplete Autoencoder
Undersampling
Unigram Tokenization
Unsupervised Learning
Vanishing Gradient
Variance
Weight Initialization
Word Embeddings
Word Tokenizer
Word2Vec Embedding
WordPiece Tokenization
XGBoost