Pre-LLM machine learning — and still the right tool for huge swaths of real problems. Tabular prediction, computer vision, recommendations, fraud, forecasting. Cheap, fast, interpretable.
← Back to AI Landscape| Library | Best for |
|---|---|
| scikit-learn | The Swiss army knife — regression, classification, clustering, pipelines, all in clean Python. |
| XGBoost | Gradient-boosted trees — wins most tabular Kaggle competitions. |
| LightGBM | Microsoft's GBT — faster training, similar accuracy. |
| CatBoost | Yandex's GBT — best-in-class with categorical features. |
| PyTorch | The default deep-learning framework today; flexible, research-friendly. |
| TensorFlow / Keras | Google's DL framework; strong production tooling (TFX, TF Serving). |
| JAX | Composable function transforms for high-performance ML / research. |
| Hugging Face Transformers | Pre-trained model hub for NLP, vision, audio. |
| statsmodels | Classical statistics — linear models, hypothesis tests, time series. |
| Prophet / NeuralProphet | Forecasting library from Meta. |
| OpenCV | Computer vision toolkit (faces, edges, tracking). |
| Surprise / implicit / RecBole | Recommender systems. |
The bread & butter — predicting churn, fraud, prices, conversion from a row of features. Gradient-boosted trees (XGBoost / LightGBM / CatBoost) almost always win, with neural nets close behind on very large datasets.
import xgboost as xgb model = xgb.XGBClassifier(n_estimators=500, max_depth=6) model.fit(X_train, y_train) pred = model.predict_proba(X_test)[:, 1]
Tokenization, TF-IDF, word2vec / GloVe, named-entity recognition, topic modeling. Most of these still matter as building blocks — and BERT-family encoders remain the best choice for cheap, fast classification & embeddings.
ARIMA, Prophet, neural forecasting (N-BEATS, Temporal Fusion Transformer). Most production forecasting still combines a tree model on lagged features with seasonal decomposition.
| Task | Metric |
|---|---|
| Regression | RMSE, MAE, R² |
| Binary classification | AUC-ROC, precision, recall, F1, log-loss |
| Multi-class | Accuracy, macro F1, confusion matrix |
| Ranking / RecSys | NDCG, MRR, MAP, hit-rate |
| Forecasting | MAPE, sMAPE, MASE |
| Detection | mAP @ IoU thresholds |