PKBoost Logo - Adaptive Gradient Boosting

PKBOOST

Adaptive Gradient Boosting for a Changing World

Built in Rust. Tuned by Shannon. Designed for Drift.

Quick Start

python
from pkboost import PKBoostClassifier

model = PKBoostClassifier.auto()
model.fit(X_train, y_train)
predictions = model.predict_proba(X_test)

Benchmark Results

Credit Card Fraud (0.2% positive class)

PKBoost 87.8% PR-AUC
LightGBM 79.3% PR-AUC
XGBoost 74.5% PR-AUC

PKBoost: 87.8% • LightGBM: 79.3% • XGBoost: 74.5%

87.8%

PR-AUC on Fraud Detection

2s

Adaptation Time

Auto

Parameter Tuning

Design Goals

Traditional Limitations

  • Manual retraining on drift
  • Poor minority class recall
  • Extensive hyperparameter tuning
  • Performance degradation under drift

PKBoost Approach

  • Automatic adaptation via metamorphosis
  • Shannon entropy for rare patterns
  • Auto-tuned from data statistics
  • Stable performance under drift

Concept Drift

Adapts to distribution shifts without manual intervention through automatic metamorphosis.

Class Imbalance

Shannon entropy guidance for improved minority class detection in imbalanced datasets.

Auto-Tuning

Hyperparameters automatically configured based on dataset characteristics.

Use Cases

Fraud Detection

Credit cards, insurance claims, transactions

Medical Diagnosis

Rare disease detection, risk assessment

Cybersecurity

Intrusion detection, malware classification

Anomaly Detection

Equipment failure, quality control