Orion-MSP Logo

Orion-MSP: Multi-Scale Sparse Attention for Tabular In-Context Learning

Orion-MSP is a tabular foundation model for in-context learning. It uses multi-scale sparse attention and Perceiver-style memory to process tabular data at multiple granularities, capturing both local feature interactions and global dataset-level patterns.

Key Features

  • Multi-Scale Sparse Attention: Processes features at three levels (scales 1, 4, 16) using windowed, global, and random attention patterns, reducing quadratic complexity to near-linear.
  • Hierarchical Feature Understanding: Captures patterns from individual cells to feature groups through scale-aware attention.
  • Perceiver-Style Memory: Cross-component memory that compresses dataset information for efficient processing across samples
  • Memory-Efficient: Block-sparse masking enables efficient processing of large tabular datasets
  • Scikit-learn Compatible: Drop-in replacement with .fit() and .predict() methods

Architecture

Orion-MSP consists of four main components:

  • Column-wise Embedding: Distribution-aware feature embeddings using Induced Set Attention Blocks (ISAB)
  • Multi-Scale Row Interaction: Sparse attention with windowed, global, and random patterns across multiple scales
  • Cross-Component Memory: Perceiver-style memory for efficient dataset-level context
  • Dataset-wise ICL: Enhanced predictor leveraging enriched representations for few-shot tabular classification

Performance

Performance comparison across three benchmark suites—TALENT, OpenML-CC18, and TabZilla. Ranks are mean ranks based on accuracy (lower is better). Metrics: ACC = Accuracy, F1 = Weighted F1. 1st; 2nd.
Models All TALENT OpenML-CC18 TabZilla
Rank RankACCF1 RankACCF1 RankACCF1
XGBoost 6.70 6.020.84030.8360 5.890.85580.8537 6.070.86120.8326
CatBoost 6.43 5.570.83360.8259 6.250.85880.8520 7.130.85790.8384
Random Forest 7.38 6.150.82850.8209 6.360.85470.8497 8.420.83580.8399
LightGBM 6.78 6.110.83310.8245 6.180.85810.8493 5.250.86180.8211
TabICL 4.96 4.090.84710.8379 4.690.86670.8623 5.890.87340.8698
OrionBiX 5.37 4.590.83460.8260 4.980.86530.8596 4.890.87280.8628
OrionMSP 3.58 3.260.84610.8360 4.120.87220.8676 3.840.88210.8786
TabPFN 4.61 3.720.85140.8412 4.760.87140.8663 4.860.87520.8716
Mitra 11.77 10.380.39210.2868 10.520.36140.2522 11.210.31520.1830
ContextTab 9.70 9.840.54740.4596 6.280.86390.8581 7.130.83890.8334
TabDPT 5.42 5.190.84080.8318 4.640.86720.8625 3.940.88140.8775

Orion-MSP is the most consistent top performer across all three benchmarks, achieving the best overall rank.

  • On TALENT, it ranks *1 overall, while TabPFN edges the highest ACC/F1 by a hair.
  • On OpenML-CC18, Orion-MSP attains the top ACC/F1 (0.8722/0.8676), narrowly ahead of TabPFN and TabDPT.
  • On TabZilla, it leads with the highest ACC/F1 and the best rank.
  • Classical baselines (XGBoost/LightGBM/CatBoost/RF) trail noticeably, highlighting Orion-MSP’s robustness across diverse tabular tasks.
Performance variation by dataset size across all benchmark suites. Rank = mean rank by accuracy (lower is better). ACC = Accuracy; F1 = Weighted F1. Size buckets: Small (<1K), Medium (1K–10K), Large (>10K).
Models Small (<1K) Medium (1K–10K) Large (>10K)
RankACCF1 RankACCF1 RankACCF1
XGBoost 7.700.81680.7964 6.880.83630.8314 5.410.89690.8920
CatBoost 7.880.81240.7935 6.470.83400.8264 5.480.87970.8733
Random Forest 8.550.79880.8187 7.160.82850.8221 7.300.86940.8628
LightGBM 7.800.81430.7789 6.940.83140.8226 5.630.88270.8764
TabICL 6.040.83010.8338 4.770.84860.8398 4.610.88020.8743
OrionBiX 6.320.83300.8150 5.480.83480.8260 4.420.87290.8670
OrionMSP 5.930.82320.8194 3.700.84940.8402 3.040.88430.8768
TabPFN 6.500.83250.8131 3.810.85570.8462 5.730.87830.8713
Mitra 13.880.43340.3236 11.590.36000.2553 11.110.38370.2754
ContextTab 9.600.75780.7363 9.520.62100.5566 10.220.63880.5638
TabDPT 5.480.83330.8271 5.400.84240.8339 5.260.88310.8765

OrionMSP is the most consistent top-ranked model as data grows (especially Medium/Large), while TabPFN peaks on Medium and GBDTs (e.g., XGBoost) catch up in raw ACC/F1 on Large.

Performance vs. feature dimensionality. Rank = mean accuracy rank (lower is better). ACC = Accuracy; F1 = Weighted F1. Groups: Narrow (<10), Medium (10–100), Wide (>100). 1st ; 2nd within each group.
Models Narrow (<10) Medium (10–100) Wide (>100)
RankACCF1 RankACCF1 RankACCF1
XGBoost 6.770.82220.8159 6.900.84820.8410 4.790.91400.9039
CatBoost 5.630.81450.8067 6.880.84410.8344 5.500.91570.9084
Random Forest 7.150.80050.7044 7.440.84100.8235 7.520.90340.8936
LightGBM 6.150.81280.7907 6.920.84580.8326 7.470.89990.8908
TabICL 5.140.82080.8119 4.610.86270.8549 6.460.91010.8936
OrionBiX 4.640.81120.8043 5.460.85100.8417 6.730.88590.8849
OrionMSP 3.760.83940.8314 4.090.85720.8478 5.690.88600.8837
TabPFN 5.300.81870.8092 4.070.86760.8589 6.1410.91290.9111
Mitra 11.250.37370.2683 11.840.38860.2781 13.030.25210.1497
ContextTab 9.520.63910.5719 9.590.64800.5843 10.970.60170.5651
TabDPT 4.660.82620.8189 5.450.85660.8483 7.230.88450.8820

OrionMSP excels on narrow and stays strong on medium width, while TabPFN dominates medium-width features and GBDTs (XGBoost/CatBoost) shine on wide feature spaces.

Usage

from orion_msp.sklearn import OrionMSPClassifier

# Initialize and use
clf = OrionMSPClassifier()
clf.fit(X_train, y_train)
predictions = clf.predict(X_test)

This code will automatically download the pre-trained model from Hugging Face and use a GPU if available.

Installation

From the source

Option 1: From the local clone

cd orion-msp
pip install -e .

Option 2: From the Git Remote

pip install git+https://github.com/Lexsi-Labs/Orion-MSP.git

Citation

If you use Orion-MSP, please cite our paper:

@article{bouadi25orionmsp,
  title={Orion-MSP: Multi-Scale Sparse Attention for Tabular In-Context Learning},
  author={Mohamed Bouadi and Pratinav Seth and Aditya Tanna and Vinay Kumar Sankarapu},
  year={2025}
  eprint={2511.02818},
  archivePrefix={arXiv},
  primaryClass={cs.AI},
  url={https://arxiv.org/abs/2511.02818}, 
}
Downloads last month
11
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Collection including Lexsi/Orion-MSP