Skip to content

AI-sandbox/iLTM

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

3 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

iLTM: Integrated Large Tabular Model

PyPI License Downloads Python Versions Hugging Face

iLTM is a foundation model for tabular data that integrates tree-derived embeddings, dimensionality-agnostic representations, a meta-trained hypernetwork, multilayer perceptron (MLP) neural networks, and retrieval. iLTM automatically handles feature scaling, categorical features, and missing values.

We release open weights of pre-trained model checkpoints that consistently achieve superior performance across tabular classification and regression tasks, from small to large and high-dimensional tasks.

iLTM architecture diagram

Install

iLTM is accessed through Python. You can install the package via pip:

pip install iltm

iLTM works on Linux, macOS and Windows, and can be executed on CPU and GPU, although GPU is highly recommended for faster execution.

Pre-trained model checkpoints are automatically downloaded from Hugging Face on first use. By default, checkpoints are stored in platform-specific cache directories (e.g., ~/.cache/iltm on Linux, ~/Library/Caches/iltm on macOS). You can specify where model checkpoints are stored by setting the ILTM_CKPT_DIR environment variable:

export ILTM_CKPT_DIR=/path/to/checkpoints

Note

The first call to iLTMRegressor or iLTMClassifier downloads the selected checkpoint. Later runs reuse the cached weights from ILTM_CKPT_DIR or the default cache location.

Tip

For interactive work on a local machine it is often worth pointing ILTM_CKPT_DIR to a fast local disk to avoid repeated downloads across environments.

Quick Start

iLTM is designed to be easy to use, with an API similar to scikit-learn.

from iltm import iLTMRegressor, iLTMClassifier

# Regression
reg = iLTMRegressor().fit(X_train, y_train)
y_pred = reg.predict(X_test)

# Classification
clf = iLTMClassifier().fit(X_train, y_train)
proba = clf.predict_proba(X_test)
y_hat = clf.predict(X_test)

# With time limit (returns partial ensemble if time runs out)
reg = iLTMRegressor().fit(X_train, y_train, fit_max_time=3600)  # 1 hour limit

Model Checkpoints

Available checkpoint names:

  • "xgbrconcat" (default): Robust preprocessing + XGBoost embeddings + concatenation
  • "cbrconcat": Robust preprocessing + CatBoost embeddings + concatenation
  • "r128bn": Robust preprocessing with 128-dim bottleneck
  • "rnobn": Robust preprocessing without bottleneck
  • "xgb": XGBoost embeddings only
  • "catb": CatBoost embeddings only
  • "rtr": Robust preprocessing with retrieval
  • "rtrcb": CatBoost embeddings with retrieval

You can also provide a local path to a checkpoint file.

Common key args:

  • checkpoint: checkpoint name or path to model file. Default "xgbrconcat".
  • device: torch device string. Default "cuda:0".
  • n_ensemble: number of generated predictors.
  • batch_size: batch size for weight prediction and inference.
  • preprocessing: "realmlp_td_s_v0" or "minimal" or "none".
  • cat_features: list of categorical column indices.
  • tree_embedding: enable GBDT leaf embeddings.
  • tree_model: "XGBoost_hist" or "CatBoost".
  • concat_tree_with_orig_features: concatenate original features with embeddings.
  • finetuning: end to end finetuning.
  • Retrieval: do_retrieval, retrieval_alpha, retrieval_temperature, retrieval_distance.

Regressor only:

  • clip_predictions: clip to train target range.
  • normalize_predictions: z-normalize outputs before unscaling.

Classifier only:

  • voting: "soft" or "hard".

Hyperparameter Optimization

iLTM performs best when you tune its hyperparameters.

Recommended search space

The package exposes a recommended search space via iltm.get_hyperparameter_search_space, a plain dictionary that maps hyperparameter names to small specs.

Tip

When running hyperparameter optimization with time constraints, you can use the fit_max_time parameter in fit() to limit training time per configuration. The model will return a partial ensemble if the time limit is reached.

The checkpoint parameter is part of this space. It is responsible for selecting one of the built in model checkpoints, which in turn sets other fields such as preprocessing, tree_embedding, and others.

The specification format is intentionally minimal so that it can be re-used in any hyperparameter optimization library or custom tuning procedure.

  • iltm.get_hyperparameter_search_space() gives you the canonical space definition.
  • iltm.sample_hyperparameters(rng) draws a single random configuration from that space for quick baselines and smoke tests.

Tip

sample_hyperparameters is mainly intended for quick baselines, smoke tests, or simple random search. For more serious tuning runs it is usually better to adapt the search space from get_hyperparameter_search_space into your optimization method of choice, and let that method decide which configurations to try.

Development

To run the tests:

pip install -e ".[dev]"
pytest tests/

Citation

If you use iLTM in your research, please cite our paper:

@article{bonet2025iltm,
  title={iLTM: Integrated Large Tabular Model},
  author={Bonet, David and Comajoan Cara, Marçal and Calafell, Alvaro and Mas Montserrat, Daniel and Ioannidis, Alexander G},
  journal={arXiv preprint arXiv:2511.15941},
  year={2025},
}

License

© Contributors, 2025. Licensed under the Apache-2.0 license.