A way to make prediction models more precise and interpretable at the same time?

causal identification causal inference feature selection prediction models ml interpretability machine learning

An interesting and potentially useful combination of causal identification algorithms and prediction modeling in the machine learning pipeline.

Luděk Stehlík https://www.linkedin.com/in/ludekstehlik/
02-13-2025

I just came across an interesting paper by Ding, Zhang, & Bos (2018), who used automatic causal identification algorithms for feature selection and achieved better prediction performance with models based on these features than with models built in a more traditional way.

The authors specifically studied the experiences and choices of 3,293 visitors at a large theme park and attempted to use them to predict their Big Five personality characteristics. First, they applied two algorithms for causal identification—the PC algorithm and the Fast Greedy Equivalence Search algorithm—to narrow down the candidate DAGs that explain the data. Then, they used the outputs from these algorithms to identify features that were causally impacted by the personality characteristic of interest. Finally, they built prediction models using only these features.

The authors found that their ML pipeline, which incorporated causal identification for feature selection, outperformed baseline models in predicting individual characteristics (specifically, they used LASSO linear regression) as the baseline model, which also performs automatic feature selection). Beyond that, these models, according to the authors, provided more human-interpretable results. A win-win.

For me, these results are quite surprising, as I would expect that if prediction performance were the only criterion, a purely “correlational” approach without causal constraints would perform better. I’m curious how these alleged benefits generalize to other types of situations—such as cases with less noise in the data. Definitely worth trying in one of my future projects.

Does anyone have experience with this specific approach to feature selection? Feel free to share your insights.

Citation

For attribution, please cite this work as

Stehlík (2025, Feb. 13). Ludek's Blog About People Analytics: A way to make prediction models more precise and interpretable at the same time?. Retrieved from https://blog-about-people-analytics.netlify.app/posts/2025-02-13-causal-identification-and-feature-selection/

BibTeX citation

@misc{stehlík2025a,
  author = {Stehlík, Luděk},
  title = {Ludek's Blog About People Analytics: A way to make prediction models more precise and interpretable at the same time?},
  url = {https://blog-about-people-analytics.netlify.app/posts/2025-02-13-causal-identification-and-feature-selection/},
  year = {2025}
}