Research Article
BibTex RIS Cite

Causal Image Processing: A Mechanism-Based Framework for Secure Visual Intelligence

Year 2026, Volume: 15 Issue: 1, 1 - 24, 19.03.2026
https://doi.org/10.55859/ijiss.1826477
https://izlik.org/JA76LH58XT

Abstract

This study introduces Causal Image Processing (CIP), a mechanism based conceptual framework designed to address fundamental weaknesses of contemporary deep learning systems in security critical visual intelligence. Unlike conventional convolutional networks, vision transformers, or diffusion models that rely on observational correlations, CIP models the image formation process through four independent causal mechanisms: content, domain conditions, sensor characteristics, and identity factors. These mechanisms are formalized within a directed acyclic graph that provides a structured representation separating the physical and semantic processes underlying image generation. CIP integrates three forms of inference within a unified architecture: predictive inference, interventional inference, and counterfactual inference. Mechanism faithful representations enforce invariance to sensor and domain variations, while counterfactual reasoning enables principled evaluation of identity consistency under hypothetical acquisition conditions. The framework also introduces learning principles based on mechanism fidelity, sparse intervention sensitivity, and causal invariance, and defines evaluation criteria centered on counterfactual consistency and tamper resistance. Overall, CIP offers a theoretical foundation for developing robust, explainable, and tamper resistant vision systems capable of operating reliably across heterogeneous sensors, environments, and acquisition conditions. The framework establishes a mechanism centered alternative to correlation driven deep learning pipelines and provides a pathway for next generation secure visual intelligence.

References

  • R. Geirhos, J.-H. Jacobsen, C. Michaelis, R. Zemel, W. Brendel, M. Bethge, and F. A. Wichmann, “Shortcut learning in deep neural networks,” Nature Machine Intelligence, vol. 2, no. 11, pp. 665–673, 2020.
  • B. Recht, R. Roelofs, L. Schmidt, and V. Shankar, “Do imagenet classifiers generalize to imagenet?” in International Conference on Learning Representations (ICLR), 2019, arXiv:1902.10811.
  • M. Raghu, T. Unterthiner, S. Kornblith, C. Zhang, and A. Dosovitskiy, “Do vision transformers see like convolutional neural networks?” in Advances in Neural Information Processing Systems, vol. 34, 2021.
  • B. Schölkopf, F. Locatello, S. Bauer, N. R. Ke, N. Kalchbrenner, A. Goyal, and Y. Bengio, “Toward causal representation learning,” Proceedings of the IEEE, vol. 109, no. 5, pp. 612–634, 2021.
  • T. Burge, “Marr’s theory of vision,” in Modularity in Knowledge Representation and Natural-Language Understanding, E. Lepore, Ed. Cambridge: MIT Press, 1989, pp. 201–230.
  • K. Zhang, Q. Sun, C. Zhao, and Y. Tang, “Causal reasoning in typical computer vision tasks,” arXiv preprint arXiv:2307.13992, 2023.
  • M. Arjovsky, L. Bottou, I. Gulrajani, and D. Lopez-Paz, “Invariant risk minimization,” arXiv preprint arXiv:1907.02893, 2019.
  • I. Gulrajani and D. Lopez-Paz, “In search of lost domain generalization,” arXiv preprint arXiv:2007.01434, 2021.
  • S. Sagawa, P. W. Koh, T. B. Hashimoto, and P. Liang, “Distributionally robust neural networks for group shifts: On the importance of regularization for worst-case generalization,” in International Conference on Learning Representations (ICLR), 2020, arXiv:1911.08731.
  • C. Zhang, S. Bengio, M. Hardt, B. Recht, and O. Vinyals, “Understanding deep learning requires rethinking generalization,” in International Conference on Learning Representations (ICLR), 2017, arXiv:1611.03530.
  • D. Liu, Y. Qiao, W. Liu, Y. Lu, Y. Zhou, T. Liang, Y. Yin, and J. Ma, “Causal3d: A comprehensive benchmark for causal learning from visual data,” arXiv preprint arXiv:2503.04852, 2025.
  • W. Li, Z. Li, X. Yang, and H. Ma, “Causal-vit: Robust vision transformer by causal intervention,” Engineering Applications of Artificial Intelligence, vol. 126, p. 107123, 2023.
  • W. Xie, X.-H. Li, C. C. Cao, and N. L. Zhang, “Vit-cx: Causal explanation of vision transformers,” arXiv preprint arXiv:2211.03064, 2022.
  • R. Sundararajan, S. Sahu, A. Namboodiri, and A. Chattopadhyay, “Trace: Training calibration-based counterfactual explainers,” Scientific Reports, vol. 12, p. 556, 2022.
  • B. Schölkopf, “Causality for machine learning,” arXiv preprint arXiv:1911.10500, 2019.
  • M. Kocaoglu, C. Snyder, A. G. Dimakis, and S. Vishwanath, “Causalgan: Learning causal implicit generative models with adversarial training,” in International Conference on Learning Representations (ICLR), 2018, arXiv:1709.02023.
  • M. Yang, F. Liu, Z. Chen, X. Shen, J. Hao, and J. Wang, “Causalvae: Disentangled representation learning via neural structural causal models,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 9593–9602.
  • I. Khemakhem, R. Monti, R. Leech, and A. Hyvarinen, “Causal autoregressive flows,” in Proceedings of the 24th International Conference on Artificial Intelligence and Statistics (AISTATS), 2021, pp. 3550–3568, arXiv:2011.02268.
  • A. Alrashidi, A. Alotaibi, M. Hussain, H. AlShehri, H. A. AboAlSamh, and G. Bebis, “Cross-sensor fingerprint matching using siamese network and adversarial learning,” Sensors, vol. 21, no. 11, p. 3657, 2021.
  • J. Pearl, Causality: Models, Reasoning, and Inference. Cambridge University Press, 2009.
  • H. Barr, K. Harrington, R. Sharpe, and C. Bruss, “Sharpshooter: Counterfactual explanations via latent projection,” arXiv preprint arXiv:2112.00890, 2021.
  • S. Lyu, Q. Zhao, Z. Zhou, M. Li, Y. Zhou, D. Yao, G. Cheng, H. Zhou, and Z. Shi, “Deep learning based domain adaptation methods in remote sensing: A comprehensive survey,” arXiv preprint arXiv:2510.15615, 2025.
  • C.-H. H. Yang, Y.-C. Liu, P.-Y. Chen, X. Ma, and Y.-C. J. Tsai, “When causal intervention meets adversarial examples and image masking for deep neural networks,” in 2019 IEEE International Conference on Image Processing (ICIP). IEEE, 2019, pp. 3811–3815.
  • M. Geiger, E. Jonas, and M. Holzleitner, “Mechanistic interpretability for ai safety: A review of methods and future directions,” arXiv preprint arXiv:2106.02826, 2021.
  • A. Goyal, A. C´ezar, M. Liu, J. Peters, and Y. Bengio, “Discovering dynamics with sparse mechanism shift modeling,” arXiv preprint arXiv:2106.05864, 2021.
  • K. Kuang, Y. Zhou, K. Sun, R. Zhang, L. Song, and H. Xiong, “Causalbench: A large-scale benchmark for causal reasoning in vision,” arXiv preprint arXiv:2303.08228, 2023.
  • A. Trockman and J. Z. Kolter, “Patches are all you need?” arXiv preprint arXiv:2201.09792, 2022.
  • J. Ho, A. Jain, and P. Abbeel, “Denoising diffusion probabilistic models,” in Advances in Neural Information Processing Systems, vol. 33, 2020, pp. 6840–6851, arXiv:2006.11239.
  • Y. Song, P. Dhariwal, M. Chen, and I. Sutskever, “Consistency models,” in Proceedings of the 40th International Conference on Machine Learning (ICML), 2023, arXiv:2303.01469.
  • P. Dhariwal and A. Nichol, “Diffusion models beat gans on image synthesis,” in Advances in Neural Information Processing Systems, vol. 34, 2021, pp. 8780–8794, arXiv:2105.05233.
  • W. Peebles and S. Xie, “Scalable diffusion models with transformers,” arXiv preprint arXiv:2212.09748, 2023.
  • G. Kotek, S. Weingartner, and K. T. Block, “From signal-based to comprehensive magnetic resonance imaging: Algebraic description of magnetization during transient response,” Scientific Reports, vol. 11, p. 17216, 2021.
There are 32 citations in total.

Details

Primary Language English
Subjects Digital Forensics, Data Security and Protection
Journal Section Research Article
Authors

Bilgehan Arslan 0000-0002-5160-4408

Submission Date November 19, 2025
Acceptance Date January 9, 2026
Publication Date March 19, 2026
DOI https://doi.org/10.55859/ijiss.1826477
IZ https://izlik.org/JA76LH58XT
Published in Issue Year 2026 Volume: 15 Issue: 1

Cite

IEEE [1]B. Arslan, “Causal Image Processing: A Mechanism-Based Framework for Secure Visual Intelligence”, IJISS, vol. 15, no. 1, pp. 1–24, Mar. 2026, doi: 10.55859/ijiss.1826477.