Image Similarity Using Logistic Regression
Meesala. Sai Kumar Computer Science & Engineering
Guru Nanak Institution Technical Campus Telangana, India meesalasaikumar1724@gmai l.com
Mohammad Nayeem Computer Science & Engineering
Guru Nanak Institution Technical Campus Telangana, India Mohammadnayeem1264@gmai l.com
J. Yamuna
Computer Science
& Engineering
Guru Nanak Institution Technical Campus Telangana, India
jimidiyamuna@gmail.com
Mrs. B. Surekha
(Assistant Professor)
Computer Science &
Engineering
Guru Nanak Institution Technical Campus Telangana, India
Surekhait21@gmail.com
ABSTRACT –
Many machine learning algorithms, such as kernel machines, nearest neighbors, clustering, and anomaly detection, rely on distances or similarities to identify patterns in data. Before using these similarities to train a model, it is crucial to ensure they reflect meaningful relationships within the data. In this paper, we propose enhancing the interpretability of these similarities by augmenting them with explanations. To achieve this, we introduce Logistic Regression & Haar Cascade, a scalable and theoretically sound method designed to systematically decompose the output of pre-trained deep similarity models for pairs of input features. Our approach can be viewed as a composition of regression- based explanations, which previous research has shown to effectively scale to highly nonlinear models. Through extensive experiments, we demonstrate that Logistic Regression consistently provides robust explanations for complex similarity models. We also apply our method to a challenge in digital humanities: evaluating
the similarity between historical documents, such as astronomical tables. In this context, Logistic Regression & Haar Cascade offers valuable insights and enhances the interpretability of a specialized, highly engineered similarity model.
Key words
Logistic Regression, Haar Cascade, Image Similarity, Machine Learning, Explainable AI, Deep Learning, Object Detection, Convolutional Neural Networks (CNN), Feature Extraction, Data Preprocessing.