Shoe prints are commonly found at the scene of a crime and can sometimes help link a suspect to the scene. Because prints tend to be partially observed or smudgy, comparing crime scene prints with reference images from a putative shoe can be challenging. Footwear examiners rely on guidelines such as those published by SWGTREAD  to visually assess the similarity between two or more footwear impressions, one reason being that reliable, quantitative methods have yet to be validated for use in real cases. To help in the development of such methods, we created a study dataset of images of outsole impressions that shared class characteristics and degree of wear and that were subject to a specific type of degradation. We also propose a method to quantify the similarity between two outsole images that extends the capabilities of MC-COMP . The proposed method is composed of three steps; (1) extracting image descriptors, (2) aligning images using the maximum clique, (3) calculating similarity values using two different classifiers; (a) degree of overlap between the two images, and (b) a score produced by a random forest. To explore the performance of the algorithm we propose, we compared degraded, crime scene-like images to high-quality reference images produced by the same or by different shoes. Even though comparisons involved matches or very close non-matches, and one of the images was blurry, the algorithm shows good source classification performance.