There is provided a method comprising: providing two anatomical images of a target individual, each captured at a unique orientation of the target individual, inputting first and second anatomical images respectively into a first and second convolutional neural network (CNN) of a classifier to respectively output first and second feature vectors, inputting a concatenation of the first and second feature vectors into a fully connected layer of the classifier, and computing an indication of distinct visual finding(s) present in the anatomical images by the fully connected layer, wherein the statistical classifier is trained on a training dataset including two anatomical images of each respective sample individual, each image captured at a respective unique orientation of the target individual, and a tag created based on an analysis that maps respective individual sentences of a text based radiology report to one of multiple indications of visual findings.