Precision and Recall, as well as their combination in terms of FMeasure,are widely used measures in computer science and generally used to evaluate the overall performance of ontology matchers in fully automatic, unsupervised scenarios. In this paper, we investigate the case of supervised matching,where automatically created ontology alignments are verified by an expert. We motivate and describe this use case and its characteristics and discuss why traditional, F-measure based evaluation measures are not suitable to choose the best matching system for this task. Therefore, we investigate several alternative evaluation measures and propose the use of Precision@N curves as a means to assess different matching systems for supervised matching. We compare the ranking of ontology matchers from the last OAEI campaign using Precision@N curves to the traditional F-measure based ranking, and discuss means to combine matchers in a way that optimizes the user support in supervised ontology matching.