Ranking Metrics
These functions are helpful in evaluation of ranking problem like as in case of Recommendation Systems, where you need to rank your recommendations based on the predicted values.
Metrics.avg_precision — Functionavg_precision(y_rec, y_rel, k = 10)Evaluates how much of the relevant documents are concentrated in the highest ranked predictions.
Calculated as ∑(Recall@i - Recall@i-1)* Precision@i for i = (1, 2, 3....k)
Here, y_rec are predicted probabilities for recommendation and y_rel defines as 1 if particular result is relevant, else 0. The shape of y_rec and y_rel are expected to be (1, N_elements)
Metrics.ranking_stats_k — Functionranking_stats_k(y_rec, y_rel, k = 10)Evaluates the relevancy of top k recommendations using precison@k, recall@k and f1_score@k. Returns result as a Dict.
Here, y_rec are predicted probabilities for recommendation and y_rel defines as 1 if particular result is relevant, else 0. The shape of y_rec and y_rel are expected to be (1, N_elements).<br>
precison_kis evaluated asRecommended_items_that_are_relevant / Total_Recommended_items.recall_lis evaluated asRecommended_items_that_are_relevant / Total_Relevant_items.f1_kis evaluated as2 * Recommended_items_that_are_relevant / (Total_Recommended_items + Total_Relevant_items).