We offer a benchmark suite together with an evaluation server, such that authors can upload their results and get a ranking. We offer a dataset that contains more than 25,000 pictures, including 15,403 images for training set, 5,000 images for validation set and 5,000 images for testing set. If you would like to submit your results, please follow the instructions on our submission page.
Note: We only display results with relatively detailed descriptions.
We use two human-centric metrics for multi-human parsing evaluation, which are initially reported by the MHP v1.0 paper. The two metrics are Average Precision based on part (APp) (%) and Percentage of Correctly parsed semantic Parts (PCP) (%).
All teams with successful submissions have a placeholder in the leaderboard, and the results of all teams will be released on 10 June. The winner of the challenge is the team with maximal numbers of top-1 ranking among all the five metrics (one in each columns). Ties are broken by the score of APp0.5.