This page describes the ranking process used by The Apache Software Foundation to evaluate mentee proposals.
The ranking process is designed to ensure the most promising looking applicants are assigned a mentor with an ASF project. The goal is not to identify the "best" projects for the ASF, but the best mentees that have applied to the ASF.
Remember, mentoring is all about the mentee and open source as a whole, it is not about the mentors or their projects.
Any ASF committer can help with the ranking, all you need to do is register as a potential mentor (you don't actually have to become a mentor).
The instructions below described ranking proposals using the Google Summer of Code application.
Projects are ranked using a scoring system as follows:
Please only offer to mentor a single project unless discussed on email@example.com first. We encourage multiple mentors per project though.
Other community members are free to comment on proposals as they see fit. In the re-evaluation phase (see below) admins will use comments provided by mentors and community members.
All scoring is done in the Google Docs spreadsheet that was sent to firstname.lastname@example.org. Every Score must be followed by an explanation. DO NOT ADD SCORES IN MELANGE! Use integer numbers only!
In this phase mentors evaluate the proposal. Mentors are asked to add an explanation for each score added using the criteria below, this helps admins in the later stages of evaluation.
Please do not add scores outside these criteria
Any questions can be addressed to the admins on email@example.com
Did the student demonstrate that they have the required skills to start the project and learn what's needed along the way? And how happy are we with the results? (0-4 points)
How do we rate the quality of the student's proposal, in terms of technical relevance, completeness, awareness of issues, realistic planning (less copy and paste)? (0-4 points)
How does the mentor rate the student's chances of success, based on an in-person (face-to-face, video, audio, IRC, email) interview? (0-3 points, or 0-1 if email interview only)
How do we rate the originality of the student's contributions to the proposal? (0-2 points, 0 if mentor provides proposal, 1 if student builds proposal from suggested ideas, 2 if proposal is original idea from student)
Did student work with the project community, on our public mailing lists, to create the proposal? (0-2 points)
Did student explicitly indicate how much time they are going to dedicate to GSoC for the project's duration, and mention any other committments that could conflict with the project (holidays, course work, etc.)? And are we happy with this commitment (it should be their major occupation during the summer)? (0-2 points)
Did student respond quickly to mentor inquiries? (0-2 points)
Are you or your student affiliated in any way (e.g. share(d) the same employer or will do so in the foreseeable future, are you his or her adviser at university, personal relationships, etc.)? (yes/no)
Other community members are free to comment on proposals inside Melange as they see fit. However, DO NOT ADD A SCORE TO THE PROPOSAL. In the re-evaluation phase (see below) admins will use comments provided by mentors and community members.
At the end of the selection phase proposals to be considered for support will have a score of between 0 and +19, and should only have scores from the assigned mentor and admins (at this stage admin scores are only corrections). The admins will check in during the selection phase periodically to look for problems by scanning the scores (i.e. anything over +19 has an error).
At the end of this selection phase we know the minimum score a proposal must have in order to be selected (i.e. if we have 30 slots, projects with a score at or above the project ranked 30th will go forward to the next phase). Before progressing to the next phase admins will read all proposals either above or near the minimum score and ensure that there are no errors in this phase of ranking.
In this phase admins evaluate all proposals above the required score and others that are just below it. We check those just below the required score to ensure that the scores have been equally assigned across all proposals (some mentors are more generous than others).
The admins will transfer the scores to melange.
At the end of this phase there is a short window for the community to verify the admins work before the final selection is made.
In general we much prefer new mentees to repeat mentees. The goal is to enable new people to learn about open source software development.
In past years we have seen some consistent patterns in quality proposals:
Towards the end of the ranking period Google will announce how many slots the ASF has been given. Shortly after that we will ask ASF committers to stop ranking and the ASF GSoC admins will work to adjust the rankings to ensure that the right number of projects are ranked above the cut-off line.
In this last stage we avoid adjusting community rankings, however, in past years there has been a cluster of students with equal rankings around the cut-off point. In this case the admins will cast a final vote on those projects to ensure that Google can assign awards as appropriate.
For example, imagine that 7 is the number of points that causes the clustering, and we have the following projects ranked at 7:
Now imagine the cut-off point for selection is currently at the third position (C).
The admins need to look a these projects and ensure they are ordered in the webapp to make the most appropriate appear at the top of the list. This ensures that they are most likely to be awarded a slot. So we may end up with:
Unfortunately this process needs to happen very quickly and there is no time for community discussion around these points. Therefore it is important that mentors and mentees provide as much information in the previous ranking stages as possible.
The problem becomes a little more complicated when we remember that the cut-off point may move as well, so Google may give us one extra slot. So we need to ensure that all rank 7 projects are correctly ordered, not just the best.
Admins work through all applications adding a +8 to all those ranked above the cut off score. This has the effect of increasing the "space" we have to work in re-ranking the clustered projects.
Admins then work through all the clustered applications adding scores as defined above (with a comment justifying the ranking).
If there is still a cluster of projects around the cut-off point then the lead admin looks at the mentors and projects involved. If a project/mentor which already has a higher scored application, the lead admin reduces the score for this one (i.e. try and spread the love)
It's not practical to ask the community to comment at this point as we are on a tight deadline. We ask that the community trusts the assigned admins.
Inevitably there will be one or two students who have been accepted in multiple organizations. This is resolved during an IRC meeting with Google admins during the final hours of ranking. In these cases the admins will attempt to contact mentors and students, but we sometimes have to make a judgment call on how to best resolve such conflicts, as things happen quite fast during this meeting.