BOINC Credit System

Within the BOINC platform for volunteer computing, the BOINC Credit System helps volunteers keep track of how much CPU time they have donated to various projects. This ensures users are returning accurate results for both scientific and statistical reasons.

Purposes for a credit system

Online distributed computing relies heavily, if not completely, on volunteer computers. For this reason, projects such as SETI@home [1] and other BOINC projects depend on a complicated balance among long-term users and the cycle of new users and retiring users.

Reasons for participation

  1. To donate to a scientific cause
    • To advance the specific field of study of a project(s)
    • To help fight disease may have an emotional connection for those participating
  2. To stress test computers
    • Processing distributed computing projects places a computer under continuous full CPU load, therefore, overclockers often use the stress to test their system's stability
  3. To team up, earn credit, and compete against other users and teams
    • Some individuals and teams run many computers and have some dedicated specifically to BOINC in hopes of climbing to the top of the world charts
  4. For personal benefit and recognition
    • Projects such as PlanetQuest plan on allowing individuals to name those planets discovered using their computers
    • Projects such as BURP, and Leiden Classical allow users to submit their own operations for use in the system. BURP allows a user to submit models to be rendered, and Leiden Classical allows users to submit physics calculations.

Cobblestones

The basis for the BOINC credit system is the cobblestone, named after Jeff Cobb of SETI@home. By definition, 200 cobblestones are awarded for one day of work on a computer that can meet either of two benchmarks:

The actual computational difficulty needed to run a given work unit is the basis for the number of credits that should be granted. The BOINC system allows for work of any length to be processed and have identical amounts of credit issued to a user. In so doing, BOINC uses benchmarks to measure the speed of a system, combining that figure with the amount of time required for a work unit to be processed. The interface then can “guess” at the amount of credit a user should receive. Since systems have many variables, including the amount of RAM, the processor speed, and specific architectures of different motherboards and CPUs, there can be wide discrepancies in the number of credits that different computers (and projects) judge a user to have earned.

Most projects require a consensus to be reached by having multiple hosts return the same work unit. If they all agree, then the credit is calculated and all hosts receive the same amount regardless of what they asked for. Each project can use their own policy depending on what they see is best for their specific needs. In general, the top and bottom claimed credits are dropped and an average of the remaining is taken. However, certain other projects award a flat amount per work unit returned and validated.

Total credit

Credits are tracked internally for computers, users, and teams. When a computer processes and returns a work unit, it receives no credit for that action alone. It must first have that work unit validated by the given, project-specific method. Once validated, the computer is granted credit, which can be less than, equal to, or greater than what was requested. This amount is immediately added to the computer, user, and team total. If a work unit is returned past the given deadline (in most cases) or is found to be inaccurate, it is marked as invalid and results in no credit. Users and teams commonly determine world rank by comparing the total number of credits accumulated. This highly favors users and teams that have been around for the longest time. This makes it extremely difficult for new users to rapidly gain ground in the rankings, even if they are running many computers. That said, given the exponential increase in computing power of the average PC, it is relatively easy to surpass inactive BOINC users who have earned all of their points on obsolete machines—even if they were at one time ranked highly. Thus, the highest ranked BOINC users will generally be the ones who are actively crunching.

Recent average credit

To find the useful amount of work provided by a computer, a special calculation called recent average credit (RAC) is used. This calculation is designed to estimate the number of credits a computer, user, and team will accumulate on an average day.

Controversies

The credit allocation has been challenged for several projects like EON [2] and Astroids@Home.[3] These concerns have led to the shut down of many such projects over time and have also lead to several alternative allocation strategies.[4]

Third-party statistics sites

BOINC projects export statistical information in the form of XML files, and make it available for anyone to download. Many different third-party statistical websites have been developed to track the progress of BOINC projects. These sites track computers, users, teams, and countries within individual projects and across many projects. Many different sites provide summary graphics, which can be used on web pages that automatically update to contain the statistical information for the specified user or team.

Active webpages

Inactive webpages (Web archive links)

References

  1. Korpela, Eric J. (2012-05-30). "SETI@home, BOINC, and Volunteer Distributed Computing". Annual Review of Earth and Planetary Sciences. 40 (1): 69–87. doi:10.1146/annurev-earth-040809-152348. ISSN 0084-6597.
  2. Chill, Samuel T; Welborn, Matthew; Terrell, Rye; Zhang, Liang; Berthet, Jean-Claude; Pedersen, Andreas; Jónsson, Hannes; Henkelman, Graeme (2014-07-01). "EON: software for long time simulations of atomic scale systems". Modelling and Simulation in Materials Science and Engineering. 22 (5): 055002. doi:10.1088/0965-0393/22/5/055002. ISSN 0965-0393.
  3. Ďurech, J.; Hanuš, J.; Vančo, R. (2015-11-01). "Asteroids@home—A BOINC distributed computing project for asteroid shape reconstruction". Astronomy and Computing. 13: 80–84. doi:10.1016/j.ascom.2015.09.004. ISSN 2213-1337.
  4. Estrada, Trilce; Flores, David A.; Taufer, Michela; Teller, Patricia J.; Kerstens, Andre; Anderson, David P. (December 2006). "The Effectiveness of Threshold-Based Scheduling Policies in BOINC Projects". 2006 Second IEEE International Conference on e-Science and Grid Computing (e-Science'06). Amsterdam, The Netherlands: IEEE: 88–88. doi:10.1109/E-SCIENCE.2006.261172. ISBN 978-0-7695-2734-5.
This article is issued from Wikipedia. The text is licensed under Creative Commons - Attribution - Sharealike. Additional terms may apply for the media files.