menu
 103,702

Geopolitical Forecasting [GF] Challenge

Can you create a method to predict future geopolitical outcomes that is better than current state-of-the-art methods?

This challenge is closed

stage:
Fall 2018: Winner Announcement
prize:
$200,000

This challenge is closed

more
Summary
Timeline
Updates29
Forum222
Teams816
FAQ
Resources
Summary

Overview

The Problem

Decision makers rely on the Intelligence Community to provide accurate and relevant forecasts, and IARPA is working to identify methods to maximize the quality of these forecasts. IARPA’s Geopolitical Forecasting (GF) Challenge invites individuals or teams to develop innovative solutions and methods for integrating crowdsourced forecasts and other data into accurate, timely forecasts of geopolitical events. Solvers will be competing against each other and in parallel with a similarly structured research program funded by IARPA. Over the course of approximately seven months, the GF Challenge solvers will be asked to develop solutions capable of producing forecasts to a series of questions like:

  • Who will win the upcoming presidential election in Egypt?
  • or What will the spot price of Brent Crude oil be on [date]?

In order to answer these questions, solvers will be given access to a continuously updated stream of forecast judgments produced by a crowd of human forecasters and will be allowed to use other data sources to produce their solutions. The challenge presents an opportunity for individuals and teams to earn prizes by creating methods that successfully forecast of a wide variety of geopolitical events, such as political elections, international conflict, disease outbreaks, and macro-economic indicators. 

Challenge Details

Solvers will be invited to the GF Challenge Platform to compete. Each solver or team will be assigned a unique application programming interface (API) token and login account for the platform. A steady stream of questions (roughly 25 per month) will be released for solvers to produce probabilistic forecasts against (specific requirements will be described in the rules). Questions may be binary (yes/no), multiple choice, or ordered outcome (binned quantity/date). In addition to the questions, solvers will receive access to a continuously updated stream of forecast judgments produced by a crowd of human forecasters. Solvers are encouraged to create solutions that use this stream of human judgments alongside other publicly available data streams and information to create their forecasting solutions.

  • Questions will be released on the platform and can be pulled through the API
  • Each forecast question will include a set of exhaustive, and mutually exclusive response options, and will have a final resolution date that may range from a day to many months in the future. Challenge participants will be asked to produce forecasts until the question resolves. 
  • Examples of potential questions*:
  • Will the WHO confirm more than 10 cases of Marburg in 2018? (Yes/No)
  • Before March 2018, with South Korea file a World Trade Organization dispute against the United States related to solar panels? (Yes/No)
  • All forecasts will need to be submitted through the API
  • Forecasts may be updated and will be scored on a daily basis for each question that is open.

Scoring

Scores will be calculated using a metric which incorporates forecast accuracy and confidence. This metric is based on the Brier score, a quadratic measure of the forecaster's closeness to ground truth for the outcome that occurs. For example, an accurate forecast with 90% confidence for an event that occurs will score four times better than a forecast with 80% confidence. The score will be calculated on a daily basis over the duration of the question; those daily accuracy scores will then be averaged. The more accurate and timely the forecast, the better score.

As questions will range in duration, milestone check-ins for questions resolved during a certain period will be used to award progress-based incentives. This means solvers who start after the challenge has launched are still able to compete for some of the prizes. Finalists who wish to compete for monetary prizes will need to provide the solution package for review the GF Challenge Team at the end of the challenge, or as appropriate, if requested. Solutions will need to be documented and all data streams used in prize-winning solutions will need to be outlined and verified by the GF Challenge Team. Please refer to guidelines for a description of scoring.

Prizes

Solvers will be able to compete on their own or form teams to compete for a total prize purse of $200,000. The prize amounts shown are dependent upon certain criteria and/or participation requirements. Challenge results and solutions will be reviewed at an IARPA forecasting Workshop in Washington, D.C. For additional details and prize requirements, please refer to the guidelines. Prizes will be distributed for the following criteria:

  • Overall Performance Prizes
    • 1st Place - $20,000
      • 1st Place could be eligible to receive the Performance Bonus: Ultimate Forecaster, an additional $40,000*
    • 2nd Place - $15,000
    • 3rd Place - $10,000
    • 4th Place - $7,000
    • 5th Place - $4,000
      • Second through Fifth Place may be eligible for a share of Performance Bonus: Star Forecasters, an additional $30,000*
  • Additional Incentives
    • Best in Domain / Region Pair (X5) - $25,000 total
    • Best Undergraduate - $4,000
    • Milestone Performance (X3) - $30,000 total
    • Interim Prizes - up to $15,000

*Bonuses have additional eligibility requirements. Please see the challenge rules for details.

Rules and Eligibility

Individuals and teams, 18 years of age or older will be able to compete for cash prizes. Some individuals or organizations may not be eligible for monetary prizes. However, these solvers may, upon IARPA approval, participate in the challenge and be eligible for ranking on the leaderboard. Approval will need to be obtained in advance of participation by emailing: gfchallenge@iarpa.gov.

 

What Can You Do Right Now?

  • Click “Follow” to receive challenge updates and follow its progress!

  • Register for the challenge by clicking "Accept the Challenge" and agreeing to the challenge rules and IP agreement.
  • Share this challenge on social media using the icons above. Show your friends, your family, or anyone you know who has a passion for discovery.
  • Start a conversation in our Forum to join the conversation, ask questions or connect with other innovators.

Guidelines

Challenge Rules Document 8.31.18 (Updated)

Challenge FAQ's 2.21.18 (Updated)

Final Submission Document 8.23.18

 

Background

Can you create a method to forecast the future? The Geopolitical Forecasting (GF) Challenge invites Solvers from around the world to develop innovative solutions and methods for integrating crowdsourced forecasts and other data into accurate, timely forecasts on worldwide issues. The challenge presents an opportunity for individuals and teams to earn prizes by creating methods that successfully demonstrate a forecast of a wide variety of geopolitical events, such as political elections, disease outbreaks, and macro-economic indicators.


Overview

Existing methods of geopolitical forecasting include human judgment-intensive methods, such as prediction markets, and data-intensive approaches, such as statistical models. GF Challenge Solvers will develop solutions that produce probabilistic forecasts in response to numerous closed-ended forecasting questions that concern specific, objectively verifiable geopolitical events containing timeframes with deadlines and locations. The effort will run in parallel to IARPA’s most current geopolitical forecasting research program-- Hybrid Forecasting Competition (HFC). Challenge Solvers will be competing on a largely overlapping set of Individual Forecasting Problems (IFPs) as HFC research teams, and given access to the same human forecaster data stream. In addition to the provided data stream, Solvers may use other data streams and their own data and models for the challenge (in accordance with applicable laws).


Challenge Process

The following three (3) stages are anticipated as part of the GF Challenge:

Stage 1: Registration and Configuration

Solvers will register on the HeroX platform and acknowledge their understanding and intent to follow the rules of the challenge. Following registration, instructions will be sent to access the application programming interface (API) within the Cultivate Labs platform. Solvers will be granted an API token and documentation for interacting with the challenge platform.
Stage 1 registrations will be accepted on a rolling basis, so Solvers may register at any point. However, solvers will be rewarded by accuracy scores that incentivize participation on their ability to forecast on more IFPs. Eligibility for prizes (documented below) have minimum participation requirements for the resolved IFPs included in each prize.

Stage 2: Development, Forecasting and Scoring

Forecasting questions, known as IFPs, will be released periodically and will be open for a specified period of time. Solvers should check daily for new IFPs. While an IFP is open, Solvers may update their forecasts daily until the IFP closes. Please see the Judging and Scoring section below for the full-time periods associated with the IFP forecasting period. Each Solver will be permitted to submit forecasts using up to 25 different methodological “slots” as described in the Rules section below. Once the IFP closes, the forecasts will be locked and scores for the IFP for each method will be calculated. See the Judging and Scoring section below for further information. Scores will be tracked on the leaderboard within the Cultivate Labs platform.
In addition to the forecast submission activities, Solvers are encouraged to work to improve their forecasting methods throughout the competition.
Throughout the Challenge, there will be three (3) Milestone Check-Ins at which Milestone Prizes will be awarded. The Milestone Prizes will be awarded based on all IFPs that resolve during the Milestone duration. The Milestone schedule is as follows (milestone prizes will be validated within 30 days of the milestone end):

  • Milestone 1: March 7, 2018 – April 30, 2018
  • Milestone 2: May 1, 2018 – June 30, 2018
  • Milestone 3: July 1, 2018 – September 7, 2018

Stage 3: Solution Evaluation and Prize Awards

At the end of the challenge, the grand total scores will be confirmed and determined for each of the prize categories. Solvers will need to provide a write-up of their solution to the Government for final review and evaluation. This final review will determine if the solutions are sound and meet the criteria of the challenge. Solvers will provide a short (4 page) explanation for their solution. Before final prize award, IARPA may ask for additional documentation.
Upon conclusion of the Government evaluation of the documentation provided by the Solvers from the Evaluation stage, prizes will be awarded as follows.

Overall Performance Prizes (Based on overall score)

Prize

Amount

Criteria

1st Place Overall

$20,000

  • Must participate in at least 70% of the IFPs.  Must have positive NBPs on more than half of the IFPs you attempt
  • Uses the single best scoring method from all of the Solver’s available methods across all IFPs

Bonus:
Ultimate Forecaster

$40,000

  • Must achieve the 1st Place Overall Prize
  • Must beat the GF Challenge Baseline
  • Must beat the HFC Top Score

2nd Place Overall

$12,000

  • Must participate in at least 70% of the IFPs.  Must have positive NBPs on more than half of the IFPs you attempt
  • Uses the single best scoring method from all of the Solver’s available methods across all IFPs

3rd Place Overall

$10,000

4th Place Overall

$7,000

5th Place Overall

$4,000

Bonus:

Star Forecasters

$30,000 split amongst all eligible Solvers/teams

  • Must achieve 2nd – 5th Place Prizes
  • Must beat the GF Challenge Baseline
  • Must beat the HFC top score
  • Bonus will be divided among the qualifying winning Solvers / teams according to the overall place as follows:
    • 4 winners: 50%, 25%, 15%, 10%
    • 3 winners: 50%, 30%, 20%
    • 2 winners: 70%, 30%
    • 1 winner: 100%

Definitions of prize criteria:

  • Net Brier Points (NBPs) – See the Judging and Scoring section point 3 for a description of Net Brier Points.
  • GF Challenge Baseline: Challenge solvers will have access to a stream of de-identified individual-level human forecast data, and an aggregate stream of human forecast data. This aggregate will be constructed using a method based on the state-of-the art from on the IARPA Aggregative Contingent Estimation (ACE) Program. The aggregate measure will serve as the baseline against which NBPs will be calculated. This baseline will be available to solvers via the Cultivate Labs platform.
  • HFC Top Score: The top scoring method coming out of the parallel HFC Perform teams will be visible on the leaderboard during the challenge.

Additional Incentives/Milestone Prizes

Prize

Amount

Criteria

Best in Domain / Region Pair

$25,000 (5 awards of $5,000 each)

  • Top score on each of the five (5) domain/region pairings at the end of the challenge
  • Minimum participation in 80% of the total IFPs in the domain/region pairing
  • Uses the best scoring method to achieve the domain/region pairing, regardless of the Solver’s overall method used in final scoring

Best Undergraduate

$4,000

  • Best score by an undergraduate, or all-undergraduate team who did not place in the Top five (5)
  • Minimum participation in more than half of the total challenge IFPs
  • Certification requirements (transcript, certified letter) from current institution must be presented in order to claim this prize

Milestone 1

$7,500 (10 awards of $750 each)

  • Top ten Solvers with the highest score in the Milestone will each get $750
  • Minimum participation in 80% of the total IFPs resolved within the milestone period

Milestone 2

$10,000(10 awards of $1,000 each)

  • Top ten Solvers with the highest score in the Milestone will each get $1,000
  • Minimum participation in 80% of the total IFPs resolved within the milestone period

Milestone 3

$12,500

Minimum participation in 80% of the total IFPs resolved within the milestone period 

Top ten Solvers with the highest score in the Milestone will each get $1,250 

Election Forecaster$10,000
  • Top Solver in election-related IFPs at the end of the challenge
  • Minimum participation in 80% of the total IFPs in the Elections topic
  • Based on Solver's best scoring method for Elections IFPs, regardless of Solver's best overall method
Spring Forecaster$2,400 (3 awards of $800 each) 
  • Top three Solvers with the highest score on IFPs that open during the month of April and resolve by June 20
  • Minimum participation in 80% of the total IFPs in the relevant interim category
  • In the event of a tie, winning Solvers will split the prize purse evenly among the top scores
Interim Prize$2,600
  • Additional incentives will be provided throughout the challenge and at the discretion of the challenge team. Prize amounts could be applied to milestone, domain/region pairings, or additional bonus incentives to the Top 5 Overall Performance Prizes. These Interim Prizes will be announced prior to the start of the Prize Calculation.
  • Minimum participation in 80% of the total IFPs in the relevant interim category

 

Payment Terms

Solvers will need to submit a W-9 tax form, or a W8-BEN form in order to receive payment. Solvers are responsible for payment of all taxes incurred from the acceptance of Prize funds. Solvers will be given 30 days to submit their paperwork upon notification of award.  IARPA and Booz Allen are not responsible for lost or stolen prize payments, or incorrect routing and payment information provided by the winning Solvers.

Rules

  • All Solvers will need to fill out a registration form providing information on their team and asserting eligibility to participate through the HeroX GF Challenge page.
  • All daily system forecast submissions must be made via the Cultivate Labs GF Challenge API and authenticated using the API token provided during the registration process. All submissions must match the format described in the API document available on the Cultivate Labs Platform.
  • IFPs and the human forecast stream will only be provided to Solvers via the  Cultivate Labs GF Challenge API.
  • GF Challenge forecast days begin at 14:01 U.S. Eastern Time (Standard or Daylight Savings as appropriate). The daily reporting deadline is 14:00:59 U.S. Eastern Time.

    Each day’s final/official forecast (due 14:00:59. ET) will be treated as a prospective forecast that covers the immediately following 24-hour period, ending at 14:00:59 ET the next calendar day. The final forecast for an IFP is defined to be the last forecast submitted on the last forecasting day before the IFP is resolved. Suppose IFP closes as scheduled at 23:59:59 ET on 31 December, then a forecast at 14:00:58 ET that day would count but a forecast at 14:01:01 ET would not count. If an IFP closes early, say 09:00:00 ET on 30 December then a forecast at 14:00:58 ET on 29 December would count but a forecast at 14:01:01 ET on 29 December would not count.
  • Each Solver is permitted to submit daily forecasts for up to 25 official competition forecasting methods or permutations (aka “slots”) at any given time. Solvers shall provide names of 50 characters or fewer for their forecasting methods, which must be used consistently throughout the challenge period. These names may not include profane or derogatory terms.
  • Solvers may submit more than one forecast update per IFP per day. For each forecasting method, only the last forecast for an IFP submitted before the daily submission deadline of a given forecast day will be treated as the Solver’s official forecast for each IFP for that forecast day.
  • Each Solver’s forecasting method will be scored across all IFPs for which forecasts are submitted, and an overall forecast accuracy score for that method will be determined. For each Solver, the Testing and Evaluation (T&E) Team will identify the single method with the best score across all IFPs as the Solver’s “best” method. In other words, each Solver’s best method will be a unitary, prospectively defined method selected from among the Solver’s 25 official methods: We will not “cherry-pick” per-IFP most accurate methods to formulate composite, per-Solver “best” methods.
  • Once a Solver submits their first forecast on an IFP they should provide daily forecast updates. If they do not, their previous forecast will be carried over each forecasting day until a new forecast is submitted.
  • For forecast days prior to the first submission by a Solver on an IFP, the assumed forecast is the aggregated human forecaster stream forecast for that IFP.
  • For Milestone Prizes, only the IFPs that are resolved during that Milestone’s time period will count for that Milestone’s prize. IFPs that are opened, but not resolved, during a Milestone will not count. They will be counted towards the Milestone period in which the IFP resolves.
  • In order to be eligible for Domain/Region, Election Forecaster, Spring Forecaster, interim, and Milestone awards, a Solver must submit forecasts for at least 80% of the IFPs considered for that prize. To be eligible for the Undergrad prize, Solvers must submit forecasts for at least 50% of the total number of IFPs in the challenge and provide certification of their status as a student from their current institution. In order to be eligible for one of the Top five (5) overall performance prizes, a Solver must submit forecasts for at least 70% of the total number of IFPs in the challenge.
  • Solvers are responsible for ensuring that submitted forecasts are compliant with the standards documented within the API documents on the Cultivate Labs Platform. Forecast submissions must be proper in that the probabilities assigned to the several outcomes for an IFP must sum to 1.0. Improper, malformed or non-compliant forecast submissions will be rejected. IARPA reserves the right to change its policies with reasonable notice, including disqualification of non-compliant data.
  • Solvers may use external data streams in their solutions. However, these external data sources will need to be documented. The Solver is responsible for ensuring that they have the necessary rights and licenses for the use of such external data sources for the purposes of this challenge.
  • The Human Forecaster data stream provided through the challenge is available for Solvers to use, free of charge, only for the purposes of this challenge. All other uses are prohibited without the express consent of IARPA. However, Solvers are not required to use this data stream as part of their solution.
  • Solvers are solely responsible for ensuring they have valid licenses or rights to any software libraries or data sources used in their forecasting systems. Any method that infringes on another party’s intellectual property will be disqualified from the competition.
  • Solvers may only register once for the challenge and can compete either as an individual or as part of a team; Solvers cannot participate as both an individual and as part of a team.

Judging and Scoring

  1. Your competition score is based on how accurate your forecasts are when compared with the stream of Government-furnished human forecasts. The Government-furnished forecast stream (GFFS) is a “crowd wisdom” aggregation of asynchronously submitted human forecasts for each IFP. Given an IFP we compute the consensus forecast as a function of the weighted average of the latest forecasts from a portion of the unique forecasters who most recently provided forecasts for this IFP. The weight assigned to each forecaster is computed from three factors. The first two are based on their performance across all IFPs: historical forecast accuracy for that forecaster and frequency with which that forecaster submits updates. The third factor is whether or not they have completed a brief, self-administered forecast training course. The weighted average of the recent forecasts is then extremized toward 0 or 1 using a tuning parameter. This method for computing the consensus forecast was determined to have good performance during the IARPA ACE Program and is a suitable benchmark against which to compare solver performance.
  2. We measure absolute forecast accuracy using a formula called the Brier score, which reflects the squared difference between the probabilities a solver or the GFFS assigns to each outcome and reality (see appendix section A.1). Lower Brier scores reflect greater accuracy: A perfect score is 0 (for instance, on a binary question, you assigned 100% to the correct outcome), the worst possible score is 2 (for instance, on a binary question, you assigned 100% to an incorrect outcome). For some IFPs the possible answers have an ordering such that some answers are closer to being correct than others. In this case, we use a modified version of the Brier score that awards more credit when assigning higher forecast probabilities to answers that are closer to the correct answer (see Appendix: A.2 Application of Brier Scoring to Ordered Outcomes). The appendix provides a complete description of generic and ordered Brier scoring.
  3. For a given IFP, we compare the accuracy of your forecasts to those of the GFFS using Net Brier Points (NBPs), defined as the difference between the GFFS’s Brier score and your Brier score for each day that the IFP is open. Negative NBPs means the GFFS was more accurate than you, positive NBPs means you were more accurate than the GFFS, and more NBPs is better. For days that pass before your first forecast on the IFP, your forecast is assumed to be the same as the GFFS consensus forecast, meaning you do not accrue (or lose) NBPs for a given IFP until you submit your first forecast. Once you submit an initial forecast, that forecast holds each day thereafter until/unless you update it. Your score for the IFP is then the sum of your NBPs for all days that an IFP is open.
  4. For most of the prize categories your score is the sum of your NBPs over all germane IFPs for that category. However, the bonus Ultimate Forecaster and Star Forecasters Prizes are scored slightly differently. These bonus prizes feature a comparison of your method to the best of the performers in the IARPA HFC program on the same IFPs. For the purposes of this bonus prize category we will compute your Mean Daily Brier Score (see Appendix: A.3 Mean Daily Brier Score ) for all IFPs and take the mean of those scores across those IFPs. If your mean Mean Daily Brier score is lower than the best HFC program method you will be eligible for this bonus prize category.
  5. To be eligible for overall prizes you must attempt at least 70% of all IFPs and you must have positive NBPs on more than half of the IFPs you attempt. For example, suppose there are 150 IFPs for the overall prize category. Suppose that you submit non-trivial forecasts on 105 of these IFPs. Furthermore, let us say that you have positive NBPs on at least 53 IFPs. You will meet all of the eligibility criteria for overall prizes.
  6. In order to be eligible for a prize the Solver must document the solution method in a short document (no more than four pages) using the provided format (coming soon). Please note that IARPA will be examining the substance and clarity of the technical details documented, not the quality of the writing.
  7. Solutions competing for prizes will be verified by the Solver or Team that all data sources and software libraries can be identified through the solution with proper attribution where necessary.

Intellectual Property

Solvers will retain the Intellectual Property (IP) rights to their solutions. IARPA will use information submitted through the challenge and application process to understand the technology.

  • IARPA will have the rights to utilize all data, forecasts, and method explanations provided for U.S. Government purposes.
  • Challenge participant names, titles, general technology descriptions, photographs, and abstracts for their submissions may be utilized in challenge-related media and promotional materials or for other internal Government uses. No sensitive intellectual property information will be shared in this manner, and must be clearly marked prior to submission.
  • Challengers will need to sign the IP Agreement from within the HeroX GF Challenge page prior to participating in the challenge.

Eligibility

Some individuals or organizations may not be eligible for prizes for reasons listed below. However, these Solvers may, upon IARPA approval, participate in the challenge and be eligible for ranking on the leaderboard. Approval will need to be obtained in advance of participation by emailing .

Any teams who forego prizes will be publicly listed on the HeroX GF Challenge page to ensure full transparency of their competition status.

General Eligibility Requirements

To be eligible to win a prize under this competition, an individual or entity:

  1. Must have completed and submitted a registration form at HeroX GF Challenge page
  2. Must be (1) an individual or team each of whom are 18 years of age and over, or (2) an entity incorporated; and
  3. May not be a federal entity or federal employee acting within the scope of their employment. An individual or entity shall not be deemed ineligible because the individual or entity used federal facilities or consulted with federal employees during a competition if the facilities and employees are made available to all individuals and entities participating in the competition on an equitable basis.

Federal grantees may not use federal funds to develop challenge applications unless consistent with the purpose of their grant award. Federal contractors may not use federal funds from a contract to develop challenge applications or to fund efforts in support of a challenge submission.

Employees of IARPA, Booz Allen Hamilton, Cultivate Labs, HeroX, Good Judgment, Inc., and MITRE are not able to compete in the challenge. Companies and contractors who are currently supporting, or who have previously supported, the above (or other) entities in efforts related to the IARPA HFC program are also not able to participate in the challenge. This restriction extends to members of such persons’ immediate families (spouses, children, siblings, parents), and persons living in the same household as such persons, whether or not related.

Federally Funded Research & Development Centers (FFRDCs) and (DoD) University Affiliated Research Centers (UARCs) not mentioned above may be eligible to submit forecasts and receive leaderboard recognition, but are not eligible to win prize dollar awards. In order to compete, an email must be sent to gfchallenge@iarpa.gov with the team name, organization represented, and team member information.


IARPA’s Hybrid Forecasting Competition (HFC) performer team members and HFC Good Judgment Forecasters, as well as members of such persons’ immediate families (spouses, children, siblings, parents), and persons living in the same household as such persons, whether or not related, are not eligible to participate in the competition.


Federal employees, employees of FFRDCs, and employees of UARCs can participate in the challenge as a private citizen, not affiliated with their organization, as long as (a) the Solver is not employed by an organization that is actively involved in the challenge or closely related programs; (b) the Solver has received approval from their organization, if applicable, to participate in their personal capacity (note, it is the Solver’s responsibility to verify this); and (c) Solvers cannot use government or their organizations’ resources, computers, or access to government information in aid of their participation, except for those resources available to all other participants on an equal basis.

Entrants must agree to assume any and all risks and waive claims against the Federal Government and its related entities, except in the case of willful misconduct, for any injury, death, damage, or loss of property, revenue, or profits, whether direct, indirect, or consequential, arising from their participation in a competition, whether the injury, death, damage, or loss arises through negligence or otherwise.

Entrants must also agree to indemnify the Federal Government against third-party claims for damages arising from or related to competition activities. Entrants are not required to obtain liability insurance or demonstrate financial responsibility in order to participate in the competition.

By participating in the competition, each entrant agrees to comply with and abide by these rules and the decisions of IARPA and/or the individual judges, which shall be final and binding in all respects.

By participating in the competition, each entrant agrees to follow all applicable local, state, federal and country of residence laws and regulations.

Companies/Teams

Companies, universities, individuals, and Teams of Solvers are able to participate in this challenge. Companies / Teams will need to elect a Team Captain, who will be the main point of contact (POC) for communications. Individuals who have registered on the HeroX GF Challenge page as an individual competitor, will not be eligible to participate in a team. All Team Members must meet the General Eligibility Requirements, self-register, and acknowledge these rules through the HeroX GF Challenge page.

  • Upon registering for the challenge, the Team Captain will need to provide a breakdown of all Team Members along with the percentage allocation between all Team Members
  • Each Team Member will need to sign an agreement that they agree to distribution of prize funds
  • A Team Captain may provide information for a company or organization instead of their personal information for tax purposes.
  • One representative from the Team, will need to set up their Team’s API. There will only be one API allowed per Team for the entire duration of the challenge
  • Once a Team has submitted their first forecast, the team members cannot change. This means that no new Solvers can be added to a Team and no Solvers can leave the Team.
  • Individual Solvers who have submitted a forecast on their own may not form a Team after their first forecast has been submitted.

Foreign Nationals and International Solvers

Except for residents of Iran, Cuba, North Korea, Crimea Region of Ukraine, Sudan or Syria or other countries prohibited on the U.S. State Department’s State Sponsors of Terrorism list, Solvers residing in any country are able to participate. In addition, Solvers are not eligible to participate if they are on the Specially Designated National list promulgated and amended, from time to time, by the United States Department of the Treasury. It is the responsibility of the Solver to ensure that they are allowed to participate in this challenge and allowed to export their technology solution to the United States.

General Liability Release

By participating in the competition, each entrant hereby agrees that:

  1. IARPA and the U.S. Government, Booz Allen Hamilton, Cultivate Labs, HeroX, and MITRE shall not be responsible or liable for any losses, damages, or injuries of any kind (including death) resulting from participation in the competition or any competition-related activity, or from entrants’ acceptance, receipt, possession, use, or misuse of any prize; and
  2. Entrants will indemnify, defend, and hold harmless IARPA and the U.S. Government, Booz Allen Hamilton, Cultivate Labs, HeroX, and MITRE from and against all third-party claims, actions, or proceedings of any kind and from any and all damages, liabilities, costs, and expenses relating to or arising from entrant’s participation in the competition.

Without limiting the generality of the foregoing, IARPA and the U.S. Government, Booz Allen Hamilton, Cultivate Labs, HeroX, and MITRE are not responsible for incomplete, misdirected, prize notifications; or for late, lost, inaccessible, network connections, or for issues with Internet Service Providers, websites, or other connections and third-party sites that the entrants choose to use for the competition; or for any technical malfunctions, failures, difficulties, or other errors of any kind or nature.

These rules cannot be modified except by IARPA. All decisions by IARPA regarding adherence to these rules are final. The invalidity or unenforceability of any provision of these rules shall not affect the validity or enforceability of any other provision. In the event that any provision is determined to be invalid or otherwise unenforceable or illegal, these rules shall otherwise remain in effect and shall be construed in accordance with their terms as if the invalid or illegal provision were not contained herein.

Warranties/Indemnification

By participating in the competition, each Solver represents, warrants, and covenants as follows:

  1. The Solver – whether an individual, team or entity – is the sole author, creator, and owner of the submission;
  2. The submission is not the subject of any actual or threatened litigation or claim;
  3. The submission does not and will not violate or infringe upon the intellectual property rights, privacy rights, publicity rights, or other legal rights of any third party;
  4. The entrant will not interfere with or attempt to influence, the events that are the subject of the IFPs in the challenge. If a Solver has a question about an activity related to their job role, please email gfchallenge@iarpa.gov for clarification.
  5. The Solver will not engage in activities that could cause political unrest, or lead to a violent event or civil unrest that could cause the outcome of the IFP to change.; and
  6. The Submission, and Solvers’ use of the Submission, does not and will not violate any applicable laws or regulations, including, without limitation, applicable export control laws and regulations of the U.S. and other jurisdictions.

If the Submission includes any third party works (such as third-party content, data sets, algorithms, or open source code), entrant must be able to provide, upon the request of IARPA, documentation of all appropriate licenses and releases for such third-party works. If entrant cannot provide documentation of all required licenses and releases, IARPA reserves the right to disqualify the applicable Submission, or seek to secure the licenses and releases for the benefit of IARPA, and allow the applicable Submission to remain in the Competition. IARPA also reserves all rights with respect to claims based on any damages caused by participant’s failure to obtain such licenses and releases.

Solvers – whether an individual, a team or an entity – will indemnify, defend, and hold IARPA, Booz Allen Hamilton, Cultivate Labs, HeroX and MITRE from and against all third-party claims, actions, or proceedings of any kind and from any and all damages, liabilities, costs, and expenses relating to or arising from entrant’s submission or any breach or alleged breach of any of the representations, warranties, and covenants of entrant hereunder.

IARPA reserves the right to disqualify any submission that IARPA, in its discretion, deems to violate these Rules. IARPA reserves the right to disqualify any Solver that is operating in a manner that disrupts the competition, does not comply with the spirit of the competition, or is found to be manipulating the competition. IARPA also reserves the right to amend these rules throughout the duration of the contest should extenuating circumstances arise.

Appendix (available in the downloadable Challenge Rules Document)

Timeline
Updates29

Challenge Updates

Milestone 2 Award Winners

July 16, 2018, 12:34 p.m. PDT by Kyla Jeffrey

Dear solvers,

We’re so excited to announce the winners of Milestone 2 of the GF Challenge!

The top 10 Solvers with the highest score in this Milestone category will each receive $1,000. As a reminder, a minimum participation in 80% of the total IFPs resolved within the Milestone 2 period was required to be eligible for this prize. The winners are:

  1. Citizensband
  2. SISLers
  3. Catskills
  4. Dobreovidiu
  5. DigitalDelphi
  6. truthLover
  7. Seb
  8. BEEFERS
  9. Pb2pv
  10. LIA

Didn’t make it this round? Don’t feel discouraged — Milestone 3 began July 1st and continues through September 7th. This Milestone will offer the top 10 Solvers with the highest score a prize of $1250 each! That’s a total prize pool of $12,500. 

Forecast on!

GF Challenge Team


Scoring & Leaderboard Update

July 12, 2018, 8:31 a.m. PDT by Kyla Jeffrey

Thank you for your patience!  The scoring audit and adjustments have been completed, we wanted to also let you know that leaderboards have been updated to exclude voided questions. You should see those adjustments already made to the boards, and inconsistencies should now be resolved.  Thanks for the feedback and continued engagement.


Scoring Audit Update

July 10, 2018, 8 a.m. PDT by Kyla Jeffrey

Solvers,

We wanted to alert you to a finding from our recent quality assurance check. While conducting a routine audit of the scoring functionality, we determined that there was an inconsistency in scoring returns.  The automated scoring/leaderboarding routine was inadvertently omitting the first day’s forecast from question and overall Brier scores.* The impact on overall Brier scores was generally minimal (<=0.01 change to overall mean of mean daily Briers). As a result, you will see periodic changes to the leaderboards as we re-run the current scores. This should continue over the next couple of days, and we will update you when we’ve completed this process.

We have also recalculated the scores for the Milestone 1 and Spring Forecaster Awards, and have determined that the Spring Forecaster Award winners remain the same, though the scores have adjusted slightly.  However, there are a couple additions to be made for Milestone 1. The new calculations show that both seb and Digital Delphi should have been recognized at that time. We will be correcting this immediately, so please watch your accounts for our email.

Moving forward: We will continue to regularly audit the scoring system to ensure the highest level of accuracy. The updated scoring will be applied to Milestone 2 and all future awards. And, we will continue to keep you up-to-date with news as soon as we have it.

In the meantime, let us know if you have questions or concerns. We are a community, and we appreciate your contributions and time.

V/r

 

*For example, if questions were released on 01/01, final forecasts submitted before 2:01pm ET 01/02 should have been treated as the first scored “day”; however, the system was not doing this: it was treating finial forecast submitted before 2:01pm ET 01/03 as first scoreable forecasts and was omitting the day ending 2:01pm ET 01/02 from scoring.


Independence Day

July 2, 2018, 8:15 p.m. PDT by Kyla Jeffrey

Dear Solvers,

Sending along a quick heads up! We won’t be releasing IFPs next Wednesday as we will all be out honoring Independence Day. So if you are in the U.S., enjoy the fireworks and friends. Others around the world, hope your day goes smoothly, and we will resume the IFP release on July 11th.


End of Milestone 2

June 30, 2018, 6 a.m. PDT by Kyla Jeffrey

Solvers,

 

We just wrapped up Milestone 2 — congratulations!  Your continued participation in this unique competition is bringing you closer to your share of the prize purse. Remember: The top 10 Solvers with the highest score in this Milestone category will each receive $1,000. A minimum participation in 80% of the total IFPs resolved within the Milestone 2 period is required to be eligible for this prize. Stay tuned for the results.

 

Milestone 3 begins July 1st and continues through September 7th. The Milestone 3 prize pool totals $12,500; the top 10 Solvers with the highest score for IFPS that resolve from July 1 – September 7 will each win $1,250.  As a reminder, minimum participation in 80% of the total IFPs resolved within this milestone period is required to be eligible for prizes. 

 

Happy forecasting!

GF Challenge Team


Forum222
Teams816
FAQ
Resources