The Good Judgment Project (GJP) is a project "harnessing the wisdom of the crowd to forecast world events". It was co-created by Philip E. Tetlock (author of Superforecasting and of Expert Political Judgment: How Good Is It? How Can We Know?), decision scientist Barbara Mellers, and Don Moore.[1][2][3] It was a participant in the Aggregative Contingent Estimation (ACE) program of the Intelligence Advanced Research Projects Activity (IARPA) in the United States.[4][5] Predictions are scored using Brier scores.[6] The top forecasters in GJP are "reportedly 30% better than intelligence officers with access to actual classified information."[7]
The Almanac of Menlo Park published a story on the GJP on January 29, 2015.[25] An article on the GJP appeared on the portal of the Philadelphia Inquirer Philly.com on February 4, 2015.[26] The book Wiser: Getting Beyond Groupthink to Make Groups Smarter has a section detailing the involvement of the GJP in the tournament run by IARPA.[27] Psychology Today published online a short article summarizing the paper by Mellers et al. that wraps up the main findings of the GJP.[28][29]
The project spawned in 2015 a book by Tetlock, one of the principal investigator, and coauthored by Dan Gardner, that divulges the main findings of the research conducted with the data from the GJP: Superforecasting - The Art and Science of Prediction.[30] Co-author Gardner had already published a book in 2010, that quoted previous research by Tetlock that seeded the GJP effort.[31] A book review in the Economist September 26, 2015, print edition, discusses the main concepts.[32] A Wall Street Journal article depicts it as: "The most important book on decision making since Daniel Kahneman’s Thinking, Fast and Slow.[33] The Harvard Business Review paired it to the book The Power of Mathematical Thinking by Jordan Ellenberg.[34] On September 30, 2015, National Public Radio aired the Colin McEnroe Show centering on the GJP and the book Superforecasting; guests of the show were prof. Philip Tetlock, the director of IARPA, Jason Matheny, and superforecaster Elaine Rich.[35]
Contents
[hide]History[edit]
The Good Judgment Project began in July 2011 in collaboration with IARPA-ACE.[8] The first contest began in September 2011.[9][10] GJP was one of many entrants in the IARPA-ACE tournament, and has repeatedly emerged as the winner in the tournament.[5] Starting with the summer of 2013, GJP contestants had access to the Integrated Conflict Early Warning System.[6]People[edit]
The co-leaders of the GJP include Philip Tetlock, Barbara Mellers and Don Moore.[1] The website lists a total of about 30 team members, including the co-leaders as well as David Budescu, Lyle Ungar, Jonathan Baron, and prediction-markets entrepreneur Emile Servan-Schreiber.[11] The advisory board included Daniel Kahneman, Robert Jervis, J. Scott Armstrong, Michael Mauboussin, Carl Spetzler and Justin Wolfers.[12] The study employed several thousand people as volunteer forecasters.[13] Using personality-trait tests, training methods and strategies the researchers at GJP were able to select forecasting participants with less cognitive bias than the average person; as the forecasting contest continued the researchers were able to further down select these individuals in groups of so-called superforecasters. The last season of the GJP enlisted a total of 260 superforecasters.Research[edit]
A significant amount of research has been conducted based on the Good Judgment Project by the people involved with it.[14][15] The results show that harnessing a blend of statistics, psychology, training and various levels of interaction between individual forecasters, consistently produced the best forecast for several years in a row.[13]Good Judgment Open[edit]
A commercial spin-off of the Good Judgment Project started to operate on the web in July 2015 under the name Good Judgment Inc. Starting in September 2015, Good Judgment Inc has been running a public forecasting tournament at the Good Judgment Open site. Like the Good Judgment Project, Good Judgment Open has questions about geopolitical and financial events, although it also has questions about US politics, entertainment, and sports.Media coverage[edit]
GJP has repeatedly been discussed in The Economist.[9][10][16] GJP has also been covered in the New York Times,[3] the Washington Post,[5][17][18] and Co.Exist.[19] NPR aired a segment on The Good Judgment Project by the title So You Think You're Smarter Than A CIA Agent, on April 2, 2014.[7] The Financial Times published an article on the GJP on September 5, 2014.[20] The Washingtonian published an article that mentioned the GJP on January 8, 2015.[21] The BBC and the Washington Post published articles on the GJP respectively on January 20, 21 and 29, 2015.[22][23][24]The Almanac of Menlo Park published a story on the GJP on January 29, 2015.[25] An article on the GJP appeared on the portal of the Philadelphia Inquirer Philly.com on February 4, 2015.[26] The book Wiser: Getting Beyond Groupthink to Make Groups Smarter has a section detailing the involvement of the GJP in the tournament run by IARPA.[27] Psychology Today published online a short article summarizing the paper by Mellers et al. that wraps up the main findings of the GJP.[28][29]
The project spawned in 2015 a book by Tetlock, one of the principal investigator, and coauthored by Dan Gardner, that divulges the main findings of the research conducted with the data from the GJP: Superforecasting - The Art and Science of Prediction.[30] Co-author Gardner had already published a book in 2010, that quoted previous research by Tetlock that seeded the GJP effort.[31] A book review in the Economist September 26, 2015, print edition, discusses the main concepts.[32] A Wall Street Journal article depicts it as: "The most important book on decision making since Daniel Kahneman’s Thinking, Fast and Slow.[33] The Harvard Business Review paired it to the book The Power of Mathematical Thinking by Jordan Ellenberg.[34] On September 30, 2015, National Public Radio aired the Colin McEnroe Show centering on the GJP and the book Superforecasting; guests of the show were prof. Philip Tetlock, the director of IARPA, Jason Matheny, and superforecaster Elaine Rich.[35]
See also[edit]
- Wisdom of the crowd
- Aggregative Contingent Estimation
- Intelligence Advanced Research Projects Activity
- SciCast
References[edit]
- ^ Jump up to: a b "Welcome to the Good Judgment ProjectTM". The Good Judgment ProjectTM. Retrieved May 5, 2014.
- Jump up ^ "Who’s who in the Good Judgment Project". The Good Judgment ProjectTM. July 27, 2011. Retrieved May 5, 2014.
- ^ Jump up to: a b Brooks, David (March 21, 2013). "Forecasting Fox". New York Times. Retrieved May 5, 2014.
- Jump up ^ "The Project". The Good Judgment ProjectTM. Retrieved May 5, 2014.
- ^ Jump up to: a b c Horowitz, Michael (November 26, 2013). "Good judgment in forecasting international affairs (and an invitation for season 3)". Washington Post. Retrieved May 5, 2014.
- ^ Jump up to: a b Dickenson, Matt (November 12, 2013). "Prediction and Good Judgment: Can ICEWS Inform Forecasts?". Predictive Heuristics. Retrieved May 24, 2014.
- ^ Jump up to: a b Spiegel, Alix. "So You Think You're Smarter Than A CIA Agent". NPR.org. Retrieved 2014-08-18.
- Jump up ^ "The idea behind the Good Judgment Project". The Good Judgment ProjectTM. July 27, 2011. Retrieved May 5, 2014.
- ^ Jump up to: a b "The perils of prediction: Adventures in punditry". The Economist. September 2, 2011. Retrieved May 6, 2014.
- ^ Jump up to: a b "Monetary policy: How likely is deflation?". The Economist. September 13, 2011. Retrieved May 6, 2014.
- Jump up ^ "Team". The Good Judgment ProjectTM. Retrieved May 5, 2014.
- Jump up ^ "Freakonomics". Sign Up for a Prediction Tournament. 2011-08-04.
- ^ Jump up to: a b Mellers, Barbara; Ungar, Lyle; Baron, Jonathan; Ramos, Jaime; Gurcay, Burcu; Fincher, Katrina; Scott, Sydney E.; Moore, Don; Atanasov, Pavel; Swift, Samuel A.; Murray, Terry; Stone, Eric; Tetlock, Philip E. (2014-05-01). "Psychological strategies for winning a geopolitical forecasting tournament". Psychological Science. 25 (5): 1106–1115. ISSN 1467-9280. PMID 24659192. doi:10.1177/0956797614524255.
- Jump up ^ Ungar, Lyle; Mellers, Barbara; Satopää, Ville; Baron, Jon; Tetlock, Philip E.; Ramos, Jaime; Swift, Sam. "The Good Judgment Project: A Large Scale Test of Different Methods of Combining Expert Predictions". Association for the Advancement of Artificial Intelligence.
- Jump up ^ Ulfelder, Jay (March 27, 2014). Using the 'Wisdom of (Expert) Crowds' to Forecast Mass Atrocities (Report). SSRN 2418980 . doi:10.2139/ssrn.2418980.
- Jump up ^ "International: Who’s good at forecasts? How to sort the best from the rest". The Economist. November 18, 2013. Retrieved May 6, 2014.
- Jump up ^ Ignatius, David (November 1, 2013). "More chatter than needed". Washington Post. Retrieved May 6, 2014.
- Jump up ^ Bender, Jeremy (April 3, 2014). "Huge Experiment Finds Regular Folks Predict World Events Better Than CIA Agents". Business Insider. Retrieved May 6, 2014.
- Jump up ^ "The Surprising Accuracy Of Crowdsourced Predictions About The Future. Do you know whether Turkey will get a new constitution? It turns out you do: A group of well-informed citizens can predict future events more often than any foreign policy expert or CIA analyst.". Co.exist. April 21, 2014. Retrieved May 6, 2014.
- Jump up ^ Harford, Tim (2014-09-05). "How to see into the future". Financial Times. ISSN 0307-1766. Retrieved 2014-09-05.
- Jump up ^ Hamilton, Keegan. "How US Agencies Are Using the Web to Pick Our Brains". Washingtonian. Retrieved 2015-01-24.
- Jump up ^ Burton, Tara (2015-01-20). "Could you be a ‘super-forecaster’?". BBC Future. Retrieved 2015-01-21.
- Jump up ^ Jensen, Nathan (2015-01-21). "Experts see a Republican Senate and fast-track authority for Obama as keys to new trade agreements". The Washington Post. ISSN 0190-8286. Retrieved 2015-01-21.
- Jump up ^ Mellers, Barbara; Michael C. Horowitz (2015-01-29). "Does anyone make accurate geopolitical predictions?". The Washington Post. ISSN 0190-8286. Retrieved 2015-01-30.
- Jump up ^ "Feature story: Bob Sawyer of Woodside discovers his latent talent in forecasting". Retrieved 2015-03-17.
- Jump up ^ Dribben, Melissa; Inquirer Staff Writer (2015-02-04). "Fortune telling: Crowds surpass pundits". Philly.com. Retrieved 2015-02-06.
- Jump up ^ Sunstein, Cass R.; Hastie, Reid (2014-12-23). Wiser: Getting Beyond Groupthink to Make Groups Smarter. Harvard Business Review Press. ISBN 978-1-4221-2299-0.
- Jump up ^ "Who's Best at Predicting the Future? (and How to Get Better)". Psychology Today. Retrieved 2015-07-11.
- Jump up ^ Mellers, Barbara; Stone, Eric; Murray, Terry; Minster, Angela; Rohrbaugh, Nick; Bishop, Michael; Chen, Eva; Baker, Joshua; Hou, Yuan; Horowitz, Michael; Ungar, Lyle; Tetlock, Philip (2015-05-01). "Identifying and Cultivating Superforecasters as a Method of Improving Probabilistic Predictions". Perspectives on Psychological Science. 10 (3): 267–281. ISSN 1745-6916. PMID 25987508. doi:10.1177/1745691615577794. Retrieved 2015-07-10.
- Jump up ^ Tetlock, Philip E.; Gardner, Dan (2015-09-29). Superforecasting - The Art and Science of Prediction. New York: Crown. ISBN 978-0-8041-3669-3.
- Jump up ^ Gardner, Dan (2010-10-12). Future Babble: Why Expert Predictions Fail - and Why We Believe Them Anyway. McClelland & Stewart.
- Jump up ^ "Unclouded vision". The Economist. 2015-09-26. ISSN 0013-0613. Retrieved 2015-09-24.
- Jump up ^ Zweig, Jason. "Can You See the Future? Probably Better Than Professional Forecasters". The Wall Street Journal. Retrieved September 25, 2015.
I think Philip Tetlock’s “Superforecasting: The Art and Science of Prediction,” ..., is the most important book on decision making since Daniel Kahneman’s “Thinking, Fast and Slow.”
- Jump up ^ Frick, Walter. "Question Certainty". Harvard Business Review. Retrieved 2015-09-26.
- Jump up ^ McEnroe, Colin; Wolf, Chion. "The Colin McEnroe Show". WNPR. National Public Radio. Retrieved October 1, 2015.
No comments:
Post a Comment