“Until recently the Automotive Aftermarket was provided data from key channel distributors indicating monthly sales activity and market share for various vehicle part categories. At the beginning of 2012 the consortium of companies that provided these data collapsed. Since then, parts suppliers and others in the Aftermarket have been searching for a new source of data to fill the void and this very issue is being discussed by industry representatives at the AAIA Fall Leadership Days conference in San Francisco this week.
In this post I propose a methodology for estimating market size and discuss how this estimate of total demand can form the basis for replacing, and perhaps improving upon, the market data previously provided by NPD.
The two key elements in producing category market size estimates are (1) vehicle registration data (referred to typically as VIO, i.e., vehicles in operation) and (2) Replacement Rates.
VIO data is available at various levels of geography (US, State, County, ZIP, Censust tract, and block group) and provided by Experian and Polk. This data is expensive but easy to acquire and utilize.”
Political Science professors from the University of Colorado recently published a study describing a model they constructed to forecast the result of the upcoming Presidential Election. They used various economic indicators, e.g., national and state unemployment rates, along with a handful of other variables to account for factors such as home state advantage and whether or not a candidate is an incumbent. I read the paper and, while I’m no political scientist, I think it’s a reasonably good piece of research. In the paper, the authors state that “the model successfully predicts every Electoral College victor“. This is a fair statement. However, in the Denver Post article the authors are quoted as saying ” it correctly forecast every winner of the electoral since 1980” and in the Huffington Post article the quote reads, “has correctly forecast every winner of the electoral race since 1980“. I realize that the differences are subtle but the quotes used by the Denver Post and the Huffington Post can lead to a complete misinterpretation.
In the research article the authors make it clear that the model was calibrated so that it successfully predicts the outcome of each election going back to 1980. But, the news articles from the Denver/Huffington Post reports the model’s accuracy as if it’s been in use since 1980 and has never been wrong in previous efforts to report election results. This is dangerously misleading.
If you’ve built a predictive model before you realize that, within reason, you can use a few tricks here and there to have the model work just the way you want it to work. Now, I’m not saying that the authors from the University of Colorado wanted a model that would forecast a Romney victory. No. What I’m saying is that the authors wanted a model that accurately forecast the winner of every previous election in their study. In other words the historical accuracy of the model wasn’t an accident (“Hey, it turns out that our model – produced in a vacuum – has worked for 30 years!”); rather, the model was developed in order that its forecasts would have been accurate. Now, the authors of the research don’t make such a claim but the journalists who reported the story fail to clarify how the study should be interpreted.
So, what’s the point of this blog post? Well, this type of reporting is potentially dangerous just like early reporting of exit polling results can be dangerous. This information may influence voters – they may choose not to vote (for either candidate – a supporter may think his candidate has the election in the bag and doesn’t bother voting or may believe his candidate doesn’t stand a chance and may not vote). In any case, journalists who report on predictive models, and polling results for that matter, need to have a better understanding of statistics so that they are able to report these studies with greater precision and with appropriate context. They do their readers a disservice when they report forecasts too loosely.
In addition, I think the researchers have a responsibility, especially in their capacity as public employees, to clarify how the model was constructed and what they mean, exactly, by stating that it has been correct going back 30 years. They were clever to publish a study that would garner plenty of media attention and I commend their effort to do salient applied work rather than something more theoretical and obscure. That said, they should be certain that their research is properly understood by the public they are paid to serve.
For the serious scholar of cartography, you’ll want to check out the History of Cartography series by J.B. Harley and David Woodward. There are several volumes including Cartography in the European Renaissance: Part 1 and Part 2.
Thanks to everyone who voted and submitted nominees! Stay tuned for the 3rd inductee!
In my initial post Waldo Tobler became the first Geographer Hall of Fame inductee. I received lots of good comments and several submitted names for consideration. Now I’d like to get reader input for the next inductee. If you’re interested in participating, please follow the link below to vote for one of the many worthy nominees:
Also, if you’d like to nominate a geographer for the next round of voting please leave a comment below or Tweet your nominee to me on Twitter (and follow me too): @justinholman.
Thanks for participating! I will share the results before the end of July.
I was having trouble downloading the PDF version of this map so I’m posting a smaller jpg version here in hopes that it might help someone struggling to access this information. This map will be more difficult to read but you should be able to click on the image and then zoom in using browser controls. If you’re familiar with the Colorado Springs road network you should be able to identify your neighborhood. The shaded area is still under mandatory evacuation notice.
Big kudos to the Colorado Springs GIS department who must be working around the clock to keep up with demand for their mapping and analytical services.
Imagine that Rodney Dangerfield was a cartographer. What would he say about this map? Please submit your caption idea and I will select a winner on Monday.
Full press release from CoreLogic with graphics can be found on PR Newswire.
This year’s explosive wildfire activity in Colorado is a reminder that during the housing boom many people built homes in risk-prone areas of the state. Today over 1 Million Colorado residents live in a “red zone”, i.e., an area that will eventually, inevitably see wildfire activity. Millions more live in locations with moderate risk. Here’s a map of fire risk in Colorado.
Terrific map from Nate Silver’s Five Thirty Eight blog at the New York Times. Apparently my move from Michigan to Colorado increased the value of my vote significantly. The map is interactive so if you visit the Five Thirty Eight blog and find the map you can mouse over your state and see how important your vote might be in November.
Interesting job post appeared in my inbox. Check it out:
Postdoctoral Research or Software Engineering Position, Computational Landscape Modeling, University of Colorado, Boulder. Applications are invited for either a post-doctoral researcher with a background in the Earth Sciences and in scientific software development, or a software engineer with experience in geoscience modeling. The successful candidate will be based at the University of Colorado, Boulder, and will collaborate with a team of scientists at Tulane University, the University of Washington, and the University of New Mexico. The project entails building stand-alone software components that can be linked to explore different questions about landscape evolution over various spatial and temporal scales. The researcher will be responsible for helping to design, implement and test software, and coordinate amongst team members. The position requires knowledge of an open-source, object-oriented programming language. Experience in C++ is especially desirable, as is a background in geomorphology or hydrology, and/or experience with model-coupling frameworks. The position will have the support of the Community Surface Dynamics Modeling System (CSDMS) infrastructure, based in Boulder, Colorado. Some travel to work with team members at different Universities will be necessary. The position is available starting in late summer or early autumn 2012. Initial funding is for one year, although the position may be extended for an additional year based on performance. To apply, please send a CV including contacts for three references and a cover letter that describes your experience with software development and numerical modeling to Greg Tucker, [email protected]. Priority will be given to applications received before July 23, 2012.
I read a recent report from Climate Central that attempts to illustrate temperature change in the US. You can find the report here. I’ll keep my review brief: I think it sucks. The main reason I don’t like the report is because the researchers decided to aggregate temperature change by State.
Why is this a problem? It doesn’t make any sense to map these data by State. Why couldn’t the researchers have taken a little time to think about an appropriate geographical scale? A quick web search using keywords “map” + “climate” would have revealed that there is an existing geography for these data called U.S. Climate Divisions. The US Climate Divisions aren’t ideal so they could have used Level III or IV Ecoregions or perhaps interpolated the data onto a high resolution spatial grid and created an animation. Or they could have called or emailed a geographer who knew what they were doing instead of stumbling blind into geographical data analysis. They could have placed a call to Glen Macdonald at UCLA who published a terrific report on drought in the US Southwest with maps of temperature change and drought indices in the US. There are dozens of other researchers who could have helped them create a good map. But, in the age of Google Maps everyone thinks they’re an accomplished cartographer.
Here’s are three better options for mapping US climate data:
Click on any of the 3 images above to link to the source. The last map above is one of several images that form an animation of climate simulations in the Pacific Northwest. The animation is used by in a Climatology course at the University of Oregon.
Using the US Climate Divisions or Ecoregions or a high resolution grid would have allowed the researchers to produce a much more nuanced and informative map. Aggregating these data by State is an oversimplification that likely disguises important spatial patterns of change and may be misleading readers.
I noticed that the 2 PhDs who were listed as authors on the report are affiliated with Duke University. Duke is obviously an excellent university but, like many universities in the United States, it has chosen not to offer a Department of Geography. What a shame.