Predicting Elections and Human Behavior: Stop Asking and Start Observing

Predicting Elections and Human Behavior: Stop Asking and Start Observing

In one of the wildest and most contentious elections in U.S. history, many Americans had trouble seeing eye to eye. About the only thing we can all agree on is that the pollsters completely blew it when it came to predicting our next president.

Some of these inaccurate poll results call into question the validity of modern-day statistics and data science.

Political polling is a beast all on its own, but the process has much in common with the AI-powered software being built by some of the world’s most influential tech companies.

The truth is, pollsters aren’t the only ones making poor predictions. There’s a lot of questionable data science masquerading as AI.

Societal implications

If the pollsters can’t accurately call an election, we’d better take a hard look at AI-fueled technologies that could have a profound effect on our lives, such as self-driving cars and heart-monitoring devices.

With all of the AI hype in today’s market, everybody labels themselves data scientists. In fact, simply updating your LinkedIn profile to include the term “data science” could increase your income overnight, with the position earning on average $123,000 per year – 113% higher than the average annual salary. However, this title doesn’t make someone a master of the craft, as even smart data scientists need access to large amounts of accurate and timely data to make reliable predictions.

So, why did the pollsters miss on this election so badly? Did they introduce bias? Or did voters misrepresent their intentions – either on purpose or because they were unsure – making it harder to get reliable data? Or, was it both?

Accurate predictions require accurate data

One major problem with the election polls is that they did not consider a broad enough and representative data set. Insufficient data often leads to bias and inaccurate results.

While most of the pollsters asked Americans who they planned to vote for in the election, the USC/LA Times Daybreak Poll took a different approach.

This poll asked people how likely they were to vote for a particular candidate. By continuously gathering new data each week that captured shifting sentiments over time, they accurately predicted Donald Trump’s victory while others continued to project Hillary Clinton as the winner.

It’s important to remember that not all data is created equal. In this election, telephone polls proved much less reliable than online polls. Perhaps voters didn’t want to acknowledge they were supporting Trump on the phone.

Here’s the thing that most people miss: Observed data is inherently more accurate than self-declared data. People don’t always reveal their true intentions when you ask them. Observing their behaviors is a much more accurate predictor of what they will do next than what they say they will do.

LinkedIn and Facebook updates are a good example of self-declared data, and it’s easy to see why this type of data is often unreliable.

When was the last time you saw somebody post on LinkedIn: “I just got fired for stealing a stapler”?

People really do get fired for theft, but they are likely to spin it on social media as: “I’m excited to report I’m exploring new career opportunities.”

This is why the Daybreak Poll outperformed the others. By paying attention to trends in sentiment over time, these brilliant researchers effectively turned self-declared data into observed data.

A better way to predict behavior

There’s a better way to do polling, and there’s a better way to do AI.

First, you must collect large amounts of timely and accurate data that is observed rather than self-declared. Then, you must analyze it with self-learning systems that continuously grow smarter over time and inject these insights directly into people’s workflows with applications that make them immediately actionable.

This is what real AI looks like, and it’s the only way to produce accurate predictions that drive better business results.

The big lesson from this election is that if we want to predict human behavior, we must stop asking and start observing.

Dave Elkington is the CEO and founder of InsideSales.com, the industry’s leading cloud-based sales acceleration technology company. All of InsideSales’ innovative products are fueled by the predictive insights of Neuralytics, a self-learning engine with more than 100 billion sales interactions and counting.

@DaveElkington

Jeremiah Johnson

Head of Engineering @ Signals

7 年

Interesting thoughts. I suppose the trick is in finding the observable behaviors and figuring out which ones matter.

回复
Matt Styles

Staff Software Engineer at Lendio

7 年

"Perhaps voters didn’t want to acknowledge they were supporting Trump on the phone." Are you trying to insinuate something here? Ha, kidding ;). I think you're spot on in that the failure included a combination of bias and a lacking data set. The media, particularly mainstream outlets, often have their own agenda which motivates them to skew poll results (and thus public perception) through the way they gather the polling data or simply through the way they report on the data. This is most evident when you consider third party candidates. Consistently marginalized and often outright ignored, excluded from polls, and yet vote totals for candidates like Jill Stein and Gary Johnson were sometimes 10x the margin that Hillary or Donald one by in the state. Just emphasizes all the more how important it is to have a broad data set that accounts for several different variables, even if you don't think the variables are significant at the time. Along those lines, you mention the value of observed behavior, so how do you find the key data points to observe? You can't observe everything, and even if you could, much of it would be at best fluff and at worst interference from real insight. What are your thoughts?

Mark A.

DATA SYSTEMS/BI/CRM/BUSINESS TECHNOLOGY/OPERATIONS LEADER: Business Systems | Operations | Analytics | Sales | Marketing | FP&A | Accounting | Professional Service | Customer Success

7 年

Excellent, cutting-edge insights. I found it particularly interesting that the USC/LA Times Daybreak Poll was highly discredited and criticized by the very pollsters who were dead wrong. It will be intriguing to see how quickly (or slowly) other pollsters shift their focus from that of asking to observing.

Martin Spanik

Vice President, Go-to-Market (GTM) & Strategic Alliances, Cybersecurity

7 年

Great article, thanks for sharing !

回复
Shayne Jensen

Solutions Architect with 10+ years experience focused on data analytics and client success. AWS Certified Solutions Architect - Associate. Experienced in Revenue Operations, Healthcare, Higher Education, and SaaS.

7 年

Great article. I think pollsters may have done better this year if they hadn't relied on last year's models. If They had observed a bit more, they may have realized the traditional demographic models from past elections had shifted and in some areas broken altogether. Observing and using accurate and timely data to develop a shifting predictive model may have allowed pollsters to use a model that corrects itself as opinions within demographics shift. With a self-generating model they may have been able to completely shrug off their assumptions based on past demographics. This election was an excellent example of why using old data isn't only inaccurate, it's misleading.

要查看或添加评论,请登录

社区洞察

其他会员也浏览了