Just because you can doesn't mean you should
When I started my career in People Analytics, I was eager to see how far data science could go in the HR space.?For all the mysteries of why people do what they do, all of a sudden we had an opportunity to make people and their behaviors a little more predictable.?We finally had access to data, technology, and expertise to answer the tough questions with statistical models instead of guesswork.?As my career has evolved, however, I realized it's not so much what you can do, but how you do it and why that matters more.
Take predicting attrition, for example.?These capabilities are now commonplace, and can even come straight out of the box from most vendors. The challenge is no longer having a model, but knowing how to use it and creating the right strategy that benefits both employees and the company.
With any project, the starting point has to come from solving business challenges.?In the case of turnover, that's easy - no company wants to see good talent walk out the door when it's preventable.?And for employees, if we can satisfy their needs and keep them happy, everyone wins.?But there is an important question you need to start with - how do you responsibly use a model that can tell you what's expected to happen? There are many things you could do:
Some of these ideas are sensible and low-risk, while others could easily have unintended consequences.?Getting a broad and diverse set of stakeholders involved is necessary to examine the plan from all angles and discuss potential good and bad outcomes. The best approach will always take into account the worst case scenarios:
There are also risks to the model's integrity to consider:
领英推荐
Only after you've planned around all these corners should you invest the time and resources needed to build and deploy a predictive attrition model. But once you reach this stage, there are another set of decisions to slow down and consider - including, what type of data is appropriate to include in your model. There is a lot of data to choose from. There are group-level factors that pose little risk in analyzing and understanding - like global factors (e.g., recessions, pandemics), industry unemployment and hiring trends, major company events (e.g., RIFs, acquisitions), leadership changes, division-wide trends, seasonality of turnover (e.g., stock vesting dates, bonus payouts), and other general inputs.
These will be helpful to include in the model, but on their own insufficient without also including individual-level data which has much more predictive utility. You should divide this type of personal information according to sensitivity of the data and potential for misuse. There is a lot of generally accessible information that can be useful without being overly intrusive - things like hire type, education, job level, job function, location, tenure, time since last promotion, number of manager changes, and perhaps other elements like participation in trainings and membership groups, if available.
Most People Analytics teams have access to more sensitive data points, as well, and these should be governed with even more scrutiny, and in accordance with why the data are being collected in the first place. While some data could be highly predictive, it may or may not be appropriate to use. These include things like age, gender, ethnicity, survey responses, salary and performance data (and changes in ratings and rewards over time), unvested stock or retention agreements, commute distance, organizational network data, or other information companies might be collecting about their employees.
Making a misstep here would be more dangerous than going to Jurassic Park during a power outage. Jeff Goldblum put it best:
"Your scientists were so preoccupied with whether or not they could, they never stopped to think if they should."
So what should you do? All this is not to discourage using data to arrive at answers and inform decisions - we have to be data informed! But we need to move thoughtfully, and in coordination with the right stakeholders, the number of which should naturally increase with the size and importance of the problem you're trying to solve. For a project like this, you should also be leveraging a structured data privacy checklist and impact assessment to evaluate and minimize potential risks of the project against the intended benefits - what data will be collected, how it will be secured, how long it will be retained, who will have access, how it will be used and why. There should be transparency with employees and alignment with informed consent practices and employee guidelines for how personal data will be used, as well as compliance with GDPR and other regulations and requirements.
Overall, what should be evident from this example, is that data science skills are necessary, but not sufficient, for People Analytics teams to be successful. The ability to think strategically, partner cross-functionally, and lead with ethics and data governance are all critical to ensure research is carried out in a way that will actually benefit employees. In the case of predicting attrition, the approach should be scalable, standardized, measurable, and strictly governed. That said, it's not a one-size-fits-all solution, so I'd love to hear from other professionals in this space what has been your experience with applying this type of model? How do you recommend using it or not using it?
Director, OD- Sensing & Analytics at Teva Pharmaceuticals
2 年Well said! That's the real reason why people analytics professional should find it hard to sleep at night. Even if you create a great model that is super reliable, you are only starting the journey to make it helpful... "Do no harm" and "with great power, comes great responsibility" should be added to the Jurassic Park quote :)
Vice President, Insight 222 LLC
3 年Fantastic article Micah! This is critical information for any #peopleanalytics practitioner, because the points Micah is raising touch on #dataethics; the surest way to destroy trust with employees (and works councils) is to miss either data collection or analysis sensitivities. Missing them closes the door to future data-based inquiries. What we've seen work really well is exactly what Micah said: be open and up front with employees about the purpose and goal of the data collection and analysis. What makes this article particularly insightful is the pragmatic depth and detail of Micah's #turnover / #attrition modeling example. Highly recommend!!
Clickhouse, Kafka, Airflow
3 年Thank you for sharing, many great points! "But we need to move thoughtfully, and in coordination with the right stakeholders, the number of which should naturally increase with the size and importance of the problem you're trying to solve." Speaks volume :)
AI-Powered People Tech
3 年Great article Micah Lueck - 100% agree that the answer is not the model. We have seen success when as much / more energy is invested in educating stakeholders in what the results of the model mean, providing them with options for actions to change the outcome. And then following through on a program in a specific piece of the organization to test "fit for purpose" - of the whole data informed strategy - before scaling this out more broadly. The people and process are as key as the data and the technology.