Skip to main content

Reducing Workforce Turnover using Anomaly Detection and NLP

Maintaining an engaged workforce is essential to any organization looking to not only minimize the costs associated with hiring new personnel but also maximize productivity through engaged employees. Our senior data scientist, Jeremiah Lowhorn, partnered with one of our clients to analyze the risk factors that lead to employee turnover and how to mitigate them. We sat down with him to learn about how he was uniquely suited to solve such a complex problem.

Can you tell us about your background and current role at Data Tapestry?


My title is senior data scientist, and I’m currently working on my second master of science, this time in information management. Before Data Tapestry, I worked as a senior software engineer at Cigna focusing mainly on big data and data science projects. Prior to that role, I was working at US cellular as a data scientist. While there, I focused mainly on time series analysis and predictive modeling.

Tell me about the problem you were asked to solve and what were the client’s expectations.


The scope of the project was essentially to figure out why physicians were turning over. They didn’t really have any other objectives outside of that. I started with a query to generate a time series for the physician data. It was a record of each physician id and every day that they were employed. Additionally, there was an opportunity to enrich the time series with salary information and other physician attributes. I was able to incorporate an additional 200 fields into the analysis.

So how did the results turn out?


Things turned out well! We were able to provide an end product that gave the client additional insights into a critical part of their business. The client wanted to look at turnover over time. Additionally, I decided to enhance that goal by looking at trend detection to answer some questions like: Have we seen a spike in turnover over time and how can we identify that? So, I built an anomaly detection algorithm detects if there’s been a larger than usual increase in turnover over a 3-4-month period.

I also saw an opportunity to incorporate the physician exit surveys which supplied self-reported reasons for leaving. This led to creating an NLP model to read physicians exit surveys and classify those responses into specific categories as well as classify the sentiment of the response.

How did you partner with the client to achieve these results?


My stakeholder did not offer a ton of direction because he trusted my recommendations. We also wanted to let the data tell the story. I defined the roadmap on how to approach the problem to maintain forward progress. After our initial results, the client and I studied the outputs to tie their organizational knowledge to the analysis. This allowed us to refine the models to achieve even better results.

If you are interested in learning more about Data Tapestry or this project, email us at: business@datatapestry.ai or click the button below to visit our website. 

Comments

Popular posts from this blog

Transforming and Accessing Data through Custom Built Pipelines

One of the biggest hurdles in data analysis is just getting access to data in the first place. At Data Tapestry, we offer end-to-end analytics services beginning with data acquisition, performing analytics, and providing end user products. Keith Shook walks us through how to maintain data security and integrity when dealing with a variety of situations.
Tell us a little about your background and your role at Data Tapestry.Currently, I’m a senior data engineer, but I actually started off as an intern ingesting data into Postgres and SQL databases. I then shifted into visualization using D3, a javascript library, but we found that Tableau was much more efficient. Since then, I’ve gained a variety of experience using scala, Hive, AWS, and building clusters.Can you walk us through a project you’ve worked on?Data engineering is pretty straightforward as far as the process goes. You get the data, ingest it into the database, and then hand it off to the data scientist. You have to be flexible…

Consulting, Designing, and Coding: the Many Roles of a Developer

At Data Tapestry, we pride ourselves in deeply understanding your business needs and delivering solutions in a hands-on or consulting capacity. Our staff not only performs analytics, but we also build and support custom software products. Philip Vacarro, our full-stack software developer, explains how he partners with multiple clients in different capacities to deliver production software solutions, advise on data architecture, and provide product support.
Can you tell us about your background and current role at Data Tapestry?
I’m a full stack software developer, and I serve as the first stop for new clients when it comes to consulting on data architecture and other products we’ve built. I’ve worked as a software engineer for Siemens in the research and development. We focused on interventional imaging. After that, I worked as a full stack developer at ORNL. We collected terabytes of data submitted from scientists all over the world.  The data was then centralized in an application so …