Skip to main content

Data Discovery and Understanding across an Organization

Much of what we do at Data Tapestry is helping our clients gain an understanding of their own data. If you work at a large organization, data access can be limited. This means you may not know what questions to ask of your data or what you can measure with your data. Hillary Rivera walks us through how she partnered with multiple stakeholders to gain insight from a variety of data sources.


Can you tell us about your background and current role at Data Tapestry?

I started out my career thinking I would eventually go to medical school, but I quickly realized I wanted to work in analytics. I studied public health and worked in healthcare data management for a while. Then I went back to school to study business analytics at UT. Since then, I’ve enjoyed working in ecommerce and software and technology. Now I’m back in healthcare working as a data scientist at Data Tapestry.

 

Tell me about how your project started with the client.

I had a unique experience with my client because I had two very disjoint responsibilities. One was to automate a report that showed how many dollars were being spent on physician liability claims. The other was to identify how physician stipend dollars were being allocated to physician leadership and to see if that allocation could be optimized based on facility volume.

 

How was each project different?

The report automation piece was very straight forward. I would receive a spreadsheet of physician claims and apply some filters to the data based on the parameters defined by the stakeholder.


The stipend dollar analysis was much more involved because it required us to define what we would consider as a stipend. There were also many subject matter experts involved because

there was no definitive knowledge of where all the data elements were located. 

 

Can you describe your relationship with your stakeholder and how you managed the project to ensure forward progress?

For the stipend project, we had a weekly check in with most of the subject matter experts and the business partners. We discussed progress and defined tasks needed for forward progress. Outside of that, I would meet with my business partner more frequently since we were both learning where all the data elements lived and how they tied together.

 

Describe the nature of the deliverable you provided and how it helped the client.

The physician claim report ended up turning into a data validation tool. So, the client ultimately wanted to reconcile what she was receiving from one system with what was being ingested by their data engineering team.  For the stipend analysis, I delivered a master data set in which all the stipend dollars could be traced back to an individual and then to a facility.  I also provided a script that provided a descriptive look at the relationship between dollars and facility volume.

If you are interested in working with Data Tapestry, visit our website at datatapestry.ai or email us at business@datatapestry.ai

Comments

Popular posts from this blog

Transforming and Accessing Data through Custom Built Pipelines

One of the biggest hurdles in data analysis is just getting access to data in the first place. At Data Tapestry, we offer end-to-end analytics services beginning with data acquisition, performing analytics, and providing end user products. Keith Shook walks us through how to maintain data security and integrity when dealing with a variety of situations.
Tell us a little about your background and your role at Data Tapestry.Currently, I’m a senior data engineer, but I actually started off as an intern ingesting data into Postgres and SQL databases. I then shifted into visualization using D3, a javascript library, but we found that Tableau was much more efficient. Since then, I’ve gained a variety of experience using scala, Hive, AWS, and building clusters.Can you walk us through a project you’ve worked on?Data engineering is pretty straightforward as far as the process goes. You get the data, ingest it into the database, and then hand it off to the data scientist. You have to be flexible…

Reducing Workforce Turnover using Anomaly Detection and NLP

Maintaining an engaged workforce is essential to any organization looking to not only minimize the costs associated with hiring new personnel but also maximize productivity through engaged employees. Our senior data scientist, Jeremiah Lowhorn, partnered with one of our clients to analyze the risk factors that lead to employee turnover and how to mitigate them. We sat down with him to learn about how he was uniquely suited to solve such a complex problem.
Can you tell us about your background and current role at Data Tapestry?
My title is senior data scientist, and I’m currently working on my second master of science, this time in information management. Before Data Tapestry, I worked as a senior software engineer at Cigna focusing mainly on big data and data science projects. Prior to that role, I was working at US cellular as a data scientist. While there, I focused mainly on time series analysis and predictive modeling.
Tell me about the problem you were asked to solve and what were t…

Consulting, Designing, and Coding: the Many Roles of a Developer

At Data Tapestry, we pride ourselves in deeply understanding your business needs and delivering solutions in a hands-on or consulting capacity. Our staff not only performs analytics, but we also build and support custom software products. Philip Vacarro, our full-stack software developer, explains how he partners with multiple clients in different capacities to deliver production software solutions, advise on data architecture, and provide product support.
Can you tell us about your background and current role at Data Tapestry?
I’m a full stack software developer, and I serve as the first stop for new clients when it comes to consulting on data architecture and other products we’ve built. I’ve worked as a software engineer for Siemens in the research and development. We focused on interventional imaging. After that, I worked as a full stack developer at ORNL. We collected terabytes of data submitted from scientists all over the world.  The data was then centralized in an application so …