Design Thinking for Data Science: A Human-Centric Approach to Solving Complex Problems

In the data-driven world, successful data science isn’t just about algorithms and statistics – it’s about solving real-world problems in ways that are impactful, understandable, and user-centered. This is where Design Thinking comes in. Originally developed for product and service design, Design Thinking is a problem-solving methodology that helps data scientists deeply understand the needs of their end-users, fostering a more human-centric approach to data solutions.

Let’s dive into the principles of Design Thinking, how it applies to data science, and why this mindset shift is valuable for creating impactful data-driven solutions.

What is Design Thinking?

Design Thinking is a methodology that encourages creative problem-solving through empathy, ideation, and iteration. It focuses on understanding users, redefining problems, and designing innovative solutions that meet their needs. Unlike traditional problem-solving methods, Design Thinking is nonlinear, meaning it doesn’t follow a strict sequence of steps but rather encourages looping back as needed to refine solutions.

The Five Stages of Design Thinking and Their Application to Data Science

Design Thinking has five main stages: Empathize, Define, Ideate, Prototype, and Test. Each stage is highly adaptable and beneficial for data science projects.

1. Empathize: Understand the User and Their Needs

Objective: Gain a deep understanding of the people involved and the problem context.

  • Data Science Application: Instead of jumping straight into data analysis, data scientists can start by interviewing stakeholders, observing end-users, and gathering insights on the problem context. This might involve learning about business needs, pain points, or specific user challenges.
  • Outcome: Developing empathy helps data scientists understand the human impact of the data solution. It frames data not just as numbers but as stories and insights that need to be translated into actionable outcomes.

Example: For a retail analytics project, a data scientist might meet with sales teams to understand their challenges with customer segmentation. They might discover that sales reps need more personalized customer insights, helping data scientists refine their approach and data features.

2. Define: Articulate the Problem Clearly

Objective: Narrow down and clearly define the problem based on insights from the empathizing stage.

  • Data Science Application: Translating observations and qualitative data from stakeholders into a precise, actionable problem statement is essential in data science. The problem statement should focus on the “why” behind the project and clarify how a solution will create value.
  • Outcome: This stage provides a clear direction for the data project, aligning it with the real-world needs and setting the foundation for effective data collection, model building, and analysis.

Example: In a predictive maintenance project for manufacturing, the problem statement could evolve from “analyze machine failure” to “predict machine failures to reduce downtime by 20%,” adding clarity and focus to the project’s goals.

3. Ideate: Generate a Range of Solutions

Objective: Brainstorm a variety of solutions, even unconventional ones, and consider multiple perspectives on how to approach the problem.

  • Data Science Application: In this stage, data scientists explore different analytical approaches, algorithms, and data sources. It’s a collaborative brainstorming session where creativity and experimentation take center stage, helping generate diverse methods for addressing the problem.
  • Outcome: Ideation leads to potential solution pathways and encourages teams to think beyond standard models or analysis techniques, considering how different data features or combinations might offer unique insights.

Example: For an employee attrition prediction project, ideation might involve brainstorming potential data features like employee tenure, manager interactions, and work-life balance. It could also involve considering various algorithms, from decision trees to deep learning, based on data availability and complexity.

4. Prototype: Build and Experiment with Solutions

Objective: Create a tangible representation of the solution, often in the form of a minimum viable product (MVP) or early-stage model.

  • Data Science Application: Prototyping in data science could involve building a quick initial model, conducting exploratory data analysis, or developing a dashboard that visualizes preliminary results. It’s about testing ideas rapidly to see which direction holds promise.
  • Outcome: Prototyping allows data scientists to see early results, gather feedback, and refine their models and visualizations. It’s a low-risk way to iterate on ideas before investing significant resources in a final solution.

Example: For a churn prediction project, the data team might create a basic logistic regression model and build a simple dashboard to visualize which factors are most influential. They can then gather feedback from the sales team on what insights are valuable and where they need more detail.

5. Test: Validate the Solution and Iterate

Objective: Test the prototype with real users or stakeholders, gather feedback, and make adjustments based on what you learn.

  • Data Science Application: Testing might involve showing stakeholders preliminary results, gathering feedback on model accuracy, or evaluating the solution’s usability. It’s about validating assumptions and refining the model or analysis based on real-world feedback.
  • Outcome: The testing phase helps data scientists ensure the model aligns with business objectives and addresses the end-users’ needs. Any gaps identified here allow for further refinement.

Example: If the initial churn model fails to predict high-risk customers accurately, data scientists can refine it by adding new features or using a more complex algorithm. Continuous feedback and iterations help the model evolve in alignment with user expectations and business goals.

How to Implement Design Thinking in Data Science Projects

  • Build Empathy: Hold interviews, run surveys, and spend time understanding end-users and stakeholders.
  • Define Clear Problem Statements: Regularly revisit the problem statement to ensure it aligns with real user needs.
  • Encourage Diverse Perspectives: Foster a team culture that values brainstorming and out-of-the-box thinking.
  • Prototype Early and Often: Don’t wait for the perfect model – use MVPs to test hypotheses and gather quick feedback.
  • Stay Iterative: Treat data science as an ongoing process, iterating on models and solutions based on user feedback and new insights.

For more details, read this interesting article written by Bill at DataScienceCentral website.

Credit: DataScienceCentral

Final Thoughts

Incorporating Design Thinking into data science transforms the way problems are approached, moving beyond data and algorithms to create solutions that are effective, empathetic, and impactful. This methodology is particularly valuable in data science, where the complexity of models can sometimes overshadow their practical applications.

By thinking more like a designer, data scientists can build solutions that not only solve technical challenges but also resonate with end-users and deliver measurable value. In an industry that’s increasingly focused on impact, adopting a Design Thinking mindset might just be the key to unlocking the full potential of data science.

T-shaped vs V-shaped path in your Analytics career

We start with learning multiple disciplines in an industry and then niche down to a specific skill that we master over the period of time to get expertise and become an authority in that space.

Typically, many including me follow a T-shaped path in the career journey where horizontal bar ‘T’ refers for wide variety of generalized knowledge / skills whereas vertical bar ‘T’ refers to depth of knowledge in a specific skill. For instance, if you’re a Data Scientist, you still do minimal Data Pre-Processing steps before doing the Exploratory Data Analysis, Model Training / Experimentation and Selection based on evaluation metrics. Although a Data Engineer or a Data Analyst, primarily works on data extraction, processing and visualization, a Data Scientist might still need to be familiar in order to get the job done on time without depending on the other team members.

Data Scientist’s vertical bar ‘T’ refers to crafting the best models for the dataset and horizontal bar ‘T’ could refer to Data processing (cleaning, transformation etc.) and visualizing the KPIs in the form of insights for the business to take informed decisions.

Strategy & Leadership consultant and author, Jeroen, comes up with a V-shaped path which makes sense in our contemporary economic situation where layoffs news are on the buzz across many MNC companies.

In terms of similarities, the author, reiterates that both models address the fact that understanding one focus area deeply and having shallow knowledge across other areas. V-shaped model refers to having one deep knowledge and a lot of adjacent knowledge areas which are not deeper but not shallow either, somewhere in between. Jeroen describes as, “It is medium-deep, medium-broad, enabling us to be versatile and agile.”

For illustration, if the Data Scientist aspires to go above and beyond the expectations, he/she can technically collaborate with Data Engineers, performs AI/ML modeling stuffs, builds reports/dashboards, generate meaningful insights, and enable end-user adoption of insights. It has a combination of hard and soft skills! Soft skills such as storytelling, collaboration with peers, project management etc., Over the period of time, as one repeats this whole process, they can get better and better (develop deeper knowledge) with model development and management, and develop adjacent soft skills to excel at work.

In my view, I think, we start with a T-shaped path and eventually, it morphs into a V-shaped career path as we put our hard-work on one skill and also develop its associated adjacent skills. And, it applies to any field that you’re in.

How long do you think it would take this transformation to attain a V-shaped path? Will this take about 10,000 hours (~a decade) as per Gladwell’s book: “Outliers” to become an expert? Maybe, yes! Sooner, the better it is!!

I’ll leave you with a three-phase approach to becoming an expert according to the author Jeroen.

Image Credits: https://www.linkedin.com/in/jeroenkraaijenbrink/