
Customer service is where your company gives your customers a sense of the products and business you are selling.
As a data scientist, can you improve the performance of customer service?
An important performance indicator is the average lead time between receiving a customer order and transmitting it to the warehouse for preparation.
In this article, we will introduce a methodology using statistical sampling to estimate this overall average lead time using 200 observations.
SUMMARY
I. Scenario
Problem Statement
You are the Customer Service Manager of an elevator parts supplier that produce and deliver engine parts for elevators.
Question
Can you estimate the average processing time with a confidence interval of 90% using your sample data?
II. Statistical Sampling
1. Notations
2. Application of the Central Limit Theorem
3. Confidence Interval
4. Final estimation
III. Conclusion
1. Other Lean Six Sigma Statistical Tools
2. Assessment of Container Loading Efficiency
Scenario
Problem Statement
You supporting the Customer Service Manager of an elevator parts supplier that produces and delivers engine parts for elevators.
Her team is in charge of order processing:
-
A Customer sends an order by phone or email with a requested delivery time (e.g., Customer orders 5 units of SKU X and would like to be delivered the same day at 10:00)
- Your team confirms the order and allocates it to the closest warehouse for preparation and shipment.
- The order is prepared and shipped from the warehouse using an express courier company.
You recently received many complaints from your customers because of late deliveries.
According to the warehouse manager, this is mainly due to delays in customer service processing orders.
During three months, you measured the order processing time of randomly selected operators and gathered 200 observations.
Can you estimate the average processing time with a confidence interval of 90% using your sample data?
Statistical Sampling
As we cannot measure the average processing time of all your operators for every order, we would like to estimate the total population average using these sample records.
Notations
To simplify the comprehension, let’s introduce some notations:
Application of the Central Limit Theorem
In a previous article, we used the Central Limit Theorem (CLT) to estimate the probability of a random variable P(X≥k), assuming that X was following a normal distribution.
The CLT also tells us:
Confidence Interval
Our objective is to know the population mean range [µ-b, µ+b] with a confidence of 90%.
And we know by the construction of the unit normal distribution that for P(-z≤Z≤z) = 0.9, we have z = 1.64
Finally, we can get our estimated range, or the population mean
Final estimation
count 200
mean 22.705
std 6.81
min 4.0
25% 18.0
50% 23.0
75% 27.0
max 41.0

We have,
n = 200
x̄ = 22.705 (min)
s = 6.81 (min)
The confidence interval is [21.96, 23.54]
Conclusion
For a confidence level of 90% and with moderate experimentation, we have a perfect estimation of the average lead time for order processing.
This approach can be used when process performance measurement is expensive and takes effort and time.
Can we trust the data?
However, it would be best to put effort into the experimental protocol to ensure that your sample data have been built based on a random selection of operators.
Have you heard about Lean Six Sigma Statistical Tools?
If you are interested in statistics for continuous improvement, check out this series of articles covering Lean Six Sigma concepts with Python.
Lean Six Sigma with Python – Kruskal Wallis Test
Lean Six Sigma with Python – Logistic Regression
Lean Six Sigma with Python – Chi-Squared Test
Can we use the same approach to evaluate other processes?
Containers Loading Optimization with Python
Due to the container shortage during COVID, the sea freight price exploded.

This put a lot of pressure on transportation management teams.
Do you want to assess the efficiency of sea container loading?
![Example of innefficient pallet loading [Right] - (Image by Author)](https://towardsdatascience.com/wp-content/uploads/2021/09/09PdpNVoRzf6jB3ya.png)
This example illustrates how inefficient loading strategies can increase costs.
The two pallets on the side won’t be loaded in the same container.
What about the additional cost? Assess the performance of your forklift drivers.
You can collect data on pallet loading (container size, number of pallets) to measure the performance and find patterns.
The insights can lead you to implement an algorithm for loading optimization like the one developed in the article linked below.
About Me
Let’s connect on Linkedin and Twitter. I am a Supply Chain Engineer who uses data analytics to improve Logistics operations and reduce costs.
For consulting or advice on analytics and sustainable supply chain transformation, feel free to contact me via Logigreen Consulting.
If you are interested in Data Analytics and Supply Chain, look at my website.
💌 New articles straight in your inbox for free: Newsletter 📘 Your complete guide for Supply Chain Analytics: Analytics Cheat Sheet