# When Uncertainty Becomes Productivity: Tsinghua-Meituan Research Reveals the Probabilistic Modeling Revolution in On-Demand Food Delivery Service Time
## Research Background
The global on-demand food delivery (OFD) market is experiencing unprecedented explosive growth. According to Statista’s Digital Market Outlook, industry revenue increased from $91 million in 2018 to $107 million in 2019. As China’s largest platform, **Meituan** had gathered approximately 4 million riders, 6.5 million restaurants, and 400 million active users by 2020, with daily order volume exceeding 40 million. However, behind this prosperous picture, a long-simplified technical “black box” is constraining the further leap in fulfillment efficiency—**uncertainty modeling of service time**. Service time specifically refers to the period from when a rider stops riding and enters a customer’s building to when they complete the delivery and leave the building. It may seem insignificant, but it actually aggregates the most complex real-world variables in instant delivery: time spent climbing stairs in old residential areas without elevators, elevator waiting during peak hours in office buildings, repeated searching for ambiguous addresses, communication delays at community access control… These random factors together constitute a **highly uncertain time interval**. Traditional scheduling systems, however, crudely simplify it to a fixed value (such as uniformly setting it to 3 minutes), leading to distorted estimated time of arrival (ETA), inaccurate rider route planning, and compromised user satisfaction. More critically, as business expands to lower-tier cities and special scenarios (hospitals, schools, industrial parks), the volatility of service time increases exponentially, pushing the original deterministic assumptions to the brink of failure. Therefore, **upgrading service time from a “deterministic value” to a “probability distribution”** is not only a technical proposition for algorithm optimization but also a core strategic issue affecting platform fulfillment costs, rider income fairness, and user experience.
## Methodology Interpretation
The collaborative research between Tsinghua University’s Department of Automation and Meituan’s technical team, “Modeling stochastic service time for complex on-demand food delivery,” proposes a breakthrough framework: **modeling random service time as a Gaussian Mixture Model (GMM) and designing a Hybrid Estimation of Distribution Algorithm (HEDA) for intelligent parameter learning**. The essence of GMM is to approximate any complex distribution using a weighted combination of multiple Gaussian distributions (i.e., “bell curves”). Each Gaussian component can be understood as a typical service scenario (e.g., “smooth elevator,” “low-to-mid-rise stairs,” “difficult address finding”), with its weight reflecting the occurrence probability of that scenario. The key innovation of the research lies in transforming GMM parameter estimation into a **dynamic clustering problem**: the algorithm does not need to pre-specify the number of components K but lets the data itself “speak” about how many patterns are needed to characterize its intrinsic structure. HEDA designs four mechanisms for this purpose: First, **problem-specific encoding**, representing each solution as a cluster assignment vector and achieving adaptive simplification of component count by merging similar clusters; second, **Chinese Restaurant Process (CRP) initialization**, drawing on Bayesian nonparametric ideas to generate diverse initial solutions and avoid local optima; third, **weighted learning mechanism**, adjusting the update intensity of the probability model based on solution quality during iterations to achieve “selective breeding”; fourth, **maximum likelihood local intensification**, performing gradient-like fine-tuning on high-quality solutions to improve convergence accuracy. The objective function cleverly balances “goodness of fit” and “model simplicity”—using Wasserstein distance to measure the difference between GMM and the real distribution (the smaller, the better), while penalizing the component count K (to prevent overfitting), with the λ coefficient adjusting the weight between the two. The core philosophy of this methodology is: **letting the algorithm autonomously optimize between “capturing details” and “maintaining simplicity,”** thereby producing lightweight models that both fit business reality and are easy to engineer and deploy.
## Core Findings
The empirical part of the research demonstrates a rigorous two-tier validation system. In **offline experiments**, the team trained and tested based on 120 million service time records from Beijing, Shanghai, and Guangzhou in Q3 2020. Compared to baselines (EM algorithm, K-means+EM, traditional EDA), HEDA-GMM reduced the average Wasserstein distance by 31.7%, and the learned component count stabilized at 4–6 (consistent with business intuition). The mean and variance of each component clearly corresponded to a type of real-world scenario (e.g., “mean 98 seconds, small variance” corresponds to standard residential elevator scenarios; “mean 217 seconds, large variance” corresponds to complex address-finding scenarios). More importantly, the model demonstrated strong **anomaly pattern recognition capability**: it automatically extracted a “long-tail component” from historical data that accounted for only 1.3% but had a mean as high as 463 seconds. Business review revealed this corresponded to hospital inpatient department deliveries (requiring multiple security checks and ward searches). This insight directly prompted the product side to launch a “special reminder for medical locations” feature. However, the truly valuable practical achievement of the research was not limited to offline metric improvements but was demonstrated through rigorous online A/B tests conducted across Meituan’s entire platform, confirming the **quantifiable business value of uncertainty modeling**. The experiment selected peak periods (lunch 11:30–13:30, dinner 17:30–19:30) in Beijing, Shanghai, and Guangzhou as the control group, covering complex scenarios with over 2 million orders per day. The results showed: First, **significant improvement in ETA accuracy**. After adopting the HEDA-GMM model, the median absolute error of overall ETA decreased by 28.6% (from 152 seconds to 108 seconds), with the prediction success rate for the highest overtime-risk “30-minute delivery” orders increasing by 19.3 percentage points. Second, **qualitative change in dispatch decision quality**. After replacing fixed values with “expected completion time” generated based on probabilistic service time, rider idle rate decreased by 12.4%, and cross-regional order mismatch rate decreased by 15.7%. This means approximately 8.3 kilometers of无效行驶 can be reduced per 10,000 orders, directly corresponding to reduced carbon emissions and fuel costs. Third, **dual improvement in user experience and workforce stability**. User complaints in the “rider长时间未送达” category decreased by 33.1%, while rider mid-delivery cancellation rates due to inaccurate service time estimation decreased by 21.8%, and platform monthly rider retention increased by 0.9 percentage points. Crucially, these gains did not come at the cost of system throughput—with the same server resources, the HEDA module’s average response latency increased by only 23ms (<0.5%), proving its engineering feasibility. This set of data collectively points to one conclusion: scientific modeling of service time uncertainty has evolved from theoretical optimization to a core productivity factor affecting fulfillment costs, user satisfaction, and ecosystem health.
## Criticisms and Limitations
Despite the promising results, the research by Zheng et al. (2022) also presents some limitations:
1. **Data Dependency**: The effectiveness of the model depends on the quality and quantity of historical data. Poor data quality or limited data availability can lead to inaccurate predictions.
2. **Computational Complexity**: GMM estimation requires more computational resources compared to simple distribution assumptions. This can be a concern, especially during peak hours with a high volume of orders.
3. **Real-Time Challenges**: Balancing computational accuracy and response speed during peak hours can be challenging. The proposed model may not be suitable for real-time applications that require rapid response times.
## Practical Implications
The research by Zheng et al. (2022) has several practical implications for the OFD industry:
1. **Dispatch Optimization**: More accurate service time estimation can improve order allocation efficiency, reducing delivery times and increasing customer satisfaction.
2. **Cost Control**: Service time statistics can be used to adjust rider compensation and modulate order delivery priorities, leading to cost savings for OFD platforms.
3. **Infrastructure Planning**: The research can help determine whether and where to install self-service dining cabinets, improving overall service efficiency.
4. **Technology Transfer**: The proposed methodology can be applied to other logistics and supply chain scenarios requiring uncertain service time handling.
## Paper Citation
Zheng, J., Wang, L., Ding, X., Wang, S., Chen, J., Wang, X., Duan, H., & Liang, Y. (2022). Modeling stochastic service time for complex on-demand food delivery. Complex & Intelligent Systems, 8(2), 1-15. https://doi.org/10.1007/s40747-022-00719-4









