the Creative Commons Attribution 4.0 License.
the Creative Commons Attribution 4.0 License.
Ensemble Kalman–Guided Model Predictive Path Integral Control for Spatially Localized Suppression of Extremes in Chaotic Geophysical Flows
Abstract. The possibility of influencing extreme weather phenomena has been discussed for decades; however, it remains far from operational practice, and there is still no established framework for designing small, spatially localized perturbations that can reliably steer chaotic geophysical flows. In this study, we propose a hybrid control method, termed ensemble-Kalmanguided model predictive path integral control (EKG-MPPI), which combines ensemble Kalman control (EnKC) with model predictive path integral (MPPI) control. Within a control simulation experiment framework, an ensemble Kalman filter is first used for state estimation, after which EnKC computes a candidate perturbation by treating the control objective as a pseudo-observation. An adaptive thresholding procedure then enforces spatial sparsity, so that the EnKC perturbation identifies candidate actuator locations and their nominal amplitudes. This information is embedded into the mean and covariance of Gaussian proposal distributions for MPPI, which subsequently refines the perturbation through sampling-based optimization with nonlinear rollouts, without linearizing the dynamics or computing gradients. Numerical experiments with the Lorenz–96 model and the surface quasi-geostrophic (SQG) model demonstrate that EKG-MPPI can suppress extremes in state variables and regional wind speed more effectively than EnKC alone, while using comparable or smaller control inputs. These results highlight EKG-MPPI as a promising building block for simulation-based assessment of localized intervention strategies in geophysical flows.
- Preprint
(1095 KB) - Metadata XML
- BibTeX
- EndNote
Status: open (until 31 Mar 2026)
-
RC1: 'Comment on egusphere-2026-419', Victor Bertret, 23 Feb 2026
reply
-
CC1: 'Reply on RC1', Kazumune Hashimoto, 21 Mar 2026
reply
Author Response to Referee #1
We thank the Referee for the thorough review and constructive comments. To address these comments, we have performed additional experiments and analyses, and in the revised manuscript we will make the following main changes:
1. Clarify the rationale for using MPPI under single-step actuation, and more explicitly position the present work as a fundamental method toward multi-step and multi-actuator extensions.
2. Improve the fairness of the baseline comparison by expanding the sensitivity study of BO-based baselines, including a BO-informed MPPI variant.
3. Add computational profiling that separates the EnKC-prior stage from the MPPI rollout stage, and discuss the effect of parallelization.
4. Improve the clarity of the notation by explaining the embedding functions f_loc, g_loc, f_mag, and g_mag when they are first introduced.Below, we respond point by point.
---
Referee Comment 1: Single-step actuation vs. sequential trajectory optimization
Referee comment:
The authors explicitly state that the control perturbation is applied only at the first rollout step, with no further input during the remaining horizon. Because the problem is reduced to optimizing a single initial impulse, the use of MPPI seems oversized. Discuss why MPPI was chosen over CMA-ES or Bayesian Optimization. If the intention is groundwork for multi-step actuation, state this more explicitly.Response:
We thank the Referee for this important point. We agree that the current implementation applies the perturbation only at the first rollout step, with zero input thereafter. Thus, at each control decision, the optimization reduces to choosing a single localized impulse. Indeed, this design was motivated by the current CSE setting: in chaotic flows, even a small localized perturbation can produce a meaningful downstream effect over the prediction horizon, while restricting the actuation to a single step reduces the effective control dimension and improves sampling efficiency.Nevertheless, we agree that a clearer justification for using the MPPI formalism is necessary. In the revised manuscript, we will emphasize that MPPI still offers several advantages in this setting:
1. Parallel sampling: MPPI evaluates a batch of rollouts whose costs and weights can be computed in parallel. This is advantageous when forward simulations are the dominant cost and can be parallelized on multi-core CPUs or GPUs.
2. Avoiding sequential acquisition loops: Bayesian optimization typically involves a sequential propose-evaluate loop, which can be less straightforward for real-time implementation.
3. Extensions to multi-step control: As the Referee points out, one motivation for adopting MPPI is that it can be generalized to multi-step control sequences and multiple actuators within the same algorithmic pipeline. The single-step setting in this manuscript is a proof of concept toward that goal.Following the Referee's suggestion, we have additionally performed a BO-based comparison for the present single-step optimization problem. In this comparison, we use a standalone BO baseline. In our BO implementation, the Expected Improvement (EI) acquisition function is used, and 10 BO iterations are carried out at each control decision. Although 10 iterations is a modest BO budget, we adopted this setting because the resulting per-decision computation time is already substantially larger than that of the proposed method.
The results show that EKG-MPPI outperforms BO in the key metrics:
- Number of exceedances (X > 12): 5547 (EKG-MPPI) vs. 6386 (BO)
- Mean input magnitude: 0.564264 (EKG-MPPI) vs. 0.611567 (BO)
- Mean computation time: 0.000253 s (EKG-MPPI) vs. 0.122630 s (BO)These results indicate that even in the single-step setting, MPPI provides a more efficient and effective solution than BO.
Furthermore, we clarify in the revised manuscript that one of the main motivations for adopting MPPI is its natural extensibility to multi-step control and multiple actuators within a unified framework. The current single-step formulation should be regarded as a proof of concept toward this broader objective.
---
Referee Comment 2: Fairness in baseline comparison (hyperparameter optimization)
Referee comment:
The comparison against vanilla MPPI lacks rigorous hyperparameter tuning. Vanilla MPPI uses fixed lambda = 0.7 while EKG-MPPI uses lambda = 0.1. Vanilla MPPI uses a static sampling distribution. EKG-MPPI adapts mean/variance via EnKC and thus has a massive advantage. Baseline MPPI should undergo hyperparameter tuning or sensitivity analysis.Response:
We thank the Referee for this important comment. We agree that the comparison with vanilla MPPI should be strengthened by examining hyperparameter sensitivity and by considering a stronger adaptive baseline. In addition to the original vanilla MPPI comparison, we also introduce a stronger MPPI baseline to examine whether the advantage of EKG-MPPI persists under a more adaptive baseline design.Specifically, we construct an additional BO-informed MPPI baseline, denoted BO-MPPI, in which Bayesian optimization is used to estimate prior information for the intervention location and magnitude, and these estimates are then supplied to vanilla MPPI.
For BO-MPPI, the BO objective balances the predicted threshold exceedance over the rollout horizon and the intervention magnitude, where the parameter alpha controls the trade-off between control effectiveness and actuation cost. Here, alpha is the BO-side trade-off parameter in the objective function, whereas lambda is the MPPI temperature parameter. These two parameters play different roles and are varied separately in our experiments.
To ensure a fair comparison with EKG-MPPI, BO-MPPI uses the same MPPI temperature lambda and the same sampling variances for intervention location and magnitude as EKG-MPPI. For each fixed value of lambda, we run BO-MPPI with alpha = 100, 10, 1, 0.1, 0.01. The BO budget is fixed at 10 iterations at each control decision, since even with this modest budget the BO-based baseline already requires a substantially longer computation time than the proposed method. We then evaluate the resulting Lorenz-96 simulations in terms of the number of threshold exceedances and the average intervention magnitude.
Separately, we repeat this comparison for lambda = 0.1, 1, 10, and provide the corresponding plots in the supplementary material. Across all tested values of lambda, the EKG-MPPI result lies below the empirical lower envelope (Pareto frontier-like trade-off curve) of the BO-MPPI results obtained by varying $\alpha$. These additional results show that the advantage of EKG-MPPI is robust to the choice of lambda and is not attributable to a single hyperparameter setting.
---Referee Comment 3: Computational scalability and profiling
Referee comment:
EKG-MPPI is the most computationally expensive. Provide a breakdown of time spent computing the EnKC prior vs. evaluating MPPI rollouts to clarify bottlenecks and scaling feasibility.Response:
We thank the Referee for this useful suggestion and agree that such a runtime breakdown is important for identifying the main computational bottlenecks and assessing scalability. We therefore measured the computation time separately for the EnKC-prior stage and the MPPI rollout/weight stage, and we have added this information to the revised manuscript.
In the Lorenz-96 experiment with K_MPPI = 40, the mean runtime per control decision was 0.000044 s for the EnKC-prior computation and 0.000222 s for the MPPI rollout/weight computation.
The total mean runtime was therefore 0.000266 s per decision step, with approximately 17% of the cost attributable to the EnKC-prior stage and 83% to the rollout and weight stage.
This indicates that the main computational bottleneck is the rollout evaluation rather than the EnKC prior itself. In other words, the additional overhead introduced by the EnKC-based prior is modest in the present setting. Moreover, the total runtime remains well below the sampling interval used in the experiment, supporting the computational feasibility of the proposed method in the current online-control setting.
---Referee Comment 4: Technical correction on embedding-function notation
Referee comment:
In Section 3.2, Step 3, the embedding functions f_loc, g_loc, f_mag, and g_mag are introduced. Add a brief sentence clarifying their exact mathematical definitions or references earlier.Response:
We thank the Referee for this helpful comment. In the revised manuscript, we will clarify that:
- f_loc and f_mag map the EnKC output to the mean of the sampling distribution (location and magnitude),
- g_loc and g_mag define the corresponding sampling variance or covariance used by MPPI.---
Summary
We again thank the Referee for the constructive feedback. We believe these revisions significantly strengthen the manuscript by clarifying the methodological rationale, improving the fairness of the comparisons, and providing a more transparent analysis of computational costs.
Sincerely,
Haru Kuroki, Kazumune Hashimoto, Yuki Uehara, Yohei Sawada, Duc Le, and Masashi Minamide
-
CC1: 'Reply on RC1', Kazumune Hashimoto, 21 Mar 2026
reply
-
RC2: 'Comment on egusphere-2026-419', Anonymous Referee #2, 11 Mar 2026
reply
This manuscript introduces EKG-MPPI, a hybrid control approach for localized suppression of extremes in chaotic geophysical flows. The method combines EnKF for state estimation, EnKC for constructing candidate perturbations, and MPPI to improve on them through nonlinear refinement. The approach is tested on the Lorenz–96 and SQG models, where the authors report reductions in extreme-event, together with comparable or smaller control inputs relative to EnKC. Overall, the paper is well motivated, the methodological idea is interesting, and the combination of EnKC and MPPI is intuitive and potentially useful for this class of problems.
The paper shows clearly how the two methods are coupled in practice. The numerical experiments on Lorenz–96 and SQG support the claim that the method works and can improve on EnKC in the examples considered. However, the empirical advantage remains somewhat difficult to assess quantitatively, because the reported performance results in both numerical sections are small and presented without uncertainty estimates. This does not reduce the interest of the proposed method, but it does make the strength of the performance claim harder to evaluate in the present version.
Below are suggestions of minor corrections that may improve the paper:-In both the Lorenz-96 and SQG sections, the conclusions are reported without uncertainty estimates. For Lorenz-96, Table 1 reports extreme event suppressions; for SQG, the conclusions are drawn from raw peak wind and mean input comparisons. But in both cases no confidence intervals or variability estimates are provided for these performance diagnostics. Since for both metrics the improvement is moderate (less than 10%), it is then difficult to assess how significant the improvement of EKG-MPPI over EnKC is.
-In the Lorenz-96 experiments, EKG-MPPI uses lambda = 0.1 and an EnKC-informed proposal distribution, while vanilla MPPI uses lambda = 0.7 and uninformed Gaussian sampling for magnitude and location. As a result, the reported difference does not isolate the effect of the proposed guidance alone. I suggest clarifying that vanilla MPPI is included as a simple reference only, and that the principal comparison is with EnKC.
-Figure 4's caption states that the evaluation metric is the percentage difference ((EKG - EnKC)/EnKC) x 100, but the figure itself shows raw maximum wind speed and raw mean input magnitude by scenario. Either this text should be removed and replaced with useful information about the figure, or the results about the percentage difference should be added.
-The figures could be made more informative. In figure 2, 3 and 5, the axes lack clear units or precise information on the quantities shown, and some legends remain too limited to guide the reader which needs to infer from the text outside of the figure. E.g. "X variables" is not defined, and most colour bars come without text. I would encourage the authors to improve the figure labels, legends, and captions so that each figure can be understood more directly on its own.
Also, here are some very small corrections that would make the manuscript easier to read and more complete:- In Algorithm 1, the Require line lists H but not H_c, although line 5 calls EnKC with H_c.
- In Eq. (1), the model equation uses q_{t-1}, while the text below defines q_t as the model error.
- In Eq. (10), the text just above introduces an ensemble {x_{t}^{alpha(i)}}{i=1}^m, but Eq. (10) then uses x{t+T_EnKC}^a as if it were a single, already defined state. It is unclear from a quick read whether this denotes one ensemble member, the ensemble mean, or another quantity.
- The switch from the generic EnKC horizon T_EnKC to T_c in the Lorenz-96 section could be made more explicit.
- The Gaussian sampling notation should be clarified. In Sect. 3, the second Gaussian parameter is introduced as a variance/covariance term, but in the numerical sections the paper sets g_mag(u_t^EnKC) = ||u_t^EnKC||_1 / 2 in Lorenz-96 and g_mag(u_t^EnKC) = ||u_t^EnKC||_1 in SQG, which seems like a natural choice for a standard deviation. It would help to clarify whether the authors mean m ~ N(mu, g_mag) or m ~ N(mu, g_mag^2).
- In the paragraph introducing Figure 5, the order “no-control, EKG-MPPI, and EnKC” should be made consistent with figure 5, where the order is “no-control, EnKC, and EKG-MPPI.”
- The introduction refers to Henderson as an early 4D-Var-based study, but the corresponding reference is missing from the bibliography. The citation should be added explicitly. From my quick research as: Henderson, J. M., Hoffman, R. N., Leidner, S. M., Nehrkorn, T., and Grassotti, C. (2005), “A 4D-Var study on the potential of weather control and exigent weather forecasting,” Quarterly Journal of the Royal Meteorological Society, 131(612), 3037-3051, doi:10.1256/qj.05.72.
- For the first CSE citation, line 34 , I think the original CSE paper from 2022 is better suited: Miyoshi, T. and Sun, Q. (2022), “Control simulation experiment with Lorenz’s butterfly attractor,” Nonlinear Processes in Geophysics, 29, 133-145.
- The manuscript currently cites Sawada (2024a) as a preprint, the authors may also update that entry to the published journal version: Sawada, Y. (2024), “Ensemble Kalman Filter Meets Model Predictive Control in Chaotic Systems,” SOLA, 20, 400-407, doi:10.2151/sola.2024-053
Overall, I believe this manuscript is suitable for publication after minor corrections. I thank the authors for an interesting and enjoyable contribution.Citation: https://doi.org/10.5194/egusphere-2026-419-RC2
Viewed
| HTML | XML | Total | BibTeX | EndNote | |
|---|---|---|---|---|---|
| 145 | 79 | 19 | 243 | 11 | 22 |
- HTML: 145
- PDF: 79
- XML: 19
- Total: 243
- BibTeX: 11
- EndNote: 22
Viewed (geographical distribution)
| Country | # | Views | % |
|---|
| Total: | 0 |
| HTML: | 0 |
| PDF: | 0 |
| XML: | 0 |
- 1
This manuscript presents a novel and elegant hybrid control strategy, termed EKG-MPPI, aimed at spatially localized interventions in chaotic geophysical flows. By combining the statistical data assimilation capabilities of Ensemble Kalman Control (EnKC) with the nonlinear, derivative-free optimization of Model Predictive Path Integral (MPPI) control, the authors provide a compelling framework for "Control Simulation Experiments". From a systems and control engineering perspective, using the EnKC to generate a sparse, physics-informed prior to guide the sampling distribution of the MPPI is a brilliant approach to solving the sample inefficiency inherent to high-dimensional state spaces. The method is entirely matrix-free and avoids the computation of adjoint models or gradients, which is highly desirable for chaotic, non-linear atmospheric models. However, while the theoretical hybridization is sound, the current implementation simplifies the temporal complexity of the control problem to a single-step actuation. This reduces the framework to a static spatial optimization problem rather than true multi-step trajectory optimization. Furthermore, the comparison with baseline methods lacks rigorous hyperparameter tuning, which somewhat weakens the performance claims. The manuscript is a strong proof-of-concept, but addressing the specific points below would significantly improve its rigor and impact.
Specific Comments :
- Single-Step Actuation vs. Sequential Trajectory Optimization: The authors explicitly state that the control perturbation is applied only at the first rollout step, with no further input provided during the remaining prediction horizon ($T_{MPPI}$). Because the problem is reduced to optimizing a single initial impulse, the use of MPP, a method designed for sequential, multi-step trajectory optimization, seems algorithmically oversized. The authors should discuss why MPPI was chosen over standard derivative-free black-box optimization algorithms (e.g., CMA-ES or Bayesian Optimization) which are highly efficient for static, single-step parameter search. If the intention is to lay the groundwork for future multi-step actuation, this should be stated more explicitly as the primary justification for using the MPPI formalism.
- Fairness in Baseline Comparison (Hyperparameter Optimization): In Section 4.1.1, the proposed EKG-MPPI is compared against a "vanilla MPPI" baseline. However, the setup for the vanilla MPPI utilizes arbitrarily fixed hyperparameters without any apparent tuning. Specifically, the authors set the temperature parameter to $\lambda = 0.7$ for the vanilla MPPI, whereas EKG-MPPI operates with $\lambda = 0.1$. Furthermore, the vanilla MPPI relies on a completely static sampling distribution (variance of 0.5 for magnitude and 20.0 for location). Because EKG-MPPI inherently adapts both the mean and variance of its sampling distribution using the EnKC output, it naturally possesses a massive advantage. To ensure a rigorous comparative analysis, the baseline MPPI should ideally undergo a degree of hyperparameter tuning or at least a sensitivity analysis similar to the one performed for EKG-MPPI in Figure 3 to firmly demonstrate that its poorer performance is due to the lack of EnKC guidance rather than sub-optimal static parameters.
- Computational Scalability and Profiling: The results in Table 2 show that EKG-MPPI is the most computationally expensive method tested. While the authors correctly point out that MPPI rollouts are "embarrassingly parallel" and benefit directly from GPU acceleration, a brief computational profiling of the EKG-MPPI execution time would be highly valuable. Specifically, providing a breakdown of the time spent computing the EnKC prior versus the time spent evaluating the MPPI rollouts would clarify where the true bottleneck lies. Given the authors' stated intention in the conclusion to scale this approach to multi-actuator and multi-step scenarios, this profiling would help readers assess the operational feasibility of scaling the framework.
Technical Corrections :
- Equation notations: In Section 3.2, Step 3, the embedding functions $f_{loc}$, $g_{loc}$, $f_{mag}$, and $g_{mag}$ are introduced. It would be helpful to the reader to add a brief sentence clarifying their exact mathematical definitions or references earlier in the text, before they are fully explicitly defined for the experiments in Equations (39) and (50).