A new type of coronavirus has started to break out around the world since 2019 and was declared to be a pandemic by WHO in March 2020 [1]. The virus can spread to a large number of people in a short period of time, which could cause symptoms such as colds, fever, and even threaten life in severe cases. Therefore, in the past few years, many institutions have been recording daily data about Covid-19 around the world, including the number of new infections, the death rate, the number of recoveries, and so on. These data are important for demographics and can also help people understand which treatments are most effective [2]. Therefore, these recordings will bring great help to the control of the global epidemic if these data can be well analyzed. Apart from the potential epidemiology benefits, we believe Covid-19 is unique in even a pure modeling standpoint as well, because we have 1) a very large amount of data globally across different regions and ethnicity, 2) a number of different variations of the virus resulting in multiple waves of confirmed cases, 3) multiple factors intertwined together resulting an complex system that might be hard to be described by differential equations. The above reason makes analyzing the data interesting but also makes it hard to design a model that could effectively capture the available information. In this report, we mainly try to tackle the third point. In the following context, we will introduce two models which will be referred to as the SIR-CDR model and the SIR-D model. Both of which are plug-and-play POMP models and the model design are based on our current study of how Covid-19 spreads.
We use the Covid-19 data in Moscow [3]. We chose this data because it has recorded and updated the number of confirmed cases, number of deaths, and number of recovered cases daily. The original dataset has the total cases since 2020-06-01 to 2022-04-12. We converted it into daily cases to better illustrate the trend. The plot of the data is shown below.
As we expected, there are several waves of cases. Since this report is mainly focused on the complex system part rather than the time varying part as described earlier in the introduction, we only use the first wave of the data before 2021-04-01 and the result is shown below.
For both the original dataset and the smaller dataset we selected, we can observe that the number of confirmed cases is almost a lagged version of the number of deaths. And it is reasonable to assume that the number of confirmed cases should equal the sum of the number of deaths and recovered cases, up to some varying lag and observational uncertainty. This motivates us to develop the SIR-CDR model which is a multivariate POMP model based on the SIR model that measures the number of confirmed cases, the number of death cases and the number of recovered cases simultaneously. By using more information in the model, we hope that the model could be fitted much more efficiently and capture the key lag information which is directly related to how people recover from Covid-19. Importantly, though deaths and recovered cases are both considered as in-transmissive cases in the standard SIR model, it should be more beneficial to model them separately as they represent the very opposite outcome of the pandemic.
We have 8 compartments:
We assume that there’s a chance of infected cases becoming asymptomatic which is unobservable throughout the event since there wasn’t a test kit yet at the early stages of the pandemic. Both the asymptomatic case and the symptomatic case are transimissive but the hospitalized cases are not transimissive. The asymptomatic case could recover on its own resulting in the R compartment, which is also unobservable. The asymptomatic case could could start to develop symptoms resulting in the Sy compartment. The symptomatic cases could recover which is observable, and could also result in death in the D compartment, and could also result in a hospital in the H compartment. The hospitalized cases could either recover or die. Another feature of the model is that we have a parameter Cap representing the capacity of the available medical treatments. We imagine this would play a big role as there had been several Covid-19 related medical shortage across the global. When the hospital compartment H is at its capacity, additional cases coming from the \[Sy\] compartment would stay at the \(Sy\) compartment until there’s more availability.
We then have 11 parameters:
Together, we have the following initialization:
\[\begin{align} S &= N*(1-\eta) \\ A &= N(\eta \alpha) \\ Sy &= N(\eta (1-\alpha)) \\ H &= 0\\ R &= 0 \\ D &= 0\\ C &= 0\\ Rr &= 0\\. \end{align}\]
Then we have the measurement model which measures three compartment simultaneously. We assume the asymptomatic cases are observable throughout the event. The symptomatic cases are observable with some reporting rate \(\rho\). The hospitalized cases are fully observable as this is one of the major pandemic and we assume official would pay extra attention recognizing such event. Then, we have our measurement model as:
\[\begin{align} \text{deaths} \sim \text{Pois}(D)\\ \text{confirmed} \sim \text{Pois}(C)\\ \text{recovered} \sim \text{Pois}(Rr)\\. \end{align}\]
And then the process model: \[\begin{align} dN_{SA} &\sim \text{Binom}(S, 1-\exp{-\beta\alpha(A+Sy)dt/N})\\ dN_{SSy} &\sim \text{Binom}(S, 1-\exp{-\beta(1-\alpha)(A+Sy)dt/N})\\ dN_{AR} &\sim \text{Binom}(A, 1-\exp{-\mu_{R}dt})\\ dN_{ASy} &\sim \text{Binom}(A, 1-\exp{-\mu_{ASy}dt})\\ dN_{SyR} &\sim \text{Binom}(Sy, 1-\exp{-\mu_{SyR}dt})\\ dN_{SyH} &\sim \text{Binom}(Sy, 1-\exp{-\mu_{SyH}dt})\\ dN_{HR} &\sim \text{Binom}(H, 1-\exp{-\mu_{R}dt})\\ dN_{SyH} &\sim \text{Binom}(Sy, 1-\exp{-\mu_{SyD}dt})\\ dN_{HD} &\sim \text{Binom}(H, 1-\exp{-\mu_{HD}dt})\\ \end{align}\]
The process model is similarly to the standard SIR model where \(dN_{AB}\) represents the flow from a compartment \(A\) to compartment \(B\), except the thresholding capacity mechanism we mentioned above. For the three measurement compartment, symptomatic cases are partially observed with rate \(\rho\) and hospitalized cases are fully observed while the asymptomatic cases are fully hidden.
The compartments related to the measurements is then:
\[\begin{align} D_{t+1} &= D_{t} + dN_{SyD, t} + dN_{HD, t} \\ C_{t+1} &= C_t + dN_{SyH, t} + \rho dN_{SyR, t} \\ Rr_{t+1} &= Rr_t + dN_{HR, t} + \rho dN_{SyR, t} \\ \end{align}\]
We then have tried many ways to tweak the model and choosing reasonable initial values, but this model turns out to be very hard to optimize. Even the particle filtering could stress some difficulties with constantly single-digit effective sample sizes. We would imagine this is because the model itself has a very specific structure and would require a lot of fine tuning to align things together. Though we couldn’t make it work, we still think this general structure could lead to modeling advantages and is worth reporting.
We then proceed to try a much smaller model, the SIR-D model which only measures the death cases. In the spirit of the SIR-CDR model and the benefits of distinguishing death cases apart from recovered cases as we discussed above, we have two separate compartments for the death cases and the recovered cases. This is different from the standard SIR model where a single compartment is responsible for accounting the in-transmissive cases which include the death cases and the recovered cases. This is less of a deal since both cases represent the very opposite outcome of the pandemic and we would want to account for them separately.
More specifically, we have 5 compartments:
We assume that there’s a chance of infected cases becoming asymptomatic which is unobservable throughout the event, since there wasn’t a test kitat the early stages of the pandemic. Both the asymptomatic case and the symptomatic case are transmissive. The asymptomatic case could recover on its own resulting in the R compartment, which is also unobservable. The asymptomatic case could start to develop symptoms resulting in the Sy compartment, which is observable. The symptomatic cases could recover, which is observable, and could also result in death in the D compartment.
We also have 9 parameters:
Together, we have the following initialization: \[\begin{align} S &= N*(1-\eta) \\ A &= N(\eta \alpha) \\ Sy &= N(\eta (1-\alpha)) \\ R &= 0 \\ D &= 0\\ \end{align}\]
And the measurement model: \[ \text{deaths} \sim \text{NegBinom}(k, D) \]
And the process model: \[\begin{align} dN_{SA} &\sim \text{Binom}(S, 1-\exp{-\beta\alpha(A+Sy)dt/N})\\ dN_{SSy} &\sim \text{Binom}(S, 1-\exp{-\beta(1-\alpha)(A+Sy)dt/N})\\ dN_{AR} &\sim \text{Binom}(A, 1-\exp{-\mu_{AR}dt})\\ dN_{ASy} &\sim \text{Binom}(A, 1-\exp{-\mu_{ASy}dt})\\ dN_{SyR} &\sim \text{Binom}(Sy, 1-\exp{-\mu_{SyR}dt})\\ dN_{SyD} &\sim \text{Binom}(Sy, 1-\exp{-\text{D_rate}\mu_{SyR}dt})\\ \end{align}\]
And the general flow between compartment is defined similarly as in the standard SIR model, where \(dN_{AB}\) represents the flow from a compartment \(A\) to compartment \(B\). One thing different from SIR is that instead of allowing arbitrary flow from the above equation, we added the constraint that all compartments can only maintain a positive value. More specifically, if \(dN_{AB}\) defines a bigger change out of the compartment \(A\), which result in a negative value of \(A\), we set \(A \leftarrow 0\) and let \(dN_{AB} \leftarrow A\) to ensure that the maximum number that can flow out of a compartment is the size of the compartment in such case. Consequently, the downstream receptive compartment only receives the restricted flow. The intuition is as simple as even sometimes the system requires a larger change but the population in each compartment has to be bigger than 0. This is implemented as a sequence if-else
control sequence in the process model.
We then perform a local search with starting value of \[\begin{align} N &= 11920000 \\ \text{D_rate} &= 0.01 \\ \eta &=0.002 \\ \beta &= 0.13 \\ \alpha &= 0.3 \\ \mu_{AR} &= 0.05 \\ \mu_{SyR} &= 0.05 \\ \mu_{ASy} &= 0.5\\ k &= 10\\. \end{align}\]
Aside from the fixed parameters we mentioned above, we use a perturbation of 0.01 to the remaining parameters of \(\beta, \mu_{AR}, \mu_{SyR}, \mu_{ASy}\). The result for 100 iterations of IF2 algorithm with 1000 particles for 20 simulations is shown below.
As we can see, some of the simulations have difficulties optimizing the likelihood and quickly drop down to some very small values. Those simulations seems to corresponds to a \(\beta\) value close to 0, which completely zeros out the flow of the compartment since it controls the flow out the source compartment S. \(\mu_{AR}\) seems to have a value around 0.1 representing roughly 10% of the asymptomatic population recovers without developing symptoms each day.
We then proceed to perform the global search with Nreps_global
starting points random draws as follow.
\[\begin{align} \beta &\sim U(0.001, 2) \\ \mu_{AR} &\sim U(0.001, 2) \\ \mu_{SyR} &\sim U(0.001, 2) \\ \mu_{Asy} &\sim U(0.001, 2) \\ \end{align}\]
The log-likelihood evaluated with 10 repetitions are shown below. We observe the similar behavior where some of the starting points raise difficulties to the IF2 algorithm, which resulted in a diverged optimization. We also see a cluster of the parameter \(\mu_{SyR}\) have very large values ranging from 50 to 150. Together with the constant difficulties which the IF2 algorithm has. This is an indication that the model is miss-specified.
The global search result from the converged runs are shown below. Gray points are the starting points for the global search and the red points are the end result. From the limited data points, we can see that the model seems to favor a \(\beta\) value of around 1 and the small\(\mu_{AR}\) and \(\mu_{ASy}\) values. For \(\mu_{SyR}\), despite having rather reasonable starting values below 2, we see a consistent trend of moving towards large values. This further confirms that our model is miss-specified
The best parameters along with its log-likelihood is shown below. Again the non-sensible large value of \(mu_{SyR}\) is very likily an indication of model specification.
Beta | Mu_AR | Mu_SyR | Mu_ASy | N | D_rate | eta | Alpha | k | loglik | loglik.se |
---|---|---|---|---|---|---|---|---|---|---|
0.9238942 | 0.1530965 | 157.4141 | 0.1396968 | 11920000 | 0.01 | 2e-04 | 0.3 | 10 | -1171.727 | 0.7091681 |
We then simulate using the above best parameter from above and the result is shown below.
As we can see, the proposed model failed to capture the general non-linear shape of the data.
From the above global search result, we obtain the parameters range to construct the profile likelihood.
Beta | Mu_AR | Mu_SyR | Mu_ASy |
---|---|---|---|
0.7180459 | 0.0151511 | 54.27137 | 0.0303764 |
1.5078589 | 0.2403424 | 228.48570 | 0.3194316 |
For a range of \(\mu_{SyR}\) values, we sample Nreps_profile
points from the other parameters for optimization to construct the profile likelihood, and the points are plotted below.
The result is shown below. Notice the range of the log likelihood is across a magnitude of several thousands, and we didn’t achieve our best log-likelihood obtained above. This is due to the reason that we fixed our target parameter \(\mu_{SyR}\) in the range of 0 and 1 but the optimized model prefers values of above 100 which is not sensible.
Then the attempted “profile likelihood” is plotted below. As we expected, all points are below the threshold and there’s no real meaningful interpretation for this graph other than another confirmation of a miss-specified model
In this report, we proposed two models trying to capture the complex relationship within the system. Specifically, the SIR-CDR model measures three key variables simultaneously while taking into the consideration of ICU capacity, which plays an important role in remedying Covid-19. The much smaller SIR-D model measures the death cases and has separate compartments for death cases and recovered cases. Although we had difficulties fitting both models due to various reasons, we believe it is still worthy to present both models, , especially the SIR-CDR model, as they have some unique aspects compared to previous Stats 531 Covid19-related projects. Besides, we have the potential chance to capture the complex system upon further tuning.