
In this article, we have presented a mathematical model to study the dynamics of hepatitis C virus (HCV) disease considering three populations namely the uninfected liver cells, infected liver cells, and HCV with the aim to control the disease. The model possesses two equilibria namely the disease-free steady state and the endemically infected state. There exists a threshold condition (basic reproduction number) that determines the stability of the disease-free equilibrium and the number of the endemic states. We have further introduced impulsive periodic therapy using DAA into the system and studied the efficacy of the DAA therapy for hepatitis C infected patients in terms of a threshold condition. Finally, impulse periodic dosing with varied rate and time interval is adopted for cost effective disease control for finding the proper dose and dosing interval for the control of HCV disease.
Citation: Amar Nath Chatterjee, Fahad Al Basir, Yasuhiro Takeuchi. Effect of DAA therapy in hepatitis C treatment — an impulsive control approach[J]. Mathematical Biosciences and Engineering, 2021, 18(2): 1450-1464. doi: 10.3934/mbe.2021075
[1] | LacreishaEjike-King, RashidaDorsey . Reducing Ex-offender Health Disparities through the Affordable Care Act: Fostering Improved Health Care Access and Linkages to Integrated Care. AIMS Public Health, 2014, 1(2): 76-83. doi: 10.3934/publichealth.2014.2.76 |
[2] | Andrea M. Auxier, Bonni D. Hopkins, Anne E. Reins . Under Construction: One States Approach to Creating Health Homes for Individuals with Serious Mental Illness. AIMS Public Health, 2015, 2(2): 163-182. doi: 10.3934/publichealth.2015.2.163 |
[3] | Bethany M. Kwan, Aimee B. Valeras, Shandra Brown Levey, Donald E. Nease, Mary E. Talen . An Evidence Roadmap for Implementation of Integrated Behavioral Health under the Affordable Care Act. AIMS Public Health, 2015, 2(4): 691-717. doi: 10.3934/publichealth.2015.4.691 |
[4] | Larrell L. Wilkinson, Saundra H. Glover, Janice C. Probst, Bo Cai, Lisa T. Wigfall . Psychological Distress and Health Insurance Coverage among Formerly Incarcerated Young Adults in the United States. AIMS Public Health, 2015, 2(3): 227-246. doi: 10.3934/publichealth.2015.3.227 |
[5] | Aswin Sugunan, Rajasekharan Pillai K, Anice George . What is the coverage of your health insurance plan? An audit of hospital billing. AIMS Public Health, 2024, 11(4): 1013-1029. doi: 10.3934/publichealth.2024052 |
[6] | Ryan M. McKenna, Jonathan Purtle, Katherine L. Nelson, Dylan H. Roby, Marsha Regenstein, Alexander N. Ortega . Examining EMTALA in the era of the patient protection and Affordable Care Act. AIMS Public Health, 2018, 5(4): 366-377. doi: 10.3934/publichealth.2018.4.366 |
[7] | Yosef Mohamed-Azzam Zakout, Fayez Saud Alreshidi, Ruba Mustafa Elsaid, Hussain Gadelkarim Ahmed . The magnitude of COVID-19 related stress, anxiety and depression associated with intense mass media coverage in Saudi Arabia. AIMS Public Health, 2020, 7(3): 664-678. doi: 10.3934/publichealth.2020052 |
[8] | Yu Chen Tseng, I Jen Wang, Christy Pu . Parents’ perception and willingness to maintain provider care continuity for their children under universal health coverage. AIMS Public Health, 2019, 6(2): 121-134. doi: 10.3934/publichealth.2019.2.121 |
[9] | James A. Swartz, Dana Franceschini, Kamryn Scamperle . Mental health and substance use disorder comorbidities among Medicaid beneficiaries: Associations with opioid use disorder and prescription opioid misuse. AIMS Public Health, 2023, 10(3): 658-677. doi: 10.3934/publichealth.2023046 |
[10] | Kristin Haltinner, Dilshani Sarathchandra . Tea Party Health Narratives and Belief Polarization: the Journey to Killing Grandma. AIMS Public Health, 2017, 4(6): 557-578. doi: 10.3934/publichealth.2017.6.557 |
In this article, we have presented a mathematical model to study the dynamics of hepatitis C virus (HCV) disease considering three populations namely the uninfected liver cells, infected liver cells, and HCV with the aim to control the disease. The model possesses two equilibria namely the disease-free steady state and the endemically infected state. There exists a threshold condition (basic reproduction number) that determines the stability of the disease-free equilibrium and the number of the endemic states. We have further introduced impulsive periodic therapy using DAA into the system and studied the efficacy of the DAA therapy for hepatitis C infected patients in terms of a threshold condition. Finally, impulse periodic dosing with varied rate and time interval is adopted for cost effective disease control for finding the proper dose and dosing interval for the control of HCV disease.
In natural language processing, automatic text summarization is widely used, e.g., to help users navigate web content and assist in building question-and-answer systems. Text summarization methods can be divided into two categories: extractive summarization and abstractive summarization. Extractive summarization models form summaries by selecting and copying text snippets from the document, while abstractive methods aim to generate concise summaries with paraphrasing. The commonly used abstractive methods include graph-based methods (e.g., graph-based lexical centrality as salience in text summarization (LexRank)) [1] and centrality-based methods (e.g., the centroid method) [2]. This work is primarily concerned with extractive summarization. Though abstractive summarization models have made strides in recent years, extractive techniques are still desirable, as they are simpler and more reliably yield semantically and grammatically correct sentences. Some authors have also made this point. For example, Li et al. used an abstractive method [3] and created the distributional semantics reward (DSR) model. The final experimental results showed that the summarization generated by DSR-facing long documents could not cover all of the critical information in the original text. See et al. used a hybrid pointer generator network and coverage for abstractive summaries [4]. Therefore, here, we used the extractive method to extract summarization.
The earliest extractive summarization method was introduced by Luhn in 1958. Luhn et al. used statistical information on words to calculate the importance of sentences [5] and selected high-scoring sentences to form the summarization. In 2000, Radev et al. built the multi-document summarization processor MEAD [6]. The model can reconstruct documents with summarization sentences and minimize the reconstruction error. Around 2004, extractive methods began to shift from "statistical methods" to "graph methods". Mihalcea et al. introduced the concept of bringing order into text with TextRank [7]. The core idea of TextRank is to determine the coverage of text units on the source document to get the score of sentence units. However, the algorithm does not consider the similarity between the sentences, resulting in information redundancy in the selected sentences.
In 2016, Ma et al. introduced an unsupervised multi-document summarization framework based on neural documents model (DocRebuild) to be able to refactor documents using summarization sentences [8] and selected summarization sentences to minimize refactoring errors. Cheng et al. proposed NN-SE and NN-WE [9], which are data-driven approaches based on neural networks and continuous sentence features. Nallapati et al. introduced a recurrent neural network-based sequence model for the extractive summarization of documents (SummaRuNNer) [10], and the model achieved state-of-the-art status within a year. Jadhav et al. built the Sentences and Words from Alternating Pointer Networks (SWAP-NET) model [11], which can establish the relationship between keywords and salient sentences. The model not only identified salient sentences and keywords, but it also combined them to form summaries. Zhou et al. have achieved neural extractive document summarization (NEUSUM) [12]. NEUSUM's integration of "selection" and "scoring" together solved the "sentence scoring" and "sentence selection" fragmentation in the extractive method. It achieved state-of-the-art results for the CNN/Daily Mail in that year. In 2020, Wang et al. created a model to dynamically update the weights of nodes by using a graph attention network iteratively [13] to extract the three highest-scoring sentences to form the summarization. Zhong et al. described the extractive summarization task as a semantic text matching task [14].
Although many scholars have recently achieved remarkable results in extracting summarization by using deep learning, it still has some limitations, particularly, the existence of exposure bias in neural network models [15]. Ranzato et al. introduced the exposure bias of neural networks [16] and pointed out that the cross-entropy is challenging to solve. He used reinforcement learning (RL) to improve exposure bias. Bahdanau et al. posted the exposure bias of neural networks [17] and alleviated this problem by using the actor-critic (AC) algorithm. Paulus et al. introduced the idea that models trained by supervised learning tend to exhibit exposure bias [18]. He used RL to circumvent this problem by focusing separately on the input and the continuously generated output. Narayan et al. pointed out that neural network training is based on cross-entropy [19], and that cross-entropy tends to produce redundant information when performing extractive summaries. Meanwhile, they suggested using RL to sequence sentences to overcome this difficulty. Mao et al. introduced the concept of multi-document summarization with maximal marginal relevance-guided RL [20] to address the redundancy of multi-document summarization in a unified manner.
To contribute to this field, we propose the graph matrix and advantage AC (GA2C) model to address the exposure bias of deep learning methods. GA2C translates "how to do extractive summarization" into "how to make action decisions" by uniting the graphs matrix with RL. Our contributions in this work could be summarized as follows: i) this is a novel application of advantage AC (A2C) on sentence selection for extractive summarization; ii) the analysis and empirical results show that GA2C alleviates the exposure bias problem; iii) the results disclosed that GA2C is better than previous models on CNN/Daily Mail data.
GA2C's graph matrix generator uses the idea of TextRank. And, the similarity between nodes is calculated using Jaccard similarity and cosine (COS) similarity. In addition, the TextCNN model is used in the implementation of the A2C algorithm.
TextRank, proposed by Mihalcea et al. in 2004, is a graph-based keyword extractive method. In addition, this method refers to the idea of PageRank [21]. Indeed, TextRank is a widely used keyword extraction method [7]. The method splits the text into many keywords and uses the keywords to form nodes and the similarities between the nodes to form edges, thus forming a graph. The score is higher when a keyword is more similar to other keywords.
In general, the overall algorithmic process is divided into the following four steps.
i) The text is processed and added to the graph to form nodes;
ii) The similarity between nodes is calculated and added to the graph to form edges;
iii) The algorithm is iterated until convergence;
iv) The nodes are ranked based on their scores at final convergence.
The weight iteration algorithm for the node Vi is shown in Eq (1).
WS(Vi)=(1−d)+d∗ ∑Vj∈∈(Vi)Wji ∑Vk∈Out(Vj) WjkWS(Vj) | (1) |
where WS(Vi) represents the weight of sentence i. Wji denotes the weight of the edge between the nodes Vj and Vi and d is the damping factor.
Jaccard similarity is used to establish the relationship between sentences to form the similarity matrix. The algorithm is mainly used to calculate the similarity between samples of a symbolic measure or a Boolean measure [22]. The Jaccard similarity is concerned with the characteristics that are common between samples. When the characteristic attributes between samples are identified by symbols and Boolean values, the magnitude of the specific weight of the difference cannot be measured, as one can only obtain a result such as "whether it is the same". In two samples, the Jaccard is equal to the concurrent set minus the intersections sets, and then divided by the concurrent sets as shown in Eq (2).
J(S1,S2)=|S1∪S2|−|S1∩S2||S1∪S2| | (2) |
where S1 and S2 denote the vector representation of the two sentences.
Cosine similarity is used to establish the relationship between sentences to form the similarity matrix. The algorithm measures the similarity between two vectors by calculating the cosine of the angle between them [23]. The value of cosine similarity is between 0 and 1. When two vectors have the same pointing, the cosine similarity is 1, which means that the similarity is 100%. When the angle of two vectors is 90°, the cosine similarity is 0, which implies that the similarity is 0. The cosine between two vectors can be found using the Euclidean dot product formula, as shown in Eq (3).
cos(S1,S2)= ∑ni=1(S1i∗S2i)√ ∑ni=1 (S1i)2√ ∑ni=1 (S2i)2 | (3) |
where S1 and S2 are the vector representations of the two sentences, and n is the dimension of the vector.
The evaluation network and decision-making network of GA2C were implemented in this study by using TextCNN. Both of them share the parameters of TextCNN. The summarization is selected by TextCNN. And, the selected summarization is evaluated by TextCNN. In 2014, Kim applied a convolutional neural network (CNN) model to the text classification tasks [24] and proposed TextCNN. The method is a variant of a CNN [25] and it aims to extract different local features by defining different filter convolutional kernel sizes. The model is mainly used in the field of text classification. Extractive summarization can be viewed as a text classification task divided into two categories.
As shown in Figure 1, the input was a 6 × 5 sentence matrix with a dimension of 5 for each word. The sentence matrix is passed through six convolutional kernels of size (3, 4, 5). The (3, 4, 5) convolutional kernels correspond to the extracted features size of (5, 4, 3). The extracted features are pooled and combined to get the representation after convolution.
The convolutional layer is the core of TextCNN. The three filters in this work have convolutional kernel sizes of 3, 4 and 5, which leads to the extraction of different features on the three levels. Its calculation formula is shown in Eq (4).
hi=f( ∑3x=1 ∑3y=1wi(x,y)×C(x,y)+bi) | (4) |
f represents the activation function, commonly used as the rectified linear unit (ReLU). wi(x,y) denotes the weight of the i-th node in the output matrix corresponding to the filter input node (x, y). C(x,y) is the value of the node (x, y) in the filter. bi denotes the bias term corresponding to the i-th node. Local feature extraction is achieved by setting three filters with convolutional kernels of (3, 4, 5). hi is the result of the convolution layer.
GA2C, which combines the A2C algorithm with graph structure composition, can reduce the impact of exposure bias. The methods used are mainly composed of A2C, TextRank and TextCNN models. The GA2C model mainly consists of a graph matrix generator, an evaluation network and a decision-making algorithm. The text is input to the graph matrix generator, after which the evaluation network and decision network are used for training. The framework structure diagram is shown in Figure 2.
The articles are fed into the graph matrix generator to generate the graph matrix, and the evaluation network and the decision-making network use the graph matrix to get Rewards, the next state and the Action. The evaluation network evaluates the decision-making network based on the Action. The decision-making network then improves the decision process based on the evaluation of the evaluation network.
TextRank provides a reference for the graph matrix generator. The process of using a graph matrix generator to generate a graph matrix is shown in Figure 3. The articles are input to the graph matrix generator, and after segmenting the articles into word vectors, WORDAVG is used to represent the sentence vectors and calculate the similarity between the sentence vectors to get the similarity matrix of the articles. The last graph is constructed using the similarity matrix and sentences node.
In this model, the A2C algorithm is used. The evaluation network corresponds to the critic of A2C. The network mainly calculates the actual values of the nodes. The optional actions are fed to the network and then provided to the hidden layer for full connectivity to obtain the value of the current state. The schematic diagram of the evaluation network is shown in Figure 4.
The figure starts from the node s9. Using the graph matrix, the action nodes connected to s9 can be derived as s1, s4 and s5. These three nodes are sent to the TextCNN and fully connected through the hidden layer, and the actual values are obtained. The actual values are calculated using the predicted values to obtain the loss of the evaluated network.
At each action step, the current state of the decision-making network is the set D of moves connected by the selected actions. The decision-making network determines the next sentence di from Set D. The evaluation network compares the selected di with the reference summarization bi to calculate the reward. In this paper, two reward calculation methods are provided.
1) ROUGE-1, ROUGE-2 and ROUGE-L as a reward:
Rewardi= ∑S∈{label} ∑gram−1∈S Countmatch(gram−1) ∑S∈{label} ∑gram−1∈S Count(gram−1) |
+ ∑S∈{label} ∑gram−2∈S Countmatch(gram−2) ∑S∈{label} ∑gram−2∈S Count(gram−2)+(1+b2)RLCS PLCSRLCS+b2RLCS | (5) |
which is equal to
Rewardi=R1+R2+Rl | (6) |
The label is the standard summarization of this method. Countmatch(gram−1) is the number of gram-1 matches. Count(gram−1) is the total number of gram-1. ROUGE-2 is the same. PLCS is the accuracy rate of longest common subsequence (LCS). RLCS is the recall rate of LCS, and b = PLCS/RLCS.
2) Another method is to use simple accumulation for reward calculation. A positive reward is given for a correct decision, and there is no reward for an error.
Rewardi= ∑k0judge(di,bi) | (7) |
where JUDGE is a reward calculation function that compares di with bi. If it is a summarization, JUDGE returns a positive reward. If it is not a summarization, there is no reward. Evaluate the network to get the total reward by discounting and adding up.
Rewardall= ∑ki=0γRewardi | (8) |
where k is the number of times the selection action is executed. γ is the discount factor, and it is a constant.
In this method, we use the A2C algorithm. The advantage function uses TD-error.
TD−errort=(rt+Vs(t+1)−Vst) | (9) |
This leads to the loss function of the evaluation network.
lossvalue=−E ∑kt=0TD−errort | (10) |
The decision-making network performs the selection of sentences. The sentences are fed to the decision network, and the probability of each current action is output after the hidden layer and SOFTMAX, and the action with the highest probability is selected as a result this time. The most basic TextCNN was used for action selection in this study to prove the framework's effectiveness.
The figure starts from the node s9. Using the graph matrix, the action nodes connected to s9 can be derived as s1, s4 and s5. These three nodes are fed to the TextCNN and passed through the hidden layer. The probabilities of the three actions are obtained, and the results are selected after SOFTMAX.
The decision-making network is the policy function πθ (s, a). TD-error records the error between the actual and predicted values of the current action. The decision-making network can use TD-error to calculate the update gradient. The formula is shown in Eq (11).
∇θG(θ)=Eπθ[TD−error∇θlogπθ(at|st)] | (11) |
where the update of θ is θ←θ+α∇θG(θ).
The loss function of the decision-making network is lossactor.
lossactor=−E ∑kt=0logπθ(at∨st) | (12) |
CNN and Daily Mail were used to evaluate GA2C [26]. This dataset contains news documents and their corresponding summarizations from CNN and Daily Mail websites. It contains 287, 227 documents for training, 13, 368 documents for validation and 11, 490 documents for testing. We followed [13] to preprocess the dataset.
There are more evaluation methods for text summarization tasks. In this study, we used recall-oriented understudy for gisting evaluation (ROUGE), which Lin and Och constructed in 2004 [26]. It is a frequently used evaluation metric in various text summarization models [27,28,29]. ROUGE is used to calculate the F1 scores of ROUGE-1, ROUGE-2 and ROUGE-L between the sentences selected by GA2C and the standard sentences for evaluation.
In the formula for ROUGE-N, N-gram refers to a contiguous substring of length N in the text. The denominator is the number of N-grams of all reference summaries. The numerator is the number of N-grams of both the model-predicted and the reference summarizations. In this study, N was taken as 1 and 2, which means that the evaluation methods are ROUGE-1 and ROUGE-2.
ROUGE−N= ∑S∈{RS} ∑gram n ∈ S Countmatch (gramn) ∑S∈{RS} ∑gram n ∈ S Count (gramn) | (13) |
where RS denotes the reference summarization.
The formula for ROUGE-L is shown in Eqs (14)-(16).
Rlcs=LCS(X,Y)m | (14) |
Plcs=LCS(X,Y)n | (15) |
The LCS function is the length of the longest common substring of X and Y. m and n means the size of the reference summarization and the GA2C summarization, respectively. Rlcs and Plcs respectively represent the recall and accuracy.
Flcs=(1+β2)RlCS PlCSRlCS+β2PlCS | (16) |
Flcs indicates the F1 score for ROUGE-L. When β = 1, precision and recall are weighted equally. In some cases, if precision is more critical, β should be adjusted to less than 1. If recall is more important, β should be adjusted to be greater than 1. In this study, β was taken as 1.
For these experiments, the parameter settings were mainly divided into a graph matrix generator, an evaluation network and a decision-making network.
1) Graph matrix generator
The sentences were cropped, and the average length of the sentences was set to 22 after calculation. The similarity between sentences and sentences was set to 0.05 to have the correct action in each candidate action. When the similarity is more significant than 0.05, the connection is created in the graph.
2) Evaluation network
When fixed-value rewards are used, this paper sets the ADD reward value to 0.1.
In this study, eight sets of experiments were conducted using the values of fixed rewards, as shown in Figure 6.
Figure 6 shows that, for the fixed-reward setting, the highest value is reached when the reward value is 0.1. The leads were 0.52 and 1.67 for ROUGE-1 and ROUGE-2, respectively.
3) Common parameters of the evaluation and decision-making networks
Where the discount factor of information entropy is 0.001, the discount factor of the reward was set to 0.99. The lexicon was 32341. The convolutional kernel size was (22, 2). The learning rate was 0.01.
GA2C's loss function consists of three parts, i.e., lossvalue, lossactor and entropy.
entropy= ∑kt=0(logπθ(at∨st)||exp(ft) ∑Cc=1 exp(fc)) | (17) |
The total loss is lossTotal:
lossTotal=lossvalue+β∗lossactor+η∗entropy | (18) |
where β and η are discount factors. The lossTotal function used in this study is consistent with many A2C applications. For example, Li et al. used A2C to diversify recommendation results [30]. The method uses a loss calculation that also adds up the actor and critic parts to perform the update. In this study, the selection summary problem was considered as an action decision problem, which means that the intelligence is trained by putting them into a graph composed of sentences, somewhat like a path-planning problem. For this reason, we applied the loss function used in the A3C Super Mario implementation and added additional entropy [31]. To demonstrate the effect of information entropy on the final effect, a set of comparative experiments was conducted. The values of ROUGE-1, ROUGE-2 and ROUGE-L for the case without η∗entropy were tested when the reward value was 0.1.
The bold in the table represents the best result. As shown in Table 1, GA2C leads to 0.45 and 0.07 for Rouge-1 and Rouge-2, respectively, for the case of added η∗entropy. This result proves that adding η∗entropy is more helpful for the convergence of training.
Model | ROUGE-1 | ROUGE-2 | ROUGE-L |
No added entropy | 40.25 | 27.14 | 35.40 |
With entropy | 40.70 | 27.21 | 35.07 |
To verify the validity of the models in this paper, GA2C was compared with some typical methods on CNN/Daily Mail data, including Lead-3, BANDITSUM [15], Refresh [19], SummaRuNNer [10], Pointer-generator + coverage [32] and deep Q-network (DQN) [33]. Refresh, which combines maximum-likelihood cross-entropy loss with RL, was chosen as the baseline of this study. The model abstracted the single-document summarization task into a sentence ranking task, and global optimization ROUGE was used for extractive summarization. Meanwhile, several classical models were used as comparison models. SummaRuNNer is an extractive model based on a recurrent neural network. The model uses a novel training mechanism that allows the introduction of labels without the training phase. Lead-3 simply uses the first three sentences of the document as a summarization, but it achieved great results. Pointer-generator + coverage incorporates the seq2seq model, a pointer-generator network, a coverage mechanism, mitigating out-of-vocabulary words and information redundancy. The DQN is an RL algorithm based on the DQN method for extractive summarization. The method learns strategies to maximize the ROUGE score in Q-value approximation. BANDITSUM uses the optimized ROUGE score to treat the extractive summarization as a contextual bandit problem.
In addition, we conducted ablation experiments from many angles, including using four activation functions, using different evaluation network AC methods and uniting different reward functions with different mapping methods.
The bold in the table represents the best result. GA2C was compared with several classical models, as shown in Table 2, with the best results indicated by bolded numbers. Compared to Refresh, GA2C led on ROUGE-1, ROUGE-2 and ROUGE-AVG by 0.70, 9.01 and 2.73, respectively. Compared with Refresh, GA2C utilized article-building graphs to better establish the connection between individual sentences and speed up the training. And, GA2C incorporates the A2C algorithm, which is superior to Refresh that only draws on the RL framework. GA2C not only avoids the problems caused by maximum likelihood similarity, it also minimizes the reliance on real tags. It can be seen that the GA2C performance achieved some improvement. This indicates that the critical sentence content extracted by the model in this study was more accurate and fluent.
Model | ROUGE-1 | ROUGE-2 | ROUGE-L | ROUGE-AVG |
SummaRuNNer | 39.60 | 16.20 | 35.30 | 30.37 |
Pointer-generator + coverage | 39.53 | 17.28 | 36.38 | 31.06 |
Lead-3 | 40.34 | 21.00 | 36.57 | 32.64 |
Refresh (baseline) | 40.00 | 18.20 | 36.60 | 31.60 |
DQN | 39.4 | 16.10 | 35.6 | 30.37 |
GA2C | 40.70 | 27.21 | 35.07 | 34.33 |
Moreover, four activation functions were used to test the GA2C model. The default here was to use the cumulative reward method and COS calculation to build the graph. The four activation functions included ReLU [34], Leaky ReLU [35], parametric ReLU (PReLU) [36] and randomized ReLU (RReLU) [37]. The four activation functions are shown in Figure 7.
ReLU: The ReLU activation function was a simple calculation. If the input was greater than 0, it directly returned the value provided as input; if the information was 0 or less, it returned the value 0.
Leaky ReLU: Leaky ReLU was used to assign a non-zero slope to all negative values, and the formula is shown in Eq (19).
yi={xiifxi≥0xiaiifxi<0 | (19) |
where ai denotes a fixed parameter in a (1, +∞) interval.
PReLU: PReLU can be considered as a variant of Leaky ReLU. In PReLU, the slope of the negative part is based on the data and not predefined.
RReLU: RReLU is also a variant of Leaky ReLU. In RReLU, the slope of the negative value is obtained randomly in training and becomes fixed in subsequent tests.
yji={xjiifxji≥0ajixjiifxji<0 | (20) |
where aji is a randomly drawn value from a uniform distribution U (I, u) and I < u; u takes values in the range of [0, 1).
Activation function | ROUGE-1 | ROUGE-2 | ROUGE-L | ROUGE-AVG |
RELU | 40.46 | 26.85 | 34.77 | 34.02 |
RRELU | 40.30 | 26.71 | 34.60 | 33.87 |
PRELU | 40.26 | 26.63 | 34.59 | 33.83 |
Leaky RELU | 40.35 | 26.82 | 34.77 | 33.98 |
The bold in the table represents the best result. By comparison, RELU achieved the best results against the other four activation functions. Compared to Leaky RELU, GA2C led on ROUGE-1, ROUGE-2 and ROUGE-AVG by 0.70, 9.01 and 2.73, respectively. RELU was a simple way to achieve good results on GA2C. In this study, we finally decided to use the RELU activation function as the default function.
Meanwhile, to compare the different evaluation network algorithms (AC and A2C) [38], experiments were conducted in this study using AC combined with graph matrices. The experimental results are shown in Table 4.
Model | ROUGE-1 | ROUGE-2 | ROUGE-L | ROUGE-AVG |
GAC | 40.41 | 26.86 | 34.71 | 33.99 |
GA2C | 40.46 | 26.85 | 34.77 | 34.02 |
The bold in the table represents the best result. The AC algorithm is slightly worse than the A2C algorithm in terms of the combined effect. Compared to AC, A2C led on ROUGE-1, ROUGE-L and Rouge-AVG by 0.70, 9.01 and 2.73, respectively. The AC algorithm has no baseline, and the feedback was positive, which can lead to some actions not converging. However, the A2C algorithm includes a baseline, making the feedback positive and negative. This made it easier to converge and train [39]. The A2C was more suitable for the model in this study.
In addition, to find the best pairing of a reward function and graph building method, we provided two ways of reward calculation: 1) cumulative way to calculate the reward of action, and 2) use of ROUGE to calculate the reward of action. And, two ways of text similarity when building graphs have been provided: (i) Jaccard similarity to calculate sentence similarity and (ii) cosine similarity to calculate sentence similarity. In this study, to try out the best combination, these two were combined separately for experiments.
Model | ROUGE-1 | ROUGE-2 | ROUGE-L | ROUGE-AVG |
GA2C-ADD-COS | 40.46 | 26.85 | 34.77 | 34.02 |
GA2C-ADD-JACCARD | 34.86 | 20.49 | 29.03 | 33.88 |
GA2C-ROUGE-COS | 40.70 | 27.21 | 35.07 | 34.33 |
GA2C-ROUGE-JACCARD | 34.76 | 20.50 | 28.98 | 28.08 |
The bold in the table represents the best result. The GA2C model used ROUGE to calculate rewards and COS to calculate sentence similarity to build graphs, which can yield the best results. The combination of ROUGE and COS performed the best. Compared to ADD and COS, it led on ROUGE-1, ROUGE-2, ROUGE-L and ROUGE-AVG by 0.24, 0.36, 0.30 and 0.31, respectively. The Jaccard was not satisfactory in combination with ROUGE or ADD, which was 6% worse than the combination with COS. This is mainly because Jaccard compared the intersection and concurrence of two sentences, which does not reflect the similar relationships between texts well. The results of the COS experiments were all higher than those of the Jaccard experiments, where ADD was 0.24, 0.36, 0.40 and 0.31 lower than ROUGE for the four metrics when the COS constructed graphs were applied, indicating that ROUGE returned a more accurate reward.
In order to solve the problems of exposure bias and low precision in extractive summarization. This paper introduced the extractive summarization model GA2C. The algorithm transforms the data into a graph matrix and trains the model using an evaluation and a decision-making network. GA2C uses an RL network for feature extraction and achieves the selection of actions by using a decision-making network. The actions selected by the decision-making network are evaluated using the evaluation network and a reward is returned. GA2C utilizes RL to reduce exposure bias greatly. The experimental results showed that the summarization content extracted by GA2C was more accurate, and that the language was more fluent. Compared to Refresh, GA2C led on ROUGE-1, ROUGE-2 and Rouge-AVG by 0.70, 9.01 and 2.73, respectively.
The model in this paper could be further improved, as follows: i) the word vector representation can be trained using Bert, Word2vec, etc.; ii) several more types of connections between sentences can be tried, such as ROUGE or other methods of establishing relationships; and iii) GA2C can be applied to datasets in different domains, such as meeting minutes, emails, etc. In addition, GA2C training is time-consuming. The training time can be reduced in the future by learning from some techniques from graph CNN training. The reward design of GA2C can also be further optimized to improve the accuracy of summary extraction.
The authors declare that there are no conflicts of interest.
[1] | WHO, Hepatitis C, 2020. Available from: https://www.who.int/news-room/fact-sheets/detail/hepatitis-c. |
[2] | M. J. Alter, HCV routes of transmission: what goes around comes around, In: Seminars in liver disease, 31 (2011), 340–346. |
[3] |
A. U. Neumann, N. P. Lam, H. Dahari, D. R. Gretch, T. E. Wiley, T. J. Layden, et al., Hepatitis C viral dynamics in vivo and the antiviral efficacy of interferon-a therapy, Science., 282 (1998), 103–107. doi: 10.1126/science.282.5386.103
![]() |
[4] |
F. Alvarez, P. A. Berg, F. B. Bianchi, International Autoimmune Hepatitis Group Report: review of criteria for diagnosis of autoimmune hepatitis, J. Hepatol., 31 (1999), 929–938. doi: 10.1016/S0168-8278(99)80297-9
![]() |
[5] | D. Das, M. Pandya, Recent Advancement of Direct-acting Antiviral Agents (DAAs) in Hepatitis C Therapy, Mini Rev. Med. Chem., 18 (2018), 584–596. |
[6] | T. Asselah, P. Marcellin, Interferon free therapy with direct acting antivirals for HCV, Liver Int., 33 (2013), 93–104. |
[7] | E. De Clercq, Current race in the development of DAAs (direct-acting antivirals) against HCV, Biochem. Pharmacol., 89 (2014), 441–452. |
[8] | M. A. Nowak, C. R. M. Bangham, Population dynamics of immune responses to persistent viruses, Science, 272 (1996), 74–79. |
[9] | S. Bonhoeffer, J. M. Coffin, M. A. Nowak, Human immunodeficiency virus drug therapy and virus load, J. Virol., 71 (1997), 3275–3278. |
[10] | R. O. Avendan, L. Estevab, J. A. Floresb, J. F. Allen, G. Gomez, J. Lopez-Estrada, A Mathematical Model for the Dynamics of Hepatitis C, J. Theor. Med., 4 (2002), 109–118. |
[11] | Y. Zhao, Z. Xu, Global Dynamics for Delayed Hepatitis C Virus Infection Model, Electron. J. Differ. Equ., 2014 (2014), 1–18. |
[12] | A. Chatterjee, J. Guedj, A. S. Perelson, Mathematical modelling of HCV infection: what can it teach us in the era of direct-acting antiviral agents?, Antivir. Ther., 17 (2012), 1171–1182. |
[13] | H. A. Elkaranshawy, H. M. Ezzat, Y. Abouelseoud, N. N. Ibrahim, Innovative approximate analytical solution for standard model of viral dynamics: hepatitis C with direct-acting agents as an implemented case, Math. Probl. Eng., 2019 (2019). |
[14] | J. Guedj, A. U. Neumann, Understanding hepatitis C viral dynamics with direct-acting antiviral agents due to the interplay between intracellular replication and cellular infection dynamics, J. Theor. Biol., 267 (2013), 330–340. |
[15] | A. N. Chatterjee, M. K. Singh, B. Kumar, The effect of immune responses in HCV disease progression, Eng. Math. Lett., 2019 (2019), 1–14. |
[16] | A. N. Chatterjee, B. Kumar, Cytotoxic T-lymphocyte Vaccination for Hepatitis C: A Mathematical Approach, Stud. Indian Place Names, 40 (2020), 769–779. |
[17] | D. Echevarria, A. Gutfraind, B. Boodram, M. Major, S. Del Valle, S. J. Cotler, et al., Mathematical modeling of hepatitis C prevalence reduction with antiviral treatment scale-up in persons who inject drugs in metropolitan Chicago, PloS One, 10 (2015), e0135901. |
[18] | N. Scott, E. McBryde, P. Vickerman, N. K. Martin, J. Stone, H. Drummer, et al., The role of a hepatitis C virus vaccine: modelling the benefits alongside direct-acting antiviral treatments, BMC Med., 13 (2015), p198. |
[19] | L. Rong, A. S. Perelson, Mathematical analysis of multiscale models for hepatitis C virus dynamics under therapy with direct-acting antiviral agents, Math. Biosci., 245 (2013), 22–30. |
[20] | S. Zhang, X. Xu, Dynamic analysis and optimal control for a model of hepatitis C with treatment, Commun. Nonlinear Sci. Numer. Simul., 46 (2017), 14–25. |
[21] | E. Mayanja, L. S. Luboobi, J. Kasozi, R. N. Nsubuga, Mathematical Modelling of HIV-HCV Coinfection Dynamics in Absence of Therapy, Comput. Math. Methods Med., 2020 (2020), 2106570. |
[22] | S. Drazilova, M. Janicko, L. Skladany, P. Kristian, M. Oltman, M. Szantova, et al., Glucose metabolism changes in patients with chronic hepatitis C treated with direct acting antivirals, Can. J. Gastroenterol. Hepatol., 2018 (2018), 6095097. |
[23] | H. Dahari, A. Lo, R. M. Ribeiro, A. S. Perelson, Modeling hepatitis C virus dynamics: Liver regeneration and critical drug efficacy, J. Theor. Biol., 247 (2007), 371–381. |
[24] | N. K. Martin, P. Vickman, M. Hickman, Mathematical modelling of hepatitis C treatment for injecting drug users, J. Theor. Biol., 274 (2011), 58–66. |
[25] | E. Ahmed, H. A. El-Saka, On fractional order models for Hepatitis C, Nonlinear Biomed. Phys., 4 (2010), 1–3. |
[26] | D. Bainov, P. Simeonov, Impulsive differential equations: periodic solutions and applications, CRC Press, 66 (1993). |
[27] | J. Lou, Y. Lou, J. Wu, Threshold virus dynamics with impulsive antiretroviral drug effects, J. Math. Biol., 65 (2012), 623–652. |
[28] | J. Lou, R. J. Smith, Modelling the effects of adherence to the HIV fusion inhibitor enfuvirtide, J. Theor. Biol., 268 (2011), 1–13. |
[29] | L. Q. Min, Y. M. Su, Y. Kuang, Mathematical analysis of a basic virus infection model with application to HBV infection, Rocky Mt. J. Math., 38 (2008), 1573–1585. |
[30] |
S. A. Gourleya, Y. Kuangb, J. D. Nagy, Dynamics of a delay differential equation model of hepatitis B virus infection, J. Biol. Dyn., 2 (2008), 140–153. doi: 10.1080/17513750701769873
![]() |
[31] | M. Zhao, Y. Wang, L. Chen, Dynamic Analysis of a Predator-Prey (Pest) Model with Disease in Prey and Involving an Impulsive Control Strategy, J. Appl. Math., 2012 (2012), 969425. |
[32] | H. Yu, S. Zhong, R. P. Agarwal, Mathematics analysis and chaos in an ecological model with an impulsive control strategy, Commun. Nonlinear Sci. Numer. Simul., 16 (2011), 776–786. |
[33] | V. Lakshmikantham, D. D. Bainov, P. S. Simeonov, Theory of Impulsive Differential Equations, Series in Modern Appl. Math., 6, 1989. |
[34] |
B. Patricia, Hepatitis C virus and peripheral blood mononuclear cell reservoirs Patricia Baré, World J. Hepatol., 1 (2009), 67. doi: 10.4254/wjh.v1.i1.67
![]() |
[35] |
C. Caussin-Schwemling, C. Schmitt, F. Stoll-Keller, Study of the infection of human blood derived monocyte/macrophages with hepatitis C virus in vitro, J. Med. Virol., 65 (2001), 14–22. doi: 10.1002/jmv.1095
![]() |
[36] |
R. J. Smith, B. D. Aggarwala, Can the viral reservoir of latently infected CD4+ T cells be eradicated with antiretroviral HIV drugs?, J. Math. Biol., 59 (2009), 697–715. doi: 10.1007/s00285-008-0245-4
![]() |
[37] |
P. van den Driessche, J. Watmough, Reproduction numbers and subthreshold endemic equilibria for compartmental models of disease transmission, Math. Biosci., 180 (2002), 29–48. doi: 10.1016/S0025-5564(02)00108-6
![]() |
[38] |
J. M. HefferNan, R. J. Smith, L. M. Wahl, Perspectives on the basic reproductive ratio, J. R. Soc. Interface., 2 (2005), 281–293. doi: 10.1098/rsif.2005.0042
![]() |
1. | Anna Bershteyn, Emma Richard, Qinlian Zhou, Maria R Khan, Joy D Scheidell, Prima Manandhar-Sasaki, Kaoon Ban, Stephen Crystal, Adam J Gordon, Amy C Justice, Kendall J Bryant, R Scott Braithwaite, Potential health benefits of integrated screening strategies for alcohol, tobacco, other substance use, depression, anxiety, and chronic pain among people living with HIV in the USA: a mathematical modelling study, 2023, 10, 23523018, e118, 10.1016/S2352-3018(22)00361-7 | |
2. | 2025, 9781394263721, 191, 10.1002/9781394263752.ch21 | |
3. | Lauren A Taylor, Berkeley Franz, Amanda Zink, Andrew Fair, Cory E Cronin, Public perceptions of US for-profit, nonprofit, and public hospitals, 2023, 1, 2976-5390, 10.1093/haschl/qxad046 | |
4. | Zibei Chen, Gaurav R. Sinha, Jordan DeVylder, Dawnsha Mushonga, Michael Mason, Lisa Fedina, Zakary Amen, Young adults in pandemic financial hardship: implications of the social determinants of mental health, 2024, 22, 1533-2985, 698, 10.1080/15332985.2024.2377959 | |
5. | Alia Alhomsi, Paula D. Strassle, Stephanie Ponce, Izzy Mendez, Stephanie M. Quintero, Miciah Wilkerson, Anita L. Stewart, Anna M. Napoles, Financial Hardship and Psychological Distress During the Pandemic: A Nationally Representative Survey of Major Racial-Ethnic Groups in the United States, 2023, 7, 2473-1242, 395, 10.1089/heq.2022.0197 | |
6. | Laura Lentini, Helia Toutounchi, Alexandra Chapleau, Adam Le, Simon Fournier, Fatemeh Emari, Robert Flamini, Andrea Rossi, Angela Gentile, Enrico Bertini, Francesco Nicita, Daniela Pohl, Sunita Venkateswaran, Stephanie Keller, Elsa Rossignol, Deborah Renaud, Danilo De Assis Pereira, Xiaoru Chen, Adeline Vanderver, Geneviève Bernard, Stress and Quality of Life of Parents of Children With POLR3-Related Leukodystrophy: A Cross-Sectional Pilot Study, 2024, 0883-0738, 10.1177/08830738241283171 | |
7. | Tatiane Santos, Richard C. Lindrooth, Shoou-Yih Daniel Lee, Kelsey Owsley, Gary J. Young, Oregon Community Benefit Reform Influenced Not-For-Profit Hospitals’ Charity Care And Medical Debt Write-Off, 2025, 44, 0278-2715, 196, 10.1377/hlthaff.2024.00760 | |
8. | Abdullahi Tunde Aborode, Oreoluwa Oginni, Modesta Abacheng, Ottoho Edima, Emily Lamunu, Tayo Nafisat Folorunso, Christian Inya Oko, Akinjo Rebecca Iretiayo, Lukman Lawal, Ruth Amarachi, Rawa Badri, Gafar Babatunde Bamigbade, Omotayo Faith Olanrewaju, Favour Obianuju Agwuna, Ridwan Olamilekan Adesola, Healthcare debts in the United States: a silent fight, 2025, 87, 2049-0801, 663, 10.1097/MS9.0000000000002865 |
Model | ROUGE-1 | ROUGE-2 | ROUGE-L |
No added entropy | 40.25 | 27.14 | 35.40 |
With entropy | 40.70 | 27.21 | 35.07 |
Model | ROUGE-1 | ROUGE-2 | ROUGE-L | ROUGE-AVG |
SummaRuNNer | 39.60 | 16.20 | 35.30 | 30.37 |
Pointer-generator + coverage | 39.53 | 17.28 | 36.38 | 31.06 |
Lead-3 | 40.34 | 21.00 | 36.57 | 32.64 |
Refresh (baseline) | 40.00 | 18.20 | 36.60 | 31.60 |
DQN | 39.4 | 16.10 | 35.6 | 30.37 |
GA2C | 40.70 | 27.21 | 35.07 | 34.33 |
Activation function | ROUGE-1 | ROUGE-2 | ROUGE-L | ROUGE-AVG |
RELU | 40.46 | 26.85 | 34.77 | 34.02 |
RRELU | 40.30 | 26.71 | 34.60 | 33.87 |
PRELU | 40.26 | 26.63 | 34.59 | 33.83 |
Leaky RELU | 40.35 | 26.82 | 34.77 | 33.98 |
Model | ROUGE-1 | ROUGE-2 | ROUGE-L | ROUGE-AVG |
GAC | 40.41 | 26.86 | 34.71 | 33.99 |
GA2C | 40.46 | 26.85 | 34.77 | 34.02 |
Model | ROUGE-1 | ROUGE-2 | ROUGE-L | ROUGE-AVG |
GA2C-ADD-COS | 40.46 | 26.85 | 34.77 | 34.02 |
GA2C-ADD-JACCARD | 34.86 | 20.49 | 29.03 | 33.88 |
GA2C-ROUGE-COS | 40.70 | 27.21 | 35.07 | 34.33 |
GA2C-ROUGE-JACCARD | 34.76 | 20.50 | 28.98 | 28.08 |
Model | ROUGE-1 | ROUGE-2 | ROUGE-L |
No added entropy | 40.25 | 27.14 | 35.40 |
With entropy | 40.70 | 27.21 | 35.07 |
Model | ROUGE-1 | ROUGE-2 | ROUGE-L | ROUGE-AVG |
SummaRuNNer | 39.60 | 16.20 | 35.30 | 30.37 |
Pointer-generator + coverage | 39.53 | 17.28 | 36.38 | 31.06 |
Lead-3 | 40.34 | 21.00 | 36.57 | 32.64 |
Refresh (baseline) | 40.00 | 18.20 | 36.60 | 31.60 |
DQN | 39.4 | 16.10 | 35.6 | 30.37 |
GA2C | 40.70 | 27.21 | 35.07 | 34.33 |
Activation function | ROUGE-1 | ROUGE-2 | ROUGE-L | ROUGE-AVG |
RELU | 40.46 | 26.85 | 34.77 | 34.02 |
RRELU | 40.30 | 26.71 | 34.60 | 33.87 |
PRELU | 40.26 | 26.63 | 34.59 | 33.83 |
Leaky RELU | 40.35 | 26.82 | 34.77 | 33.98 |
Model | ROUGE-1 | ROUGE-2 | ROUGE-L | ROUGE-AVG |
GAC | 40.41 | 26.86 | 34.71 | 33.99 |
GA2C | 40.46 | 26.85 | 34.77 | 34.02 |
Model | ROUGE-1 | ROUGE-2 | ROUGE-L | ROUGE-AVG |
GA2C-ADD-COS | 40.46 | 26.85 | 34.77 | 34.02 |
GA2C-ADD-JACCARD | 34.86 | 20.49 | 29.03 | 33.88 |
GA2C-ROUGE-COS | 40.70 | 27.21 | 35.07 | 34.33 |
GA2C-ROUGE-JACCARD | 34.76 | 20.50 | 28.98 | 28.08 |