In this paper, the state-flipped control technique is explored to investigate the stabilization of probabilistic Boolean networks (PBNs). Changing the values of many nodes from 0 to 1 (or from 1 to 0) is called the state-flipped control. The concepts of fixed point, reachable sets, and finite-time global stabilization of PBNs under state-flipped control are proposed. Several necessary and sufficient conditions for global stabilization are also derived based on the reachable sets of a given state. Furthermore, a model-free reinforcement learning (RL) algorithm, namely $Q$-learning ($Q$L), is presented to design a flip sequence for any state that steers the state to a given destination state, thereby achieving finite-time global stabilization via state-flipped control. In addition, the process of finding the minimum flip set is proposed under the semi-tensor product and $Q$L methods. Finally, the viability of the results in the paper is shown by considering a 12-gene hepatocellular cancer cell tumor network.

Stabilization of Probabilistic Boolean Networks via State-Flipped Control and Reinforcement Learning

Yerudkar A.;Vecchio C. D.
2023-01-01

Abstract

In this paper, the state-flipped control technique is explored to investigate the stabilization of probabilistic Boolean networks (PBNs). Changing the values of many nodes from 0 to 1 (or from 1 to 0) is called the state-flipped control. The concepts of fixed point, reachable sets, and finite-time global stabilization of PBNs under state-flipped control are proposed. Several necessary and sufficient conditions for global stabilization are also derived based on the reachable sets of a given state. Furthermore, a model-free reinforcement learning (RL) algorithm, namely $Q$-learning ($Q$L), is presented to design a flip sequence for any state that steers the state to a given destination state, thereby achieving finite-time global stabilization via state-flipped control. In addition, the process of finding the minimum flip set is proposed under the semi-tensor product and $Q$L methods. Finally, the viability of the results in the paper is shown by considering a 12-gene hepatocellular cancer cell tumor network.
2023
$Q$ -learning ( $Q$ L)
Asymptotic stability
Controllability
finite-time global stabilization
Probabilistic Boolean networks (PBNs)
Probabilistic logic
semi-tensor product (STP)
Stability criteria
State feedback
state-flipped control
Switches
Transient analysis
File in questo prodotto:
Non ci sono file associati a questo prodotto.

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/20.500.12070/63159
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 3
  • ???jsp.display-item.citation.isi??? ND
social impact