Volume 6 Supplement 3
Theoretical aspects and modelling of cellular decision making, cell killing and information-processing in photodynamic therapy of cancer
© Gkigkitzis; licensee BioMed Central Ltd. 2013
Published: 11 November 2013
The aim of this report is to provide a mathematical model of the mechanism for making binary fate decisions about cell death or survival, during and after Photodynamic Therapy (PDT) treatment, and to supply the logical design for this decision mechanism as an application of rate distortion theory to the biochemical processing of information by the physical system of a cell.
Based on system biology models of the molecular interactions involved in the PDT processes previously established, and regarding a cellular decision-making system as a noisy communication channel, we use rate distortion theory to design a time dependent Blahut-Arimoto algorithm where the input is a stimulus vector composed of the time dependent concentrations of three PDT related cell death signaling molecules and the output is a cell fate decision. The molecular concentrations are determined by a group of rate equations. The basic steps are: initialize the probability of the cell fate decision, compute the conditional probability distribution that minimizes the mutual information between input and output, compute the cell probability of cell fate decision that minimizes the mutual information and repeat the last two steps until the probabilities converge. Advance to the next discrete time point and repeat the process.
Based on the model from communication theory described in this work, and assuming that the activation of the death signal processing occurs when any of the molecular stimulants increases higher than a predefined threshold (50% of the maximum concentrations), for 1800s of treatment, the cell undergoes necrosis within the first 30 minutes with probability range 90.0%-99.99% and in the case of repair/survival, it goes through apoptosis within 3-4 hours with probability range 90.00%-99.00%. Although, there is no experimental validation of the model at this moment, it reproduces some patterns of survival ratios of predicted experimental data.
Analytical modeling based on cell death signaling molecules has been shown to be an independent and useful tool for prediction of cell surviving response to PDT. The model can be adjusted to provide important insights for cellular response to other treatments such as hyperthermia, and diseases such as neurodegeneration.
Conventional cancer therapies include radiation and chemotherapies, surgery, and a combination of any or all of those therapies. The treatments themselves have important side effects, even life-threatening. Chemotherapy is known to impose difficulties because drugs often produce harmful side effects and x-rays sometimes damages normal tissue. Photodynamic therapy offers an alternative, less invasive treatment for such illnesses such as several types of cancers. It involves the use of three basic components : a photosensitizer, a light-absorbing molecule that is activated by the second element, light of a corresponding wavelength, and third, molecular oxygen is consumed during the photochemical reaction to produce cytotoxic agents, thus destroying neoplastic tissue. It is accepted that cell photo-killing (induced in cultured cells) may involve all three main cell death morphologies described, i.e. apoptotic, necrotic and autophagy cell death . Dynamic modeling of cell fate exists, for apoptosis/necrosis  and for autophagy , in a single cell model. In a previous work  we established a model of oxygen transport and cell killing in Type II PDT. This model can be directly linked to these cell fate models, to provide a coherent model of the major biochemical events in PDT on the basis of major components and the main features of the intracellular interactions. Based on existing system biology models [3–5] it is possible to develop a detailed a molecular interaction diagram that summarizes the major biochemical features of the photochemical processes, together with a corresponding system of molecular interactions, rate equations, reaction constants and initial conditions. In the modeling and simulation sections we briefly summarize these major facts of the cell biochemistry of Type II PDT, since the focus of this contribution is the study of the cell decision mechanism of a single cell model in response to PDT treatment and the probability of cell survival.
Information can be defined in terms of its ability to increase the probability of something being true  and it is carried on a channel which is a physical mechanism for communication. A channel is distinguished by having a limit on its ability to carry information and by the fact that it is susceptible to random interference, called noise . Whenever energy is transferred, information is transferred. In PDT, light energy is absorbed by the photosensitizers and then transferred to oxygen and other molecules, through a cascade of reactions in the environment of a cell . The PDT treatment parameters act as a "source" generating the input information that the system of molecular network and interactions within a cell must communicate to the "receiver" or the cell . Information is encoded by the parameters of the light and the photosensitizer doses as the source "words" or "code" (death signals) and is transformed into a form, through activated photosensitizers, that can be transmitted through the "channel" of molecular interactions. When decoded by molecular "thresholds", the input information can be converted to a channel output that has the form of a cell's state in terms of necrosis, apoptosis, autophagy or survival . The performance-efficiency of such a bio-communication system and its usefulness for modeling the experimental data is quantified through the assignment of numerical values to the variations and errors that the system may produce. In particular, the mechanism that governs the generation of the source signal and the distortion measure that penalizes the bio-coding errors and determines the fidelity of the reproduction of the cell killing signal need is identified through measurable quantities - functions. Ultimately, the goal is to design and model an optimal treatment strategy that, through the scheme of intracellular biochemical reactions, may lead to reproduction of the PDT death signal output after processing by the cell, with an average distortion that does not exceed a specified upper level D, for a single tumor cell model (or in general, a tumor cell population).
The treatment pattern of the a priori setting parameters (light density, photosensitizer concentration, etc.), is related to the data bio-compression of the death signal through molecular interactions, and the classification of the signal as to cell death or cell survival is done with a possible statistical error that is assigned a numerical penalty: the distortion function or distortion measure d. The distortion function d does not in itself wholly determine a cell decision. What is important is the relationship between the distortion function and the prior probability distributions of cell death signaling molecules. It is possible to have two different distortion functions which lead to the same decision when the prior probability distributions associated with each, compensate for the details of each distortion function. Combining the three elements of the prior probability (distributions of molecular concentrations), the cell data (distortion tolerance of the cellular system), and the distortion function then allows cell fate decisions to be based on minimizing the mutual information between input and output. The minimization of the mutual information as an application of rate distortion theory to decision making mechanisms in biology has been adapted for testing a framework for designing and analyzing binary decision-making strategies in cellular systems , for the information-theoretic characterization of the optimal gradient sensing response of cells  and for the rate distortion approach to protein symmetry  among other applications. The mutual information of two random variables in general, measures the information that × (input) and Y (output) share: it measures how much knowing one of these variables reduces the entropy of the other. This reduction of the entropy will be compensated by the cell either by interaction with the environment, while it is in a vulnerable state and its survival probability decreases, or by cell death.
There are two kinds of cell division: mitosis and meiosis. Mitosis is essentially a duplication process: It produces two genetically identical "daughter" cells from a single "parent" cell. All cells must replicate their DNA prior to cell division. This assures that each new cell produced receives all of the genetic material necessary to survive and reproduce. Therefore certain information that is formless and does not change or die nor is it composed of matter, is "carried" from one cell to the next and is "reproduced" and it is always present in every cell structure.
John Von Neumann posed and solved the following question: what kind of logical organization is sufficient for an automaton to control itself in such a manner that it reproduces itself? . A cellular automaton is specified by giving a finite list of states for each cell, a distinguished state called the blank state, and a rule which gives the state of a cell at time t+1 as a function of its own state and the states of its neighbors at time t. It consists of a cellular space and a transition function defined over this space. Finite automata constitute the basis of Turing machines . Von Neumann was the first to provide an algorithmic model of a self-reproducing automaton, the Universal Constructor, a self-replicating machine in a cellular automata environment and (in a brief summary) he proved that the construction of this sort of automaton would necessitate the solution to four fundamental problems [11, 12]:
α. to store instructions in a memory;
β. to duplicate these instructions;
γ. to implement an automatic factory ("Universal Constructor"), able to read the memory instructions, and, based on them, to construct the components of the system;
δ. to manage all these functions by means of a central control unit.
A self-reproducing system must contain the program of its own construction. This program is a sort of consistent and complete abstract image of the system. In other words, self-reproduction needs programming and processors (software -for information based replication- and hardware). The solution to these problems mentioned above may be found in living things as observed by modern biology. An efficient mechanism of information storage and an elegant mechanism of duplication of the DNA molecule may be the one and only perfect solution to the twin problems of information storage and duplication for self-replicating automata . But more importantly, Von Neumann understood that any information-based replicator must contain inside itself (among other indispensable things) a symbolic representation of itself, an "image" of itself. The relation between the replicator (hardware) and the image (a structure of symbols, the software) is a functional relation of dependence, since the symbolic representation consists of directives and instructions that must be interpreted by the replicator machinery for constructing a copy of itself. As reported by Luis Rocha in his 2012 Fall lecture notes "Biologically-inspired computing", Indiana University, Von Neumann proposed this scheme before the structure of the DNA molecule was uncovered by Watson and Crick, though after the Avery-MacLeod-McCarty experiment which identified DNA has the carrier of genetic information.
With respect to Von Neumann's Universal Constructor we need to notice first that the four principles (α,β,γ,δ) mentioned above are irreducible in complexity and secondly that the concept of the symbolic representation-based self-reproduction implies a language (a symbol system, a syntactic code to be used to map instructions into construction commands for replication. In copying a description, the syntactic aspects are replicated. This indicates that the appropriate framework for the study of such systems could be information theory as discussed below. It is important to decipher the meaning of information available to a cell as something that determines its activity. Information has no mass, energy, or spatial extension, it cannot be seen, touched, or smelled. Nevertheless it is a distinct, objective entity. The cell, as an information system has the ability to discriminate and select between cell fates (which is what we call cell decision making). In fact, the manifestation of information can be found in the existence of alphabets (where as alphabet we interpret the set of physical states that can be realized in some system), the combination of codes (where as a code we consider a collection of the letters of alphabets that follow some pattern-words) and the variety of codes that determine the state of the system.
A brief summary of major molecular pathways and biochemical events induced by PDT
Rate distortion theory
And this determines the distortion measure.
Assign a value ε that determines the accuracy of the algorithm.
Initialize the exponential function .
Initialize the probability distribution of the decision as a binary distribution as in equation (4).
Given compute the conditional probability distribution that minimizes the mutual information while satisfying the condition of equation (6).
Update the probability that minimizes the mutual information by using equation (5).
Calculate the distortion D given by the equation (9).
- g)Calculate the rate distortion function given by :(14)
Iterate steps (a-g) for all simulation times t, and this yields for one set values of PDT treatment parameters: photo-density ρ, drug concentration [S0] and initial molecular oxygen concentration .
(optional) Vary (, [S0], [F]) where F is the fluence, is the molecular oxygen concentration in the cell, [S0] is the concentration of ground state photosensitizer, and ([F]is the fluence (see ) and obtain corresponding survival curve for optimization of fluence/drug dose modeling parameters for the biomolecular mechanism studied and the given choice of molecular components of the stimulus vector). Fit to experimental data of survival curves to find the optimal range for the parameters.
Survival functions - predator prey models
Survival functions can be derived using predator-prey model. The predator-prey model has been used for the description of the survival probability in dynamic energy budget models  under the assumption that that the per capita death rate has two contributions, a constant loss due to random misfortunes, and a density-dependent loss due to predation, with a Holling Type II functional form. This model was designed to predict the growth and reproduction patterns of a species based on the characteristics of individual organisms, particularly the strategy used to allocate resources.. This model takes an individual-based approach where all members of the prey population are "copies" of one individual, and each "copy", could be the "model individual" itself. The use of a predator-prey model (a continuous model used for the simulation of discrete population dynamics) for the modeling of survival probability(a continuous variable) suggests the quantization of survival probability. Indeed, the quantization of probability has been proposed by other authors [18, 19]. The existence of the "chance-quantum" (c.q.), implies certain axioms [Go 43]. For example, if the probability of an event is equal to or greater than one c.q., it may ultimately occur, if an event has a calculated probability of less than one cq. it will not occur, for an event having an appreciable probability (equivalent to many cq.), a change in surrounding conditions leading to a computed change in probability of less than one cq. will in fact cause no change in the probability of the event, etc.
Survival Units Duality refers to the idea that the life a cell (or survival probability) can be discretized (quantized) in quanta of life (survival units) which are assumed here as the basic units of life in every cell. New cells are produced by existing cells, and therefore the termination of a cell does not allow to assign any morphological or biochemical characteristics to the life of the cell itself, since these characteristics can only be considered as the manifestations of the monitoring, interaction and response of the cell, as a biochemical unit undividedly united to cellular life ("life units"), to the extracellular environment. Cellular life (survival probaility) is a set of life units (survival probaility quanta), where each cellular life unit contains the whole complete life of the cell in itself, therefore allowing the cell to repair itself after any loss of survival units due to the attack of cell death inducers or other factors (Figure 3). Therefore, the survival probability of a cell is a set that contains itself within each of its elements (survival probablity quanta or units). This idea is not new in mathematics. This is in accordance with Von Neumann's idea that any information-based replicator must contain inside itself a symbolic representation of itself, an "image" of itself.
Russell's paradox and information
A set is a collection of objects, or elements. Sets are defined by the unique properties of their elements and sets and elements may not be mentioned simultaneously, since sets are determined by their elements and therefore one notion has no meaning without other. Bertrand Russell, while working on his "Principia Mathematica" (Principles of Mathematics) in 1903, he discovered a paradox that arised from Frege's set theory that leads to a contradiction . It says "the set of all sets which are not members of themselves contains itself." In mathematical terms, let , then S ∈ S ⇔ S ∉ S. Although the precise rules for set formation have been under intense investigations and several different logical systems have been proposed, sets that contain themselves as elements, like S, are definitely ruled out, as "abnormal". Based on the work Russell and Whitehead, Kurt Gödel was able to show that a theorem could be stated within the context of Russell and Whitehead's system that was impossible to prove within that system . Gödel's Incompleteness Theorem states that there are mathematical statements that can never be proved, in any consistent system of axioms such as the arithmetic system.
The need for the distinction between two kinds of collection can be found back in the work of Schroder and Cantor :
"If we start from the notion of a definite multiplicity of things, it is necessary, as I discovered, to distinguish two kinds of multiplicities (by this I always mean definite multiplicities). For a multiplicity can be such that the assumption that all of its elements "are together" leads to a contradiction, so that it is impossible to conceive of the multiplicity as a unity, as "one finished thing". Such multiplicites I call absolutely infinite or inconsistent multiplicities.... If on the other hand the totality of the elements of a multiplicity can be thought of without contradiction as "being together", so that they can be gathered together into "one thing", I call it a consistent multiplicity or a "set".
Cantor's conclusions are the ancestors of today's distinction between classes and sets, as they appear in the work of Von Neumann . For von Neumann all sets are classes, but not all classes are sets. And those classes that are not sets - the so-called proper classes -cannot themselves be members . In Von Neumann's axiomatization theory, some major advantages are : There are extensions for the predicates 'set', 'non-self-membered set', 'well-founded set', 'ordinal'. There is a well-determined collection of all the Zermelo-Fraenkel sets; and there is a domain for quantification over sets. Further, the Axiom of Choice is provable in von Neumann's system. Several issues, both technical and intuitive, have been reported with respect to this system. A discussion can be found in , and here we only mention the consequence of this theory, that the concept of class has no extension (based on the axioms of this system, there is no class of all classes, and therefore the problem has just been pushed back). Therefore the resolution of this paradox remains unresolved.
In mathematical logic, it is suggested that problems that are essentially the same must be resolved by the same means, and similar paradoxes should be resolved by similar means. This is the principle of uniform solution . Two paradoxes can be thought to be of the same kind when (at a suitable level of abstraction) they share a similar internal structure, or because of external considerations such as the relationships of the paradoxes . The question rises as to the existence of other paradoxes that are of the same kind with Russell's paradox. Russell focused more on the underlying structure of the paradoxes and saw them all as paradoxes of impredicativity. The "inclosure schema" was proposed by Priest, as a formal schema that can be used to classify paradoxes . Although the schema will not be analyzed in this work, the conclusion is very interesting: Russell's paradox is of one kind with the "sorites" paradox (the paradox of the "heap"). This paradox was introduced by to Eubulides of Miletus (4th century BC), a pupil of Euclid, and appears when one considers a heap of sand, from which grains are removed. Is it still a heap when only one grain remains? If not, when did it change from a heap to a non-heap? These two paradoxes are neighboring paradoxes, and it has been suggested that we should not just consider the internal structure of the paradoxes, although that is undoubtedly important, but we also consider the external relationships--the relationships to other nearby paradoxes . The way nearby neighbors (paradoxes of one kind) respond or fail to respond to proposed treatments tells us something about what makes the whole family tick and about their structural similarity . The question "when is the cell dead?" indicates confusion between cessation of organic coherence and cellular activity. When a cell irrevocably loses its organization, it's dead. The point when it becomes irrevocably damaged is related to the sorites problem.
It is known that continuous time Markov processes, are used for the formulation of stochastic predator prey models that are based on within individual variation . Within individual variation, used under the name of "demographic stochasticity", has been used in the theory of adaptive dynamics. The theory of adaptive dynamics aim at describing the dynamics of the dominant trait in a population, that is called the 'fittest' trait. The main approach is through stochastic or individual centered models which in the limit of large population, can be transformed into integro-differential equations or partial differential equations . Stochastic simulations, using a finite size population, involve extinction phenomenon operating through demographic stochasticity which acts drastically on small populations . These simulations involve a unit for minimal survival population size, which corresponds to a single individual. In general though, typical stochastic and deterministic simulations do not fit and give rather different behaviors in terms of branching patterns. It has been observed that the notion of demographic stochasticity does not occur in general in deterministic population models, and an alternative proposed has been proposed in order to include a similar notion in these models: the notion of a survival threshold , which allows some phenotypical traits of the population to vanish when represented by too few individuals. In particular, through the investigations of simple and standard Lotka Volterra systems that describe the time of the distribution of phenotypic traits in time, it is shown that the inadequacy of deterministic models to handle extinction phenomena through demographic stochasticity, can be corrected by the introduction of a survival threshold, leading to a mimicking effect of the extinction probability due to demographic stochastcity in small sub-populations, while hardly influences the dynamics of large sub-populations . In this framework, the above principle implies (at the extreme) that densities correspond to less than one individual are undesirable , indicating that the link between the continuous (large populations) and the discrete (small sub populations), between the existence (survival) and the vanishing (extinction - demographic stochasticity) is correlated with the existence of a survival threshold in the model.
Furthermore, this hybrid approach of survival, as continuous-discrete function with a survival threshold assigned to a population, raises the following question: Is there an internal quantization scheme that relates the continuous models for large populations with survival thresholds to small populations' discrete models? The existence of both features, of continuity and quantization in a single process, appears in the study of the conditional survival probabilities of a firm (the computation of the conditional survival probability of the firm from an investor's point of view, i.e., given the "investor information"). Callegaro and Sagna used a quantization procedure, to analyze and compare the spread curves under complete and partial information in new and more general settings in their work on applications to credit risk of optimal quantization methods for nonlinear filtering. The theory of quantization probability they used was based on an earlier study of local quantization behavior of absolutely continuous probabilities . This study analyzes the L r quantization error estimates for L r (P) codebooks for absolutely continuous probabilities P and and Voronoi partitions satisfying specific conditions. But the origins of the theory developed there can be traced back to electrical engineering and image processing and in particular in digitizing analog signals and compressing digital images . Therefore, in the heart of the study of survival probabilities we find a theory for the quantization as analog-to-digital conversion and as data compression. Analog signal is a continuous signal which transmits information as a response to changes in physical phenomenon and uses continuous range of values to represent information, where digital signals are discrete time signals generated by digital modulation and use discrete or discontinuous values to represent information. The quality of a quantizer can be measured by the goodness of the resulting reproduction of a signal in comparison to the original. This is accomplished with the definition of a distortion measure that quantifies cost or distortion resulting from reproducing the signal, and the consideration of the average distortion as a measure of the quality of a system, with smaller average distortion meaning higher quality .
This is precisely the framework we adopt in this work to study and analyze the process of cell survival during treatment (in our framework). This suggests an organic connection among an axiomatic system foundation, a predator prey rate equation and information theoretic signal processing.
Mathematical modeling scheme
Where is the expression level of a molecular concentration and is a sigmoidal function that varies from 0 (when ) to 1 (when ). The parameter σ controls the steepness of the sigmoidal function at its inflection point. W i is the net effect on molecule i of all molecules in the network. The coefficient ω ij is less than 0 if molecule j inhibits the expression of molecule i, more than 0 if molecule j activates molecule i, or equal to 0 if there is no effect of molecule j on molecule i. This equation has the great advantage that it is subject to all the powerful analytical and simulation tools of nonlinear ODEs, yet, in the limit of large σ, it behaves like a discrete Boolean network . When σ ≫ 1, tends to flip (on a timescale ≈ γ-1) between 0 and 1, and the dynamical system approximates a Boolean network .
Modeling and simulation
The total time for the simulations was up to 30,000 sec to monitor post-treatment cell killing. We used the stiff solver (ode15s) by MATLAB (The Math Works, Natick, MA) to obtain the solution vector as a function of illumination and observation times, from the start of illumination at t = 0 to 1800 (s) (end of illumination time) and from 1800 to 30,000 (s). Experimental verification of these quantities that describe the levels of all these molecular concentrations can be very difficult if not impossible and they relate indirectly to the ultimate consequence of PDT for cell killing. In  we introduced a cell killing model that related the molecular concentrations of the singlet oxygen and the unoxidized receptors to the cell survival ratio, which can be measured with an in vitro cell model. The same software (MATLAB) is used for producing the simulations for the decision mechanism of a single cell model design. The output of the time dependent Blahut Arimoto algorithm is the cell survival probability (Figure 3). The distortion measure d that quantifies how disadvantageous a decision y is, in response to the stimulus vector is defined by the equation if for some i, and if for some i, and a small number otherwise. The thresholds for the normalized concentrations were all set to 0.5. This distortion measure penalizes a cell survival error more than cell death error for given stimuli, by one order of magnitude. For the range of the Lagrange multipliers the equation s = -e-nwas used and in the simulations n varied over a finite set of integers (a sample of n values from 1 to 20 was taken for the simulations below). The initial survival probability was set equal to 0.9. The treatment parameters for the PDT model that was introduced in our previous work , was linked to the input of this algorithm.
The effort to link biochemical pathways and molecular interactions to the behavior of whole cells and to infer causality from statistical correlation in large data sets in photo-chemotherapy is a matter of considerable difficulty, and to account for all biological variation is a very challenging goal. The existence of more than one PDT tissue destruction mechanism in vivo for the treatment of intraocular retinoblastoma like tumor, has been suggested and documented in  where an early direct cell damage was followed by a subsequent late damage occurring in the tumor tissue left in situ after treatment, resulting in a biphasic pattern in the cell survival curve as a function of time. In , experiments on Chinese hamster cells with phthalocyanine dyes and split light fluence indicated that cells can repair sublethal photo cytotoxic damage during the course of several hours. Although direct cytotoxicity to the tumor cells has been shown to be relatively small after PDT and to increase with time after treatment , examples of in vitro mammalian cell curves as functions of exposure time for different photosensitizer concentrations show that for an acute high dose treatment (vast majority of PDT treatments) the cell survival ratio decreases to less than 1% in the course of a few minutes.
Discrepancies may be due to many factors such as light attenuation passing through the skin resulting in a relatively lower energy dose to some cells than others or the fact that the tumor vasculature is a primary target of PDT. The local micro-environment might have significant impact on PDT response. Vascular effects can be secondary to cell death or conversely, cell death can be secondary to vascular shutdown. Another factor that might affect the final outcome is the triggering of the immune responses, local or systemic.
According to Langton, the 'logical form' of an organism can be separated from its material basis of construction, and that 'aliveness' will be found to be a property of the former, not of the latter . It is the major assumption in the field of Artificial Life (AI) that life is a property of the organization of matter, rather than a property of the matter itself. Organization reduces uncertainty through a process of information collection, management and use. A conceptual and mechanistic system biology mathematical model that is based on information theory can yield valuable insights since cellular behavior cannot be summarized in population averages . The Blahut Arimoto model has several features that are consistent with the experimental results. For the parameter s, estimation can be performed using experimental data, and a range of values can be recovered. The shapes of the survival curves and their correlation with the parameter s will depend on the structure of the rate equations, the type of cell decision algorithm adopted and the accuracy of the experimental data. Different values of the parameters will be predictive of different model curve topologies. Although the origin of the cell parameter s in a cell population for this biochemical model remains non-identifiable from the biophysics point of view, high-likelihood predictions can still be made by appropriate choice and calibration of this parameter. The survival probability predicted by the rate distortion function and calculated by the Blahut Arimoto algorithm, and the variability in the graphs resulting from different values of the parameters provide a framework for the interpretation of self-renewal capabilities of the cell and its ability to generate drug resistance.
The model presented in this report is applicable to the study of cell killing mechanisms in other cases such as hydrogen peroxide H2O2 induced cell death in neurodegenerative diseases. Neuronal death observed in neurodegenerative disorders has been shown to be related to free radical damage and the mechanisms by which reactive oxygen species may damage or kill neurons have been investigated, with a series of experiments designed to document events associated with H2O2 induced cell death in primary neuronal culture . Moreover, this model provides a conceptual frame for the study of hyperthermia induced cell death. Hyperthermia also induces apoptosis in a wide range of cancer cells . The way hyperthermia initiates the intrinsic pathway of apoptosis is yet not completely elucidated, but it is known to involve the transmission of the temperature elevation signal to the mitochondrion through proteins belonging to the Bcl-2 family. Recent preclinical developments show the importance of heat shock proteins and other proteins interfering and regulating the intrinsic and extrinsic pathways of apoptosis . It has been suggested that intracellular de novo synthesis and polymerization of both RNA- and DNA-molecules as well as protein synthesis are decreased in vitro at temperatures between 42 and 45°C in a dose dependent manner. Whereas RNA- and protein synthesis recover rapidly after termination of heat exposure, DNA-synthesis is inhibited for a longer period [39, 40]. The heat shock induces an aggregation of denatured proteins at the nuclear matrix owed to the insolubility of cellular proteins after heat-induced protein unfolding, entailing an enhancement of the nuclear protein concentration. Increase of the nuclear protein content by heat consequently affects several molecular functions (including DNA-synthesis and -repair) when a certain thermal dose is exceeded. A variation of the threshold dose among distinct cell lines is to be expected. Cells are surrounded by electromagnetic fields and the ion distributions inside and outside the cells are also at different concentrations depending on their charge and their type (healthy cells or tumor cells). Heat is transported by means of the extracellular space, ionically bound, to the intracellular space. After the compensation capacity of the cell is exceeded, it is natural to expect that a very small change in the temperature within and outside the cell membrane will be sufficient to affect, or block the metabolic processes or even to denature the proteins (Celsius 42+, Koln, Germany, modalities and procedural technologies for clinical hyperthermia). Now, heat is not a property of a system or body, but instead is always associated with a process of some kind, and is synonymous with heat flow and heat transfer. It has the characteristic feature that it increases the entropy. According to Shannon's theory, entropy measures the information contained in a message (entropy is often used as a characterization of the information content of a data source). Assuming that heat increases the conformational entropy which implies the thermic denaturation of proteins and that heat shock proteins down regulate antioxidants and therefore up regulate oxidative stress as demonstrated by several experimental studies, leading to lipid peroxidation of the lipid components, a model similar to one presented above for PDT can be derived for making binary cell fate decisions (death/survival), as a result of heat exposure (hyperthermia). There are some obvious similarities with PDT such that heat toxicity is the result of oxidative stress. A computational model that will be based on information theory (information enters the cell in this case in the form of increasing entropy, a word with informational context "disorganize") can be used not only to quantitatively describe how the heat shock signal is transformed to a cell death probability (the activation of the thermal signal processing occurring when any of some predefined molecular stimulants increases higher than a predefined threshold) but also to elucidate the main factors causing cell death after heat exposure from a qualitative viewpoint.
Information implies both facts and transmission of facts. The definition of information is content neutral and in Shannon's distortion theory, information is interpreted as what reduces uncertainty. It also presupposes knowledge of a priori probabilities. These probabilities need to be designed or calculated in a way that they will reflect the varieties of environmental stimuli. It is important to decipher the meaning of information available to a cell as something that determines its activity. Information has no mass, energy, or spatiotemporal extension. Nevertheless it is a distinct, objective entity. This entity can be traced through detectable differences. For example, the cell, as an information system has the ability to discriminate and select between cell fates (which is what we call cell decision making). In fact, the manifestation of information can be found in the existence of alphabets (where as alphabet we interpret the set of physical states that can be realized in some system), the combination of codes (where as a code we consider a collection of the letters of alphabets that follow some pattern-words) and the variety of codes that determine the state of the system.
In the work of James G. Miller  on living systems, it was postulated that by the information input of its charter or genetic input, or by changes in behavior brought about by rewards and punishments from its suprasystem, a system develops a preferential hierarchy of values that gives rise to decision rules which determine its preference for one internal steady-state value rather than another. This was defined as the purpose of the system, which will also have an external goal related to its purpose. Therefore the goal of the system is determined by a system on a higher level. This was confirmed in the study of a cell model in the work of Perkins and Swain  who characterized cellular decision-making as having three main tasks: a cell must (1) estimate the state of its environment by sensing stimuli; (2) make a decision informed by the consequences of the alternatives; and (3) perform these functions in a way that maximizes the fitness of the population. Porter and Iglesias suggested the distortion theory framework, providing a complementary perspective on decision-making, regarding these three tasks as a single process. According to Porter and Iglesias , the distortion measure d : defines accurate sensing (task 1) by how heavily it penalizes small mistakes, and it quantifies the disadvantages of alternative decisions (task 2); the expected distortion describes how accurate sensing must be (task 1) and how much disadvantage the cell can afford in making a decision (task 2); the resulting optimal strategies fulfill task 3 by making choices with decisiveness proportional to the information available. The existence of suprasystem that determines the goal of the system of a cell (which can be either the tumor cell population for tumor growth or the healthy tissue surrounding the cell performing regulatory functions such as immune dynamics, angiogenesis, etc. or an unidentified entity) is reflected on the structure of the distortion measure. The condition of an "observer" distinguishable from the system that determines the goal of the system is a prerequisite for the definition of information in cybernetics by Wiener, which is founded on the issues of control and communication.
Q measures associations between variables, and not the direction of the transmission: 'This means that nothing is gained formally by distinguishing transmitters from receivers, therefore it goes beyond the Shannon framework of linear transmissions . An interaction is regularity, a pattern, a dependence present only in the whole set of events, but not in any subset. It is symmetric and undirected, so directionality no longer needs to be explained by, e.g. causality . Positive interaction implies synergy, and variables has been associated with the non-separability of a system in quantum physics  and with the origin of synergy in relationships between neurons . Q measures the amount of influence on the relationship between × and Y, resulting from the introduction of Z . It is the amount of information that is common to all variables but not present in any subset. The interaction information may provide the appropriate framework for the study of the Radiation induced Bystander effect (unirradiated cells exhibit irradiated effects as a result of signals received from nearby irradiated cells), a well-established consequence of exposure of living cells to radiation . Although cell to cell communications in normal and carcinogenic cells have been discussed extensively [49, 50], and it is believed that in general cell to cell regulatory signals are conducted by chemical and electrical signals (Gap Junctional Intercellular Communication (GJIC) or Distant Signaling Intercellular Communication (DSIC)) , it remains an assumption that these signals are propagated by a Brownian diffusive motion, because this yields to relatively satisfactory results in simulations of bystander effects . To understand the bystander synergistic effect in the case of radiation, we observe that if X is a cell state cell (target cell), and Y is some parameter that represents radiation, and Z is a cell state (neighbor cell), Q(X, Y, Z) can be understood as the difference between the decrease in entropy of Z achieved by the joint attribute XY (cell irradiation event) and the expected decrease in entropy with the assumption of independence between X and Y.
In this study a model of a cell decision mechanism is proposed, which captures certain observed characteristics of a cell behavior during photo-irradiation and pharmacological treatment (Type II PDT) using rate distortion theory to quantify the goals of a binary decision process (cell survival - cell death). The main components of the model are, the time dependent distribution of molecular stimuli, the distortion function (or measure), the conditional probability of the cell decision strategy, the cell survival probability, the expected distortion and the rate distortion function which quantifies a limit on how well the goals can be achieved given the stimulation. The results are independent of the biological mechanism by which the cell strategy is implemented and the Blahut Arimoto algorithm is used to derive optimal pathways. The model requires knowing the probability distribution of the stimuli as its input. For a variety of Lagrange multipliers, there is a corresponding variety of optimal pathways, but an approximation of the distortion function around which the pathway is optimized, is possible, based on algebraic properties of the algorithm (the distortion constraint) and numerical and experimental data . Intracellular molecular interactions can be studied with the purpose of extracting useful conclusions, by using computational methods. In this report we used a previous developed systems biology model that includes detailed molecular pathways induced by PDT treatment leading to cell death, which we coupled to a cell decision making algorithm that is based on the mutual information between cell death stimulation and cell response as the output of a bio molecular communication channel. This line of research can be relevant to future improvement and management of cancer treatment methodologies. The cell survival probability is modeled as the output of an optimization process of transmitting the death signal through a communication channel with a possible environmental and/or inherent distortion. Modeling results can be compared directly to experimental results that are based on the levels of measurable molecular concentrations and cell survival ratios, for optimization of the unknown parameters or/and for design of different in vitro studies of PDT. This modeling establishes a framework that may also be able to address questions such as why do cell types, although they share the same genome, they are in general stable entities represented by different observable cell fates with certain rules that govern their molecular dynamics and do not gradually change into other forms.
The author wants to thank biomedical physicist Dr. Xin-Hua Hu of ECU Physics Department for his direction and great help with the references and to acknowledge his assistance in the implementation of the Matlab codes. He also wants to thank medical physicist Carlos Austerlitz for his support and discussions during the preparation of this paper.
The publication costs for this article were funded by the corresponding author, Ioannis Gkigkitzis.
This article has been published as part of BMC Medical Genomics Volume 6 Supplement 3, 2013: Selected articles from the IEEE International Conference on Bioinformatics and Biomedicine 2012: Medical Genomics. The full contents of the supplement are available online at http://www.biomedcentral.com/bmcmedgenomics/supplements/6/S3.
- Sharman WM, Allen CM, van Lier JE: 'Role of activated oxygen species in photodynamic therapy'. Methods in enzymology. 2000, 319: 376-400.View ArticlePubMedGoogle Scholar
- Buytaert E, Callewaert G, Vandenheede JR, Agostinis P: 'Deficiency in apoptotic effectors Bax and Bak reveals an autophagic cell death pathway initiated by photodamage to the endoplasmic reticulum'. Autophagy. 2006, 2 (3): 238-240. 10.4161/auto.2730.View ArticlePubMedGoogle Scholar
- Albeck JG, Burke JM, Spencer SL, Lauffenburger DA, Sorger PK: 'Modeling a snap-action, variable-delay switch controlling extrinsic cell death'. PLoS biology. 2008, 6 (12): 2831-2852.View ArticlePubMedGoogle Scholar
- Tyson JJ, Baumann WT, Chen C, Verdugo A, Tavassoly I, Wang Y, Weiner LM, Clarke R: 'Dynamic modelling of oestrogen signalling and cell fate in breast cancer cells'. Nat Rev Cancer. 2011, 11 (7): 523-532. 10.1038/nrc3081.PubMed CentralView ArticlePubMedGoogle Scholar
- Gkigkitzis I, Feng Y, Yang C, Lu JQ, Hu XH: 'Modeling of Oxygen Transport and Cell Killing in Type-II Photodynamic Therapy'. Photochem Photobiol. 2012, 88 (4): 969-977. 10.1111/j.1751-1097.2012.01145.x.View ArticlePubMedGoogle Scholar
- Howland JL: 'Introduction to cell physiology: information and control'. 1968, Macmillan, 1968Google Scholar
- Gkigkitzis I, Hu X-H: 'A model of cellular decision making in photodynamic therapy of cancer'. Bioinformatics and Biomedicine (BIBM), 2012 IEEE International Conference on: 4-7 October 2012. 2012, 1-5. 10.1109/BIBM.2012.6392704.View ArticleGoogle Scholar
- Porter JR, Andrews BW, Iglesias PA: 'A framework for designing and analyzing binary decision-making strategies in cellular systems'. Integr Biol (Camb). 2012, 4 (3): 310-317. 10.1039/c2ib00114d.View ArticleGoogle Scholar
- Andrews BW, Iglesias PA: 'An information-theoretic characterization of the optimal gradient sensing response of cells'. PLoS computational biology. 2007, 3 (8): e153-10.1371/journal.pcbi.0030153.PubMed CentralView ArticlePubMedGoogle Scholar
- Wallace R: 'A rate distortion approach to protein symmetry'. Biosystems. 2010, 101 (2): 97-108. 10.1016/j.biosystems.2010.05.002.View ArticlePubMedGoogle Scholar
- Von Neumann J, Burks AW: 'Theory of self-reproducing automata'. 1966, University of Illinois Press, 1966Google Scholar
- Denton M: 'Evolution : a theory in crisis'. 1986, Adler & Adler, 1st U.S. edn. 1986Google Scholar
- Gougeon ML, Kroemer G: 'Charming to death: caspase-dependent or -independent?'. Cell death and differentiation. 2003, 10 (3): 390-392. 10.1038/sj.cdd.4401199.View ArticlePubMedGoogle Scholar
- Luo Y, Kessel D: 'Initiation of apoptosis versus necrosis by photodynamic therapy with chloroaluminum phthalocyanine'. Photochemistry and photobiology. 1997, 66 (4): 479-483. 10.1111/j.1751-1097.1997.tb03176.x.View ArticlePubMedGoogle Scholar
- Blahut RE: 'Computation of Channel Capacity and Rate-Distortion Functions'. Ieee T Inform Theory. 1972, 18 (4): 4601-View ArticleGoogle Scholar
- Berger T: 'Rate distortion theory; a mathematical basis for data compression'. 1971, Prentice-Hall, 1971Google Scholar
- Ledder G, Logan JD, Joern A: 'Dynamic energy budget models with size-dependent hazard rates'. J Math Biol. 2004, 48 (6): 605-622. 10.1007/s00285-003-0263-1.View ArticlePubMedGoogle Scholar
- Burns RC, Lawson AD: 'Quantized Probability Circuit Design Principles Applied to Linear Circuits'. Ieee T Reliab. 1964, R 13 (2): 16-&.View ArticleGoogle Scholar
- Lomsadze YM: 'Relativistically Invariant Formulation of Theory of Quantized Probability Amplitude Field'. Nucl Phys. 1962, 37 (1): 147-&.View ArticleGoogle Scholar
- Link G, ebrary Inc: 'One hundred years of Russell's paradox mathematics, logic, philosophy'. 'Book One hundred years of Russell's paradox mathematics, logic, philosophy'. 2004, Walter de Gruyter, ix: 662-(Ed.)^(Eds.)Google Scholar
- Van Heijenoort J: 'From Frege to Gödel; a source book in mathematical logic, 1879-1931. 1967, Harvard University Press, 1967Google Scholar
- Simmons K: 'Sets, classes and extensions: A singularity approach to Russell's paradox'. Philos Stud. 2000, 100 (2): 109-149. 10.1023/A:1018666804035.View ArticleGoogle Scholar
- von Neumann J: 'An axiomatization of set theory'. Aut Aut. 1997, 280-81: 107-123.Google Scholar
- Priest G: 'Beyond the limits of thought'. 2002, Clarendon/Oxford University Press, 2002View ArticleGoogle Scholar
- Chesson P: 'Predator-Prey Theory and Variability'. Annu Rev Ecol Syst. 1978, 9: 323-347. 10.1146/annurev.es.09.110178.001543.View ArticleGoogle Scholar
- Perthame B, Gauduchon M: 'Survival thresholds and mortality rates in adaptive dynamics: conciliating deterministic and stochastic simulations'. Math Med Biol. 2010, 27 (3): 195-210. 10.1093/imammb/dqp018.View ArticlePubMedGoogle Scholar
- Masutani K: 'Effects of survival thresholds upon one-dimensional dynamics of single-species populations'. B Math Biol. 1993, 55 (1): 1-13. 10.1007/BF02460292.View ArticleGoogle Scholar
- Graf S, Luschgy H, Pages G: 'The Local Quantization Behavior of Absolutely Continuous Probabilities'. Ann Probab. 2012, 40 (4): 1795-1828. 10.1214/11-AOP663.View ArticleGoogle Scholar
- Gray RM, Neuhoff DL: 'Quantization'. Ieee T Inform Theory. 1998, 44 (6): 2325-2383. 10.1109/18.720541.View ArticleGoogle Scholar
- Tyson JJ, Novak B: 'Functional Motifs in Biochemical Reaction Networks'. Annu Rev Phys Chem. 2010, 61: 219-240. 10.1146/annurev.physchem.012809.103457.PubMed CentralView ArticlePubMedGoogle Scholar
- Winther J: 'Photodynamic therapy effect in an intraocular retinoblastoma-like tumour assessed by an in vivo to in vitro colony forming assay'. British journal of cancer. 1989, 59 (6): 869-872. 10.1038/bjc.1989.184.PubMed CentralView ArticlePubMedGoogle Scholar
- Ben-Hur E, Rosenthal I, Leznoff CC: 'Recovery of Chinese hamster cells following photosensitization by zinc tetrahydroxyphthalocyanine'. Journal of photochemistry and photobiology B, Biology. 1988, 2 (2): 243-252. 10.1016/1011-1344(88)80007-1.View ArticlePubMedGoogle Scholar
- Selman SH, Kreimer-Birnbaum M, Chaudhuri K, Garbo GM, Seaman DA, Keck RW, Ben-Hur E, Rosenthal I: 'Photodynamic treatment of transplantable bladder tumors in rodents after pretreatment with chloroaluminum tetrasulfophthalocyanine'. The Journal of urology. 1986, 136 (1): 141-145.PubMedGoogle Scholar
- Langton CG, Center for Nonlinear Studies (Los Alamos National Laboratory), Santa Fe Institute (Santa Fe N.M.), Apple Computer Inc: 'Artificial life : the proceedings of an interdisciplinary workshop on the synthesis and simulation of living systems, held September, 1987 in Los Alamos, New Mexico'. 1989, Addison-Wesley Pub. Co., 1989Google Scholar
- Gunawardena J: 'Signals and systems: Towards a systems biology of signal transduction'. P Ieee. 2008, 96 (8): 1386-1397.View ArticleGoogle Scholar
- Whittemore ER, Loo DT, Watt JA, Cotman CW: 'A detailed analysis of hydrogen peroxide-induced cell death in primary neuronal culture'. Neuroscience. 1995, 67 (4): 921-932. 10.1016/0306-4522(95)00108-U.View ArticlePubMedGoogle Scholar
- Rong Y, Mack P: 'Apoptosis induced by hyperthermia in Dunn osteosarcoma cell line in vitro'. Int J Hyperthermia. 2000, 16 (1): 19-27. 10.1080/026567300285394.View ArticlePubMedGoogle Scholar
- Beere HM: 'Death versus survival: functional interaction between the apoptotic and stress-inducible heat shock protein pathways'. The Journal of clinical investigation. 2005, 115 (10): 2633-2639. 10.1172/JCI26471.PubMed CentralView ArticlePubMedGoogle Scholar
- Streffer C: 'Aspects of metabolic change after hyperthermia'. Recent results in cancer research Fortschritte der Krebsforschung Progres dans les recherches sur le cancer. 1988, 107: 7-16. 10.1007/978-3-642-83260-4_2.PubMedGoogle Scholar
- Falk MH, Issels RD: 'Hyperthermia in oncology'. Int J Hyperthermia. 2001, 17 (1): 1-18. 10.1080/02656730150201552.View ArticlePubMedGoogle Scholar
- Miller JG: 'Living systems'. 1995, University Press of Colorado, 1995Google Scholar
- Perkins TJ, Swain PS: 'Strategies for cellular decision-making'. Molecular systems biology. 2009, 5: 326-PubMed CentralView ArticlePubMedGoogle Scholar
- McGill WJ: 'Multivariate Information Transmission'. Psychometrika. 1954, 19 (2): 97-116. 10.1007/BF02289159.View ArticleGoogle Scholar
- Leydesdorff L: 'Interaction information: linear and nonlinear interpretations'. Int J Gen Syst. 2009, 38 (6): 681-685. 10.1080/03081070902993038.View ArticleGoogle Scholar
- Jakulin A, Bratko I: 'Analyzing attribute dependencies'. Lect Notes Artif Int. 2003, 2838: 229-240.Google Scholar
- Cerf NJ, Adami C: 'Entropic Bell inequalities'. Phys Rev A. 1997, 55 (5): 3371-3374. 10.1103/PhysRevA.55.3371.View ArticleGoogle Scholar
- Brenner N, Strong SP, Koberle R, Bialek W, van Steveninck RRD: 'Synergy in a neural code'. Neural Comput. 2000, 12 (7): 1531-1552. 10.1162/089976600300015259.View ArticlePubMedGoogle Scholar
- Lorimore SA, Wright EG: 'Radiation-induced genomic instability and bystander effects: related inflammatory-type responses to radiation-induced stress and injury? A review'. International journal of radiation biology. 2003, 79 (1): 15-25. 10.1080/713864986.View ArticlePubMedGoogle Scholar
- Trosko JE, Ruch RJ: 'Cell-cell communication in carcinogenesis'. Frontiers in bioscience : a journal and virtual library. 1998, 3: d208-236.Google Scholar
- Nikjoo H, Khvostunov IK: 'A theoretical approach to the role and critical issues associated with bystander effect in risk estimation'. Human & experimental toxicology. 2004, 23 (2): 81-86. 10.1191/0960327104ht422oa.View ArticleGoogle Scholar
This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.