Bio/statistics Handout 5  How matrix products arise

 

·        Genetics:  Suppose that a given stretch of DNA coding for cellular product is very mutable, so that there are some number, N, of possible sequences that can appear in any given individual (this is called ‘polymorphism’) in the population.   To elaborate, a strand of DNA is a molecule that appears as a string of small, standard molecules that are bound end to end.  Each of these standard building blocks can be one of four, labeled C, G, A and T.  The order in which they appear along the strand determines any resulting cellular product that the given part of the DNA molecule might produce.  For example, AAGCTA  may code for a different product than GCTTAA. 

As it turns out, there are stretches of DNA where the code can be changed without damage to the individual.  What with inheriting genes from both parents, random mutations over the generations can then result in a population where the codes on the given stretch of DNA vary from individual to individual.  In this situation, the gene is called ‘polymorphic’.  Suppose that there are N different possible codes for a given stretch.  For example, if one is looking at just one particular site along a particular DNA strand, there could be at most N = 4 possibilities at that site, namely C, G, A or T.  Looking at two sites gives N = 4´4 = 16 possibilities.

Lets label the possible sequences by integers starting from 1 and going to N.  At any given generation t, let pj(t) denote the frequencey of the j’th sequence appearing in the population at that generation.  Given that the j’th sequence appears in an individual, there is some probability, P(i|j), that the i’th sequence appears in their heir.  Here, let us assume that the population is constant and that these P(i|j) are independent of the generation number.  (This is all very simplistic except in very special circumstances.)   In this situation, we can say the following:  The probability of any given sequence, say i, appearing in generation t+1 can be written as a sum:

 

pi(t+1) = P(i|1) p1(t) + P(i|2) p2(t) + · · · + P(i|N) pN(t)

(5.1)

This is to say that the probability of sequence i appearing in an heir is equal to the probability that sequence 1 appears in the parent times the probability of a mutation that changes sequence 1 to sequence i, plus the probability that sequence 2 appears in the parent times the probability of a mutation that changes sequence 2 to sequence i, etc.

      Now,we can write this using our matrix formalism by thinking of the numbers {pj(t)}1≤j≤N as defining a column vector, (t), in RN, and likewise the numbers {pi(t+1)}1≤i≤N as defining a second column vector, (t+1) in RN.  If I introduce the N´N matrix A whose entry in the j’th column and i’th row is P(i|j), then (5.1) says in very cryptic shorthand:

 

(t+1) = A(t) .

(5.2)

Now, we can sample the population at time t = T = now, and thus determine (T), or at least the proxy that takes pi(T) to be the percent of people in the population today that have sequence i.  One very interesting question is to determine (T´) at some point far in the past, thus T´ << T.  For example, if we find (T´) such that all pi(T´) are zero but a very few, this then indicates that the population at time T´ was extremely homogeneous, and thus presumably very small. 

To determine (T´), we use (5.2) in an iterated form:

 

(T) = A(T-1) = AA(T-2)) = AAA(T-3) = ··· = A···A(T´) ,

(5.3)

where the final term has T-T´ copies of A multiplying one after the other. 

On a similar vein, we can use (5.2) to predict the distribution of the sequences in the population at any time T´ > T by iterating it to read

 

(T´) = A···A(T)

(5.4)

      Here, the multiplication is by T´-T successive copies of the matrix A.

By the way, here is a bit of notation:  The sequence {(t)}t=0,1,… of vectors of probabilities is an example of a Markov chain.  In general, a Markov chain is a sequence of probabilities, {P(0), P(1), P(2), …, } where the N’th probability P(N) depends only on the probabilities with numbers that are less than N.

 

·        How bacteria find food:  If you put certain sorts of bacteria on one side of a petri dish and put some sugar some distance away on the other, the bacteria will migrate towards the sugar.  Apparently, the sugar diffuses in the petri dish, so that there is slight concentrations everywhere, with most of the concentration in the original spot.  The bacteria are sensitive to the different levels at their front and rear ends and tend to move in the direction of the greater level.  At the expense of borrowing from Math 21a, the bacteria sense the direction of the gradient of the sugar concentration.  Even so, their movement from low to high concentration has a certain randomness to it; at any given step there is some probability that they will move in the wrong direction. 

What follows is a simplistic model for this:  Imagine bacteria moving in steps along the x-axis along the segment that stretches from x = 1 to x = N.  Here, N is some large integer.  Suppose the sugar placed initially where x = N, and the bacteria is placed initially at x = 1.  Our model also supposes that the bacteria moves one unit per second, with probability q Î (0, 1) of moving to the right at any given step, and probability 1-q of moving to the left unless it is at the end of the interval.  If it is at the x = 1 end, it moves to the right with probability q and stays put with probability 1-q.  If it is at the x = N end, it stays put with probability q and moves to the left with probability 1-q.  In our model, q is independent of position, x Î {1, . . . , N}.  Such would be roughly the case in the real petri dish where the sugar concentration gradient is relatively constant.  We should take q >  in the case that the bacteria is attracted to the sugar. 

For each time step t, and j Î {1, . . . , N}, let pj(t) denote the probability that the bacteria is at position j at time t.  For example, p1(0) = 1 and pj(0) = 0 for j > 1.  Our model then says that the probability of finding the bacteria at position j ≠ 1 of N at time step t > 0 is equal to the probability that it is at position j-1 at time t-1 times the probability that it moves one step to the right, plus the probability that it is at position j+1 at time t-1 and moves one step to the left.  Thus,

 

pj(t+1) = q pj-1(t) + (1-q) pj+1(t)   when  2 ≤ j ≤ N-1

(5.5)

      By the same token,

 

p1(t+1) = (1-q) p1(t) + (1-q) p2(t)   and    pN(t+1) = q pN-1(t) + q pN(t) .

(5.6)

Let us introduce the vector (t) in RN whose j’th component is pj(t).  Then (5.5) and (5.6) assert that (t+1) is obtained from (t) by the action of a linear transformation: (t+1) = A(t), where A is the N´N matrix whose only non-zero entries are:

 

A11 = 1-q,   Ajj-1 = q for  2 ≤ j ≤ N,    Ajj+1 = 1-q  for 1 ≤ j ≤ N-1,   and   ANN = q

 (5.7)

     For example, in the case that N = 4, the matrix A is

 

(5.8)

In any case, we can iterate the equation (t+1) = A(t) to find that

 

(t) = A···A(0)

(5.9)

where A···A signifies t copies of A multiplied one after the other.  By the way, a common shorthand for some n copies of any given matrix, M, successively multiplying one after the other is Mn. 

 

·        Growth of nerves in a developing embryo:  Here is a very topical question in the study of development:  Nerve cells connect muscles and organs to the brain.  When an organism develops, how do its nerves know where to connect?  A given nerve cell stretches out extremely long and thin ‘appendages’ called axons.  Any given axon may end abutting a muscle cell, or another nerve cell in a chain of nerves that ends in the brain.  How do the axons ‘know’ where they are supposed to attach?  A simple model proposes that the tip of the growing axon in an embryo is guided by chemical gradients in much the same fashion as the bacteria in the previous discussion is guided to the sugar.

 

·        Enzyme dynamics:  An enzyme is a protein molecule that facilitates a chemical reaction that would take a long time to occur otherwise.  Most biological reactions are facilitated by enzymes.  One typical mode of operation is for the enzyme to facilitate a reaction between molecules of type a and molecules of type b to produce a new molecule of type g.  It can do this if it simulaneously attracts the a and b molecules.  In doing so, the enzyme holds the two sorts near each other for enough time that they can bind together to form the more complicated g molecule.  If this g molecule does noot have a strong attraction to the enzyme, it will break away and so free the enzyme to attract another a and another b to make more g. 

Now, it is often the case that the enzyme only works efficently if it is folded in the right conformation.  Folded in the wrong way, the parts of the molecule that attract either a or b molecules might find themselves covered by parts that are indifferent to a or b.  Due to random collisions and other effects, any given molecule, and thus our enzyme, will change its fold configuration.  Suppose that there is some probability, q, that the enzyme is folded correctly to attract a and b in any given unit of time, and thus probability (1-q) that it is not.   Suppose in addition, that when folded correctly, the enzyme makes 2 units of g per unit of time.  Meanwhile, suppose that the freely diffusing g falls apart or is degraded by other enzymes at a rate of 1 unit per unit time. 

For any integer j, let pj(t) denote the probability of finding level j of g after some t units of time. Then the scenario just outlined finds

 

pj(t+1) = q pj-1(t) + (1-q) pj+1(t)  when  j ≥ 1  and   p0(t+1) = (1-q)(p1(t) + p0(t))    

(5.10)

       as long as a and b are well supplied.

 

Exercises:

 

Exercises 1-3 concern the example above where the position of the bacteria at any given time t is labeled by an integer, x(t), in the set {1, . . . , N}.

 

1.      Suppose T is a positive integer. 

a)  What is the sample space for the collection of positions, {x(0), x(1), . . . , x(T)}, of the bacteria at the times t = 0, 1, …, T.

b)  Fix some j Î {1, …, N} and t Î {0, 1, …, T-1} and let A denote the event that x(t+1) = j.  For each i Î {1, …, N}, let Aj denote the event that x(t) = i.  Explain why the equation in (5.5) has the form P(A) = P(A|A­1)P(A1) + ··· + P(A|An)P(AN).

 

2.      Recall that events A and B are said to be independent when P(AÇB) = P(A)P(B).

a)  Explain why this definition is equivalent to the assertion that the conditional probability of A happening given B is equal to the probability of A happening.

b)  Suppose that k Î {2, . . . , N-1} and that t ≥ 2.  Derive a relation between the respective probabilities that x(t) = k-1 and x(t) = k+1 so as to insure that the event that x(t+1) = k is independent of the event that x(t) = k-1.

 

3.      Suppose now that q+ and q- are non-negative numbers whose sum is less than 1.  Change the bacteria model so that if the bacteria is at position k Î {2, . . . , N-1} at time t, then it moves to position k+1 with probability q+, to position k-1 with probability q- and stays put with probability 1-q+-q-.   If the bacteria is at position 1 at time t, then it stays put with probability 1-q+ and moves to position 2 with probability q+.  If the bacteria is at position N at time t, then in stays put with probability 1-q- and moves to position N-1 with probability q-

a)  Write down the analog of (5.5) for this model. 

b)  Write down the analog of the matrix in (5.8) for the N = 4 case.

c)   Write down the probability that the bacteria is at position N at t = N-1 given that the bacteria starts at position 1 at time 0.

 

4.      Write down a version of (5.10) that would hold in the case that the enzyme when folded correctly produces some L = 1, or L > 2 units of g per unit time.