Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                

Research Reports Mdh/Ima No. 2007-4, Issn 1404-4978: Date Key Words and Phrases

Download as pdf or txt
Download as pdf or txt
You are on page 1of 16

Research Reports MdH/IMa No.

2007-4, ISSN 1404-4978

EFFICIENT IMPLEMENTATION OF THE AI-REML ITERATION FOR VARIANCE COMPONENT QTL ANALYSIS

arXiv:0709.0625v2 [q-bio.QM] 11 Feb 2008

KATERYNA MISHCHENKO, SVERKER HOLMGREN, AND LARS RONNEG ARD Abstract. Regions in the genome that aect complex traits, quantitative trait loci (QTL), can be identied using statistical analysis of genetic and phenotypic data. When restricted maximum-likelihood (REML) models are used, the mapping procedure is normally computationally demanding. We develop a new ecient computational scheme for QTL mapping using variance component analysis and the AI-REML algorithm. The algorithm uses an exact or approximative low-rank representation of the identity-by-descent matrix, which combined with the Woodbury formula for matrix inversion results in that the computations in the AI-REML iteration body can be performed more eciently. For cases where an exact low-rank representation of the IBD matrix is available a-priori, the improved AI-REML algorithm normally runs almost twice as fast compared to the standard version. When an exact low-rank representation is not available, a truncated spectral decomposition is used to determine a low-rank approximation. We show that also in this case, the computational eciency of the AI-REML scheme can often be signicantly improved.

1. Introduction Traits that vary continuously are called quantitative. In general, such traits are affected by an interplay between multiple genetic factors and the environment. Most medically and economically important traits in humans, animals and plants are quantitative, and understanding the genetics behind them is of great importance.
Date : August 22, 2008. Key words and phrases. Quantitative Trait Mapping, restricted maximum-likelihood, average information matrix, identity-by-descent matrix, Woodbury formula, matrix inversion, truncated spectral decomposition, low-rank approximation, computational eciency, cpu time.

nneg Kateryna Mishchenko, Sverker Holmgren and Lars Ro ard The dissection of quantitative traits is generally performed by statistical analysis of

genetic and phenotypic data for experimental populations. Such analysis can reveal quantitative trait loci, QTL, in the genome that aect the trait. In the eld of QTL analysis (Lynch and Walsh, 1998), the use of variance component models are commonly used, see e.g. (Blangero et al., 2001). Using such models, the statistical analysis is performed using maximum-likelihood (ML) or restricted maximum-likelihood (REML) estimators which have the advantage that they do not use specic assumptions on the design or balance of the data. These types of schemes are often considered to be statistically ecient since they utilize all available data. On the other hand, the corresponding algorithms are also relatively computationally demanding since non-linear optimization problems must be solved using an iterative procedure. When searching for the most likely locations of the QTL, REML optimization problems must be solved for many genetic locations within an outer global optimization procedure. Furthermore, if the signicance of the results is established using an experimental procedure like parametric bootstrap (Davison and Hinkley, 1997), hundreds or thousands of QTL searches must be performed. This implies that the computational complexity for the numerical methods for solving the REML problems must be minimal for the QTL mapping to be viable. During the last two decades, specialized algorithms for variance component analysis in animal breeding settings have been developed and implemented in codes like ASREML, DMU and VCE, see (Druet and Ducrocq, 2006) and references therein. The algorithms used in these codes have in common that they use a specic Newton-type iteration, theaverage information restricted maximum likelihood algorithm, AI-REML. However, the structure of the problems in traditional animal breeding applications is dierent from that of QTL analysis problems, and the computational schemes should be examined and possibly modied before they are applied in a QTL analysis setting. In this short paper we perform this type of investigation and develop a new ecient computational scheme for QTL mapping using variance component analysis and the AI-REML algorithm. This type scheme should then be applied within an ecient and robust optimization method for maximizing the likelihood in the AI-REML method, and also within an ecient and

Efficient Implementation of the AI-REML Iteration

robust global optimization algorithm for the search for the outer optimization problem where the most likely position of the QTL is determined. 2. The AI-REML algorithm A general linear mixed model is given by (1) y = Xb + Rr + e,

where y is a vector of n observations, X is the n nf design matrix for nf xed eects, R is the n nr design matrix for nr random eects, b is the vector of nf unknown xed eects, r is the vector of nr unknown random eects, and e is a vector of n residuals. In the QTL analysis setting, we assume that the entries of e are identically and independently distributed and there is a single observation for each individual in
2 the pedigree. In this case, the covariance matrices are given by var (e) = Ie and 2 var (r ) = Aa , where A is referred to as the identity-by-descent (IBD) matrix. Using

these assumptions, we have that (2)


2 2 var (y ) V = a A + e I 1 A + 2 I.

Here, 1 0 and 2 > 0 and we assume that the phenotype follows a normal distribution with y MV N (Xb, V ). At least two dierent procedures can be used for computing estimates of b and 1,2 . In the standard codes mentioned above, the parameters are computed from the mixed-model equations (MME) (Lynch and Walsh, 1998) using a the inverse of A. To be able to use this approach, A has to be positive denite, or it must be modied so that this property holds. Also, the computations can be performed eciently if A is sparse. For the QTL analysis problems, the IBD matrix A is often only semi-denite and not necessarily sparse. Here the values of 1,2 are given by the solution of the minimization problem (3) Min L , s.t. 1 0 (4) 2 > 0

nneg Kateryna Mishchenko, Sverker Holmgren and Lars Ro ard

where L is the log-likelihood for the model (1), (5) L = 2ln(l) = C + ln(det(V )) + ln(det(X T V 1 X )) + yP T y,

and the projection matrix P is dened by (6) P = V 1 V 1 X (X T V 1 X ) 1 X T V 1 .

In the original AI-REML algorithm, the minimization problem is solved using the standard Newton scheme but where the Hessian is substituted by the average information matrix H AI , whose entries are given by (7)
AI Hi,j = yT P

V V P Py i j

, i, j = 1, 2.

The entries of the gradient of L are given by (8) L = tr i V P i yT P V Py i , i = 1, 2 .

3. Efficient implementation of AI-REML optimization The main result of this paper is an algorithm that allows for an ecient implementation of the iteration body in the AI-REML algorithm for variance component analysis in QTL mapping problems. In (Mishchenko et al.), we examine optimization schemes for the REML optimization and dierent approaches for introducing the constraints for 1,2 in the AI-REML scheme. For cases when the solution to the optimization problem is close to a constraint, dierent ad-hoc xes have earlier been attempted to ensure convergence for the Newton scheme. In (Mishchenko et al.), we show that by introducing an active-set formulation in the AI-REML optimization scheme, robustness and good convergence properties were achieved also for cases when the solution is close to a constraint. When computing the results below, we use the active-set AI-REML optimization scheme from (Mishchenko et al.). From the formulas in the previous section, it is clear that the most computationally demanding part of the AI-REML algorithm is the computation of the matrix P , and especially the explicit inversion of the matrix V = 1 A + 2 I , where A is a constant

Efficient Implementation of the AI-REML Iteration

semi-denite matrix and 1,2 are updated in each iteration. If V is regarded as a general matrix, i.e. no information about the structure of the problem is used, a standard algorithm based on Cholesky factorization with complete pivoting is the only alternative, and O (n3 ) arithmetic operations are required. If A is very sparse, the work can be signicantly reduced by using a sparse Cholesky factorization, possibly combined with a reordering of the equations. This can be compared to the approach taken e.g. in (Bates, (2004)), where a sparse factorization is used for A when solving the mixedmodel equations. However, as remarked earlier A is only semi-denite and not very sparse for the QTL analysis problems and this approach is not an option here. The IBD matrix A is a function of the location in the genome where the REML model is applied. The key observation leading to a more ecient algorithm for inverting V is that the rank of A at a location with complete genetic information only depends on the size of the base generation. At such locations, A will be a rank-k matrix where k n in experimental pedigrees with small base generations. If the genetic information is not fully complete, A can still be approximated by a low-rank matrix. The error in such an approximation can be made small when the genetic distance to a location with complete data is small. Let a symmetric rank-k representation of A be given by

(9)

A = ZZ T ,

where Z is an n k matrix. We now exploit this type of representation to compute V 1 . Case 1: An exact low-rank representation of A is available a-priori. In general, the inverse of a matrix plus a low-rank update can be computed using the Woodbury formula, see e.g. (Golub and C.Van Loan, (1996)),

(10)

T 1 T 1 T B 1 = (C + S1 S2 ) = C 1 C 1 S 1 ( I + S2 C S1 ) 1 S2 C 1

nneg Kateryna Mishchenko, Sverker Holmgren and Lars Ro ard

Applying (10) with C = I2 , S1 = Z and S2 = 1 Z , we get the following formula for computing V 1 , (11)
1 1 V 1 = I2 2 Z (I + Z T Z )1 Z T ,

1 where = 1 2 . The k k matrix Z T Z can be computed once, before the Newton

iteration is started. Also, the matrix (I + Z T Z ) is only of size k k , and its inverse can be computed using a standard factorization method in O (k 3 ) arithmetic operations. Case 2: No low-rank representation of A is given a-priori. If only the matrix A is given, a low-rank representation/approximation can be computed once before the Newton iteration is started. Then, the Woodbury formula is again used to get an ecient computational algorithm for the iterations. Computing a low-rank approximation of a matrix is a problem with a long history (Eckart and Young, 1936), and the standard tool is the truncated singular value decomposition, see e.g. (Golub and C.Van Loan, (1996); Press et al., 1997), which computes the optimal approximation in Frobenius norm. In our case, A is symmetric and positive semi-denite, and the SVD is equivalent to the standard spectral decomposition. Also, the spectral decomposition can be written as A = W W T = (W )(W )T , where = diag(i ) , i = 1, . . . , n. Hence, we compute a low-rank representation of A using a truncated spectral decomposition, (12) A = (Wt t )(Wt t )T ,

where t is a diagonal matrix of k eigenvalues and Wt is a rectangular matrix of size n k containing the corresponding orthogonal eigenvectors. Here, t and Wt are obtained by eliminating the n k smallest eigenvalues and the corresponding eigenvectors from the standard spectral decomposition. Such idea is used in Genetic Principal Component Analysis, see e.g. (Kirpatrick and Meyer, 2004). Again using the Woodbury formula (10), now with C = I2 , S1 = Wt t and 1 according to, S2 = 1 Wt t , we compute V (13) 1 = I 1 1 (Wt V 2 2 t )(I + t )1 (Wt t )T .

Efficient Implementation of the AI-REML Iteration Or in equivalent form, (14) 1 = I 1 1 Wt (1 + I )1 W T . V 2 2 t t

A similar approach for matrix inversion has been used for problems in Gaussian process regression with applications in e.g. machine learning, see (Schwaighofer and Tresp, 2002; Rasmussen and Williams, 2005). However, as far as we know, an approach based on a truncated spectral decomposition and formula (13) has not been been pursued for Newton iterations for REML-problems before. When performing the computation described by (14), the matrix Wt is constant and
1 can be precomputed before the iteration starts. The matrix ( t + I ) is diagonal, so

its inverse is trivially computable. The factorization (12) is a rank-k representation of A. This is exact if A has at least 1 = V 1 . For a general IBD matrix A, more than k n k zero eigenvalues, and then V 1 is an approximation of V 1 . eigenvalues will normally be non-zero, resulting in that V

The accuracy of this approximation depends on the quality of the truncated factorization (12) for A. For the QTL analysis problems, we need to chose the rank k such that the accuracy of the solution to the REML problem, i.e. the variance components 1,2 , is sucient. Assume that the eigenvalues are ordered such that 1 2 n . Several criteria for choosing k can be considered. One option is to simply set k to a predetermined value, either motivated by the known value of the rank at a fully informative position in the genome or determined as a given percentage of n. Such a choice has the advantage that it is possible to pre-determine the computational work needed for computation of 1 in (14). Another easily applied criterion for determining k is given by V (15) k = k ( ) = {max i : i 1 },

where the threshold is for example chosen as = 0.001. This means that all eigenvalues of the matrix A that are smaller than 0.0011 are neglected and the corresponding

nneg Kateryna Mishchenko, Sverker Holmgren and Lars Ro ard

columns of W in the spectral decomposition are deleted. In the numerical experiments presented in the next section, we use the truncation criterion (15). For computing the truncated spectral decomposition, we currently use a standard eigenvalue solver, compute all eigenvalues and eigenvectors, and then ignore the data connected to eigenvalues smaller than given by the truncation criterion. For problems with large data matrices where it is known that only a small number of eigenvalues will be large, an alternative can be to use e.g. the Lanczos iteration for computing these. Potentially, this can reduce the arithmetic work for computing the truncated decomposition before the iteration starts. Once V 1 has been computed using either (14) or (11), formula (6) is used to compute P , and the entries of the Hessian H AI and the gradient of L are computed using the formulas (7) and (8). The matrices involved have common terms which are computed once at the beginning of the current iteration to reduce the total complexity.

4. Numerical Results In this section we present numerical experiment computed using experimental IBD matrices of size 767 767 drawn from a chicken population (Kerje et al., 2003). For these rst pilot investigations of the computational performance we have chosen to implement the algorithms in Matlab, and we present execution time measurements for the Matlab scripts. At a later stage, we will implement the most promising algorithms also in C, which will reduce the execution time further. In this case, we expect that the relative gain from using the new algorithms will be larger than indicated below. Matlabs native implementation of standard inversion is presumably more ecient than our Matlab scripts for the new algorithms, which are interpreted at runtime. Case 1: We rst present numerical results for the case when the IBD matrices A have exact low-rank representations given a-priori. Let TC be the measured computational time for the computation of V 1 within the AI-REML scheme using standard inversion, and let TW be the corresponding time when using the low-rank representation and

Efficient Implementation of the AI-REML Iteration

formula (11). In Figure 1 we show TC /TW , i.e. the speedup for the computations of V 1 , as a function of the rank k of the matrices A.
The speedup achieved for computing V1 by using the lowrank representation of A 25

20

15 T /T
W C

10

100

200 300 400 500 Rank k of the lowrank representation of A

600

700

Figure 1. The speedup achieved for the computations of V 1 in the AI-REML scheme achieved by using the scheme based on (11) instead of standard inversion From Figure 1, it is clear that for IBD matrices of size 767 767, the implementation using the low-rank representation is faster if the rank k is smaller than approximately 350. Also, for k < 50, the speedup is more than one order of magnitude. In Table 1 we present the total CPU-times for the AI-REML scheme and the CPUtime for the computations of V 1 within this scheme for dierent values of k and number
tot of iterations equal to 9. The total CPU-times for the AI-REML are denoted by TC and tot TW , while the CPU-times for the computation of V 1 are denoted by TC and TW .

From Table 1 it is clear that for the problems studied here, computing V 1 using standard inversion accounts for approximately half of the computational time in the AI-REML scheme. For small values of k , the time needed for inverting V 1 using the low-rank representation can eectively be ignored, and the improved algorithm runs approximately twice as fast as the original scheme using standard inversion.

10

nneg Kateryna Mishchenko, Sverker Holmgren and Lars Ro ard

tot TC

TC

tot TW

TW

700 10.28 5.45 20.3 15.5 600 10.28 5.45 15.7 10.9 500 10.28 5.45 12.6 7.81 400 10.28 5.45 10.4 5.55 300 10.28 5.45 8.20 3.37 200 10.28 5.45 6.62 1.86 150 10.28 5.45 6.25 1.42 100 10.28 5.45 5.69 0.86 50 25 10.28 5.45 5.23 0.41 10.28 5.45 5.08 0.25

Table 1. Total CPU-time for the AI-REML computations and CPUtime for computing V 1 for the schemes using standard inversion and a low-rank representation

Case 2: We now turn to case 2. Here a low-rank representation of A is not available a-priori, and a truncated eigenvalue decomposition must be computed before the AI-REML iteration starts. We present results from computations performed using 18 dierent IBD matrices corresponding to genetic locations where a low-rank representation is not given. The rst set of results shows the computational time TC,T W for computing V 1 within the AI-REML scheme. In this case we compare standard inversion (C) to the scheme given by (13), based on truncated spectral decomposition (TW). In Table 2, we show the CPU-time and the number of AI-REML iterations for the two schemes. For the scheme based on the truncated spectral decomposition, the table also shows the rank k as determined by the truncation criterion (15) and the relative errors (in %) for the variance components (e1,2 ) arising from using a low-rank approximation of A. For the matrices A11 and A12 , the variance component 1 is zero, and it is not possible to compute the relative error. For these matrices, the absolute errors in

Efficient Implementation of the AI-REML Iteration

11

C A TC A1 A2 A3 A4 A5 A6 A7 A8 A9 4.18 4.16 3.50 3.48 2.81 4.89 2.88 4.93 8.36 # it 7 7 6 6 5 8 5 8 13 23 4 4 9 5 5 5 5 7 TT W 2.20 1.12 1.25 1.87 1.81 2.60 1.66 1.22 0.671 1.22 0.140 0.344 0.467 k 294 166 169 280 291 294 286 163 54 50 46 71 57 = 0.001

TW =0.005 TT W k # it e1 [%] e2 [%] 7 6 8 15 11 6 5 7 13 23 4 5 9 7 5 8 6 6 6.7 7.3 12.7 40.8 30.4 8.1 5.9 3.9 4.5 0.05 0.02 16.9 3.1 22.2 6.3 8.0 4.7 1.8 3.0 7.5 5.4 3.1 2.1 0.64 0.09 0.05 0 0 0.02 1.6 1.1 2.0 3.2 3.7

#it e1 [%] e2 [%] 7 7 7 6 6 8 5 8 12 23 4 5 9 5 5 5 5 8 0.82 0.03 0.11 1.14 0.94 0.64 0.84 0.44 0.04 0.07 0.004 0.53 0.36 0.19 0.96 0.69 1.67 0.377 1.19 1.83 1.56 0.76 0.53 0.18 0.03 0.02 0 0 0 0.26 0.41 0.39 0.95 0.91

0.531 82 0.203 37 0.307 35 1.05 71

0.920 82 0.438 81 0.313 70 0.314 43 0.328 15 0.527 10 0.093 13 0.108 14 0.173 12 0.327 51 0.250 51 0.327 44 0.311 60 0.485 86

A10 15.3 A11 1.98 A12 2.00 A13 5,45 A14 2.86 A15 2.77 A16 2.83 A17 2.80 A18 4.23

0.873 192 0.884 198 0.749 173 1.28 2.62 270 303

Table 2. CPU-time for computing V 1 within the AI-REML iteration, and errors arising from exploiting a low-rank approximation of A 1 is very small for both values of . From the results in Table 2, it is clear that the accuracy of the variance components does not only depend on the value of the parameter , but also on the properties of the matrix A. When an approximation of V 1 is employed in the computations, the optimization landscape is aected in dierent ways for dierent IBD matrices A, resulting in dierent behavior for the optimization

12

nneg Kateryna Mishchenko, Sverker Holmgren and Lars Ro ard

scheme and dierent errors in the location of the optima. However, for the problems studied here, choosing = 0.001 results in that the variance components are determined with sucient accuracy for all the IBD matrices examined. In the next set of experiments we study the speedup and error introduced by using an approximate inverse of V in some more detail for the two IBD matrices A1 and A13 . As can be seen from Table 2, A1 is harder to approximate with a low rank matrix than A13 . In Tables 3 and 4, we vary the value of the parameter in the truncation criterion (15) and show the resulting rank k of the approximation of A, the number of iterations in the AI-REML optimization scheme, the speedup TC /TT W , and the relative errors in the variance components. 107 106 105 104 5 104 103 5 103 k 699 662 556 294 82 29 # it 7 7 7 7 7 36 no conv. TC /TT W e1 [%] e2 [%] 0.46 0.51 0.71 1.97 7.68 3.28 0.000 0.002 0.070 0.82 6.75 95.2 0.000 0.001 0.054 1.67 4.69 6.70 -

Table 3. Results for the IBD matrix A1 for dierent values of the truncation parameter

In Figure 2 we nally present total timings for solving the AI-REML problems using the new scheme exploiting a truncated spectral decomposition, including the time needed for computing the truncated factorization prior to the AI-REML iterations. We also compare these timings to the corresponding results for the standard scheme using direct inversion. In the gure, the timings for all matrices A1 A18 are shown as a function of the number of AI-REML iterations required. When several matrices result in the same number of iterations, the cpu time for each matrix and the average result are shown.

Efficient Implementation of the AI-REML Iteration

13

107 106 105 104 5 104 103 5 103 102 5 102

k 134 134 123 57 12 9 7 5 -

# it 9 9 9 9 9 9 9 9 no conv.

TC /TT W e1 [%] e2 [%] 6.41 6.50 7.12 15.0 34.6 37.3 44.7 55.7 0.000 0.000 0.000 0.004 0.009 0.024 0.109 2.3 0.000 0.000 0.000 0.006 0.025 0.025 0.027 0.046 -

Table 4. Results for the IBD matrix A13 for dierent values of the truncation parameter
Average Total CPU time 30 INV TSDWOOD average TSDWOOD 25

20 Total CPU time (sec)

15

10

10

12 14 16 Number of iterations

18

20

22

24

Figure 2. Total CPU-times for solving the AI-REML problems using the scheme based on standard inversion of V and the new scheme using a truncated spectral decomposition

14

nneg Kateryna Mishchenko, Sverker Holmgren and Lars Ro ard Theoretically, the average cpu time for the scheme exploiting the truncated spectral

decomposition should be a straight line. Deviations are due to the inconsistence of time measurement in Matlab. From Figure 2, we see that the standard scheme using direct inversion is faster when number of iterations is small, i.e. less than 9 iterations. The reason for this is that the spectral decomposition of the matrix A is relatively costly, and it must be amortized over a number of iterations before the faster iterations begin to pay of. When the number of iterations required is signicantly larger than 9, which is often the case for real-world problems, the new algorithm is signicantly faster also when no low-rank representations of the IBD matrices are not available a priori. 5. Conclusions In this paper we present a family of algorithms that allow for an ecient implementation of the iteration body in the AI-REML algorithm for variance component analysis in QTL mapping problems. Combined with the improved optimization scheme in (Mishchenko et al.), the new algorithms form a basis for an ecient and robust AIREML scheme for evaluating variance component QTL models. The most costly operation in the AI-REML iteration body is the explicit inversion of the matrix V = 1 A + 2 I , where the IBD matrix A is constant and positive semidenite, and 1 0 and 2 > 0 are updated in each iteration. The key information enabling the introduction of improved algorithms is that the rank of A at a location in the genome with complete genetic information only depends on the size of the base generation. At such locations, A will be a rank-k matrix where k n, and by exploiting the Woodbury formula the inverse of V can be computed more eciently than by using a standard algorithm based on Cholesky factorization. If the genetic information is not fully complete, a general IBD matrix A can still be approximated by a low-rank matrix and the error in such an approximation can be made small when the genetic distance to a location with complete data is small. More importantly, there might be a possibility for setting up a low-rank representation of the matrix A also at genetic

Efficient Implementation of the AI-REML Iteration

15

locations where the information is not complete. This is a topic of current investigation (R onneg ard and Carlborg, 2007). We present results for IBD matrices A from a real data set for two dierent settings; Firstly, we show that if a low-rank representation of A is available a-priori, the inversion of V using the new algorithms is performed faster than using a standard algorithm if the rank k is smaller than approximately 350. Also, for k < 50, the speedup is more than one order of magnitude. Then we also show that, even if a low-rank representation is not directly available and a low-rank approximation of A needs to be computed before the AI-REML iterations, signicant speedup of the variance component model computations can still be achieved. For QTL mapping problems, the eciency of our new method will increase when the ratio between the total pedigree size and base generation size increases, the density and informativeness of markers increases. Hence, the relative ecacy of the method will continuously increase in the future with deeper pedigrees and more markers. Also, we are currently developing a scheme for directly constructing a low-rank representation of the IBD matrix A also between markers, which will result in that the eigenvalue factorization prior to the AI-REML iterations is not needed any more. References
Bates, D. (2004). Sparse matrix representations of linear mixed models. Technical Report. http://www.stat.wisc.edu/ bates/reports/MixedEects.pdf. Blangero, J., J.T. Williams, and L. Almasy. 2001. Variance component methods for detecting complex trait loci, Advances in Genetics 42, 151181. Davison, A. C. and D. V. Hinkley. 1997. Bootstrap methods and their application., Cambridge University Press, U.K. Druet, T. and V. Ducrocq. 2006. Innovations in software packages in quantitative genetics. World Congress on Genetics Applied to Livestock Production,Belo Horizonte, Brazil. Eckart, C. and G. Young. 1936. The approximation of one matrix by another of lower rank, Psychometrica 1, no. 3, 211218. Golub, G. and C.Van Loan. (1996). Matrix Computations, Third edition, The Johns Hopkins University Press.

16

nneg Kateryna Mishchenko, Sverker Holmgren and Lars Ro ard

Carlborg, L. Jacobsson, K. Schutz, C. Hartmann, P. Jensen, and L. Andersson. 2003. Kerje, S., O. The twofold dierence in adult size between the red junglefowl and white leghorn chickens is largely explained by a limited number of qtls, Animal Genetics 34, 264274. Kirpatrick, M. and K. Meyer. 2004. Direct estimation of genetic principal components: Simplied analysis of complex phenotypes, Genetics 168, 22953206. Lynch, M. and B. Walsh. 1998. Genetics and analysis of quantitative traits, Sinauer Associates, Inc. Mishchenko, K., S. Holmgren, and L. R onneg ard. Newton-type methods for reml estimation in genetic analysis of quantitative traits. http://arxiv.org/abs/0711.2619. Press, W. H., S. A. Teukolsky, W. T. Vetterling, and B. P. Flannery. 1997. Numerical recipes in c. the art of scientic computing, Second edition, Cambridge University Press. Rasmussen, C.E. and C. Williams. 2005. Gaussian processes for machine learning, MIT Press. Carlborg. 2007. Separation of base allele and sampling term eects gives new R onneg ard, L. and O. insights in variance component qtl analysis, BMC Genetics 8, no. 1. Schwaighofer, A. and V. Tresp. 2002. Transductive and inductive methods for approximate gaussian process regression, NIPS.

lardalen University, Box 883, SE-721 23 Department of Mathematics and Physics, Ma ster Va as, Sweden E-mail address : kateryna.mishchenko@mdh.se Division of Scientific Computing, Department of Information Technology, Uppsala University, Sweden E-mail address : sverker@it.uu.se Linnus Center for Bioinformatics, Uppsala University, Sweden E-mail address : lars.ronnegard@lcb.uu.se

You might also like