To save content items to your account,
please confirm that you agree to abide by our usage policies.
If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account.
Find out more about saving content to .
To save content items to your Kindle, first ensure no-reply@cambridge.org
is added to your Approved Personal Document E-mail List under your Personal Document Settings
on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part
of your Kindle email address below.
Find out more about saving to your Kindle.
Note you can select to save to either the @free.kindle.com or @kindle.com variations.
‘@free.kindle.com’ emails are free but can only be saved to your device when it is connected to wi-fi.
‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.
Sometimes nonparametric bootstrap inference can be performed by using a saddlepoint inversion to replace the resampling that is normally required. This chapter considers this general concept and pursues the extent to which the general procedure can implemented.
The main set of examples have been taken from Butler and Bronson (2002) and Bronson (2001). They involve bootstrap inference for first passage characteristics and time to event distributions in the semi-Markov processes discussed in chapter 13. Saddlepoint approximations of nonparametric bootstrap estimates are given for the mean and standard deviation of first passage and as well as for the density, survival, and hazard functions associated with the passage events. Thus saddlepoint approximations become estimates for these moments and functions and represent approximations to the estimates that would normally be provided were a large amount of resampling to be implemented with “single” bootstrap resampling. In performing these computations however, no resampling has been used to compute the estimates, only saddlepoint inversions. The saddlepoint estimates utilize data from the process through an empirical estimate T (s) of the unknown system transmittance matrix T (s). Empirical transmittance T (s) is used in place of T (s) when implementing the saddlepoint inversions.
In addition to bootstrap estimates, confidence intervals for the moments and confidence bands for the functions may be computed by using the double bootstrap. Often the resampling demands inherent in the double bootstrap place its implementation beyond the range of practical computing (Booth and Presnell, 1998).
Saddlepoint methods are applied to many of the commonly used test statistics MANOVA. The intent here is to highlight the usefulness of saddlepoint procedures in providing simple and accurate probability computations in classical multivariate normal theory models. Power curves and p-values are computed for tests in MANOVA and for tests of covariance. Convenient formulae are given which facilitate the practical implementation of the methods.
The null distributions of many multivariate tests are easily computed by using their Mellin transforms which admit saddlepoint approximations leading to very accurate pvalues. The first section concentrates on the four important tests of MANOVA. Very accurate p-value computations are suggested for (i) the Wilks likelihood ratio for MANOVA, (ii) the Bartlett–Nanda–Pillai trace statistic, (iii) Roy's largest eigenvalue test, and (iv) the Lawley-Hotelling trace statistic. Saddlepoint methods for Wilks' likelihood ratio test were introduced in Srivastava and Yau (1989) and Butler et al. (1992a); approximations for (ii) were introduced in Butler et al. (1992b) and were also discussed in section 10.4; p-values for Roy's test do not use saddlepoint methods and are based on the development in Butler and Paige (2007) who extend the results of Gupta and Richards (1985); and p-values for Lawley–Hotelling trace are based on numerical inversion of its Laplace transform in work to be published by Butler and Paige.
The second section considers tests for covariance patterns. Saddlepoint approximations for p-values are suggested for the following tests: the likelihood ratio tests for (ii) block independence, (iii) sphericity, (iv) an intraclass correlation structure; and (v) the Bartlett–Box test for equal covariances. These applications were developed in Butler et al. (1993), and Booth et al. (1995).
Exponential families provide the context for many practical applications of saddlepoint methods including those using generalized linear models with canonical links. More traditional exponential family applications are considered in this chapter and some generalized linear models are presented in chapter 6. All these settings have canonical sufficient statistics whose joint distribution is conveniently approximated using saddlepoint methods. Such approximate distributions become especially important in this context because statistical methods are almost always based on these sufficient statistics. Furthermore, these distributions are most often intractable by other means when considered to the degree of accuracy that may be achieved when using saddlepoint approximations. Thus, in terms of accuracy, both in this context and in the many others previously discussed, saddlepoint methods are largely unrivaled by all other analytical approximations.
The subclass of regular exponential families of distributions was introduced in section 2.4.5 by “tilting” a distribution whose MGF converges in the “regular” sense. Tilting combined with Edgeworth expansions provide the techniques for an alternative derivation of saddlepoint densities that is presented in this chapter. This alternative approach also gives further understanding as to why the saddlepoint density achieves its remarkable accuracy. Both unconditional and conditional saddlepoint approximations are considered for the canonical sufficient statistics in a regular exponential family. The computation of p-values for various likelihood ratio tests and the determination of some uniformly most powerful unbiased tests are some of the practical examples that are considered.
Regular exponential families
Definitions and examples
Let z be the data and denote Z as the random variable for the data before observation.
When the joint MGF of (X, Y) is available, the conditional MGF of Y given X = x can be approximated by using the sequential saddlepoint method. Apart from providing conditional moments, this approximate conditional MGF may also serve as a surrogate for the true conditional MGF with saddlepoint methods. In such a role, it can be the input into a singlesaddlepoint method, such as the Lugananni and Rice approximation, to give an approximate conditional distribution. Thus, the resulting sequential saddlepoint approximation to Pr(Y = y|X = x) provides an alternative to the double-saddlepoint methods of sections 4.1-4.2.
Computation of a p-value for the Bartlett–Nandi–Pillai trace statistic in MANOVA provides a context in which the sequential saddlepoint approximation succeeds with high accuracy but the competing double-saddlepoint CDF approximation fails. Among the latter methods, only numerical integration of the double-saddlepoint density successfully replicates the accuracy of the sequential saddlepoint CDF approximation; see Butler et al. (1992b).
Another highly successful application of sequential saddlepoint methods occurs when approximating the power function of Wilks' likelihood ratio test inMANOVA. This example is deferred to section 11.3.1.
Sequential saddlepoint approximation
Suppose (X, Y) is a m-dimensional random vector with known joint CGF K(s, t) where s and t are the respective components. The goal is to use the joint CGF to determine conditional probabilities and moments of Y given X = x.
The p* densitywas introduced in Barndorff-Nielsen (1980, 1983) and has been prominently featured as an approximation for the density of the maximum likelihood estimate (MLE). Its development from saddlepoint approximations and its role in likelihood inference for regular exponential families are discussed in section 7.1. Section 7.2 considers group transformation models, such as linear regression and location and scale models. In this setting the p* formula provides the conditional distributions ofMLEsgiven a maximal ancillary statistic and agrees with expressions developed by Fisher (1934) and Fraser (1979). In curved exponential families, the p* formula also approximates the conditional distribution of MLEs given certain approximate ancillaries. This development is explored in section 7.3 with a more detailed discussion given for the situation in which the ancillaries are affine.
The p* density in regular exponential families
In the context of a regular exponential family, the p* density is the normalized saddlepoint density for the MLE. In its unnormalized form in the continuous setting, p* is simply a Jacobian transformation removed from the saddlepoint density of the canonical sufficient statistic. Consider the regular exponential family in (5.1) with canonical parameter θ ε ε ⊆ ℜm and canonical sufficient statistic X. Statistic X is also the MLE û for the mean parameterization μ = c′(θ) as discussed in section 5.1.2.
Among the various tools that have been developed for use in statistics and probability over the years, perhaps the least understood and most remarkable tool is the saddlepoint approximation. It is remarkable because it usually provides probability approximations whose accuracy is much greater than the current supporting theory would suggest. It is least understood because of the difficulty of the subject itself and the difficulty of the research papers and books that have been written about it. Indeed this lack of accessibility has placed its understanding out of the reach of many researchers in both statistics and its related subjects.
The primary aim of this book is to provide an accessible account of the theory and application of saddlepoint methods that can be understood by the widest possible audience. To do this, the book has been written at graduated levels of difficulty with the first six chapters forming the easiest part and the core of the subject. These chapters use little mathematics beyond the difficulty of advanced calculus (no complex variables) and should provide relatively easy reading to first year graduate students in statistics, engineering, and other quantitative fields. These chapters would also be accessible to senior-level undergraduate mathematics and advanced engineering majors. With the accessibility issue in mind, the first six chapters have been purposefully written to address the issue and should assure that the widest audience is able to read and learn the subject.
The presentation throughout the book takes the point of view of users of saddlepoint approximations; theoretical aspects of the methods are also covered but are given less emphasis than they are in journal articles.
Statistical inference is considered in five practical settings. In each application, saddlepoint approximations offer an innovative approach for computing p-values, mid-p-values, power functions, and confidence intervals. As case studies, these five settings motivate additional considerations connected with both the saddlepoint methodology and theory that are used for making statistical inference. Each additional topic is addressed as it arises through its motivating example.
The first application concerns logistic regression, with an emphasis on the determination of LD50. p-value computation, test inversion to determine confidence intervals, and the notion of mid-p-values are considered. Saddlepoint treatment of prospective and retrospective analyses are also compared.
The second application deals with common odds ratio estimation. Both single- and double-saddlepoint methods are applicable and these two approaches are compared with several examples. Power function computations and their properties are discussed in connection with both the single- and double-saddlepoint approaches. Time series data may also be analyzed and an example dealing with an autoregression of Poisson counts is given.
The final two applications are concerned with modeling nonhomogeneous Poisson processes, and inference with data that has been truncated. The former example reconsiders the Lake Konstanz data analyzed in Barndorff-Nielsen and Cox (1979) and the latter deals with the truncation of binomial counts that occur with genetic data.
The general use of saddlepoint methods for testing purposes in exponential families was introduced in Davison (1988), however these applications had already been anticipated much earlier in the discussion by Daniels (1958) concerning the logistic setting as well as in Barndorff-Nielsen and Cox (1979). Test inversion to determine confidence intervals was first considered in Pierce and Peters (1992).
In 1993 the ITU recommended that in spectrum prices should follow a set of principles [1].
All spectrum users should pay a charge.
Non-discrimination – the spectrum charge should be calculated fairly, i.e. if two users are using the same amount of spectrum in the same way, both should pay the same charge.
The spectrum charge should be proportionate to the amount of bandwidth used.
The charges should reflect the spectrum's value to society, i.e. if need be, frequencies used for public services should be subject to lower charges.
The cost of spectrum regulation should not be borne by the state.
Spectrum users should be consulted about intended adjustments in spectrum charges.
The pricing structure should be clear, transparent and comprehensive, without unnecessarily lengthening the licensing process.
The pricing structure should reflect the scarcity of available spectrum and the level of demand for spectrum in different frequency bands.
The spectrum charge should be calculated so as to recover the costs of spectrum regulation. Spectrum pricing should not seek to maximise revenue for the government.
The ability to levy spectrum charges should be anchored in law.
As discussed in the previous chapter, the contemporary approach to the setting of incentive based spectrum prices places a greater emphasis on economic factors. While some of the principles above remain relevant to the setting of spectrum prices, the 1993 ITU recommendations contain contradictions. For example, it is generally not possible to set spectrum prices to reflect scarcity while at the same time recover only the administrative costs of regulating spectrum.
Much of the discussion in previous chapters has revolved around problems of spectrum management likely to be encountered in developed countries. It is thus pertinent to ask how the situation differs for developing countries.
If anything, their dependence on spectrum-using technologies is even greater. Lacking fixed networks to deliver communications services such as voice telephony and broadcasting, they are heavily reliant on spectrum for commercial and non-commercial services. This is illustrated by recent ITU data, which show the growth of penetration (per 100 inhabitants) of fixed and mobile lines in the least developed countries. It shows that mobile lines were roughly the same in number as fixed lines in 2001 (see Figure 16.1), but by 2004 they outnumbered fixed lines by three to one. Over the 2000–4 period the number of television receivers, mostly relying on terrestrial distribution, also increased by 50%. These data emphasise the importance of getting spectrum policy right.
Consequences for spectrum management
What special aspects of spectrum management are important in developing countries? It is helpful first to identify the differentiating factors between the two environments that are relevant.
Developing countries are characterised by a lower per capita income, which reduces consumption of all items including spectrum-using services.
Conversely, a lack of alternative platforms places a high priority on the development of wireless systems; there is also growing evidence that mobile communications can improve business efficiency, widen markets and promote income growth in developing countries.
At the same time, much spectrum in developing countries is not yet assigned, or assigned wastefully to government departments, especially defence forces.
In earlier chapters we have stated that there is a need for, and a benefit associated with, regulating radio spectrum use. In practice the costs of regulation are typically recovered through licence fees paid by radio spectrum users and hence there is a price associated with the use of licensed radio spectrum. For example, in the USA the FCC applies two types of fees – application fees and regulatory fees which cover the administrative cost of managing the use of spectrum, respectively. They may also serve to discourage the filing of frivolous applications. If set too high, however, fees can result in under-utilisation of the spectrum, while if set too low hoarding and congestion may arise.
The simple recovery of administrative costs via licence fees, while practised by almost every spectrum management agency around the world, fails to make use of one of the most powerful incentive mechanisms available to encourage more efficient use of radio spectrum. By varying licence fees in a suitable way, a spectrum manager can improve the economic and technical efficiency of spectrum management. The setting of incentive based prices is especially attractive in circumstances where spectrum has been assigned and/or allocated via administrative means rather than auctions. Incentive based pricing also works well in the absence of secondary trading, but as we show in this chapter, it can also work alongside spectrum trading.
Licence fees are a potent means of achieving greater efficiency for radio spectrum licensees holding non-auctioned spectrum.
In this chapter we look in more detail at different mechanisms for dividing up access to the spectrum. The previous chapter noted that spectrum is typically divided by frequency, with each user being given exclusive access to a frequency. In this chapter we look at the current process in more detail, assess all the different mechanisms for dividing spectrum and look at the impact that a change would have on the current use of radio spectrum.
Before delving deeper into spectrum access it is worth considering the question of what spectrum actually is. When a user gains a licence they do not actually gain any “spectrum”. Indeed, there is not really any such thing as radio spectrum, it is merely a representation of electro-magnetic radiation. Instead, users typically get a licence to transmit at a specific frequency and often with other conditions attached. Implicit to this is the expectation that they will be able to receive the transmitted signal without any harmful interference. We return to look at these conditions or “rights” in more detail in Chapters 7 and 8 where they become an important component of spectrum trading. As long as these rights are not compromised, other users could in principle access the same frequencies. Any method of dividing access to the spectrum that allows users to receive their transmitted signal would be viable.