To save content items to your account,
please confirm that you agree to abide by our usage policies.
If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account.
Find out more about saving content to .
To save content items to your Kindle, first ensure no-reply@cambridge.org
is added to your Approved Personal Document E-mail List under your Personal Document Settings
on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part
of your Kindle email address below.
Find out more about saving to your Kindle.
Note you can select to save to either the @free.kindle.com or @kindle.com variations.
‘@free.kindle.com’ emails are free but can only be saved to your device when it is connected to wi-fi.
‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.
Build a firm foundation for studying statistical modelling, data science, and machine learning with this practical introduction to statistics, written with chemical engineers in mind. It introduces a data–model–decision approach to applying statistical methods to real-world chemical engineering challenges, establishes links between statistics, probability, linear algebra, calculus, and optimization, and covers classical and modern topics such as uncertainty quantification, risk modelling, and decision-making under uncertainty. Over 100 worked examples using Matlab and Python demonstrate how to apply theory to practice, with over 70 end-of-chapter problems to reinforce student learning, and key topics are introduced using a modular structure, which supports learning at a range of paces and levels. Requiring only a basic understanding of calculus and linear algebra, this textbook is the ideal introduction for undergraduate students in chemical engineering, and a valuable preparatory text for advanced courses in data science and machine learning with chemical engineering applications.
This chapter discusses techniques to build predictive models from data and to quantify the uncertainty of the model parameters and of the model predictions. The chapter discusses important concepts of linear and nonlinear regression and focuses on a couple of major paradigms used for estimation: maximum likelihood and Bayesian estimation. The chapter also discusses how to incorporate prior knowledge in the estimation process.
This chapter provides an end-to-end introduction to statistics; this highlights how statistics can be used to develop models from data, to quantify the uncertainty of such models, and to make decisions under uncertainty. The chapter also discusses how random variables are the key modeling paradigm that is used in statistics to characterize and quantify uncertainty and risk.
This chapter provides a discussion on multivariate random variables, which are collections of univariate random variables. The chapter discusses how the presence of multiple random variables gives rise to concepts of covariance and correlation, which capture relationships that can arise between variables. The chapter also discussed the multivariate Gaussian model, which is widely used in applications.
This chapter discusses how to apply principles of statistics, optimization, and linear algebra in advanced techniques of data science and machine learning. The chapter shows how to use principal component analysis and singular value decomposition for analyzing complex datasets and discusses advanced estimation techniques such as logistic regression, Gaussian process models, and neural networks.
This chapter provides an overview of different theoretical random variable models that can be used to model random phenomena encountered in applications. The chapter discusses the types of behavior that different models capture and provides some preliminary discussion on how to determine model parameters from data.
This chapter discusses techniques that help us estimate parameters and summarizing statistics for random variables from data. The chapter discusses techniques such as the method of moments, least-squares, and maximum likelihood. The chapter also touches on concepts of Monte Carlo simulation, which is a technique that can be used to approximate the summarizing statistics of random variables from random samples or from data. The chapter also highlights how one can characterize the quality of such approximations using the central limit theorem and the law of large numbers.
This chapter discusses techniques to measure uncertainty/risk and to make decisions that explicitly take risk into consideration. The chapter also discusses how to use principles of statistics and optimization in advanced decision-making techniques such as stochastic programming, flexibility analysis, and Bayesian optimization.
The objective of this chapter is to extend the ad hoc least squares method of somewhat arbitrarily selected base functions to a more generic method applicable to a broad range of functions – the Fourier series, which is an expansion of a relatively arbitrary function (with certain smoothness requirement and finite jumps at worst) with a series of sinusoidal functions. An important mathematical reason for using Fourier series is its “completeness” and almost guaranteed convergence. Here “completeness” means that the error goes to zero when the whole Fourier series with infinite base function is used. In other words, the Fourier series formed by the selected sinusoidal functions is sufficient to linearly combine into a function that converges to an arbitrary continuous function. This chapter on Fourier series will lay out a foundation that will lead to Fourier Transform and spectrum analysis. In this sense, this chapter is important as it provides background information and theoretical preparation.
The objective of this chapter is to present some important relations between the Fourier Transform and correlation functions. It turns out that the cross-correlation function and autocorrelation function have some useful relationships to Fourier Transform and power spectrum of the individual functions. As a result, cospectrum and coherence (normalized statistical correlation in frequency domain) can be defined.