Catálogo de publicaciones - libros
Nonlinear Problems of Elasticity
Stuart S. Antman
Second Edition.
Resumen/Descripción – provisto por la editorial
No disponible.
Palabras clave – provistas por la editorial
No disponibles.
Disponibilidad
Institución detectada | Año de publicación | Navegá | Descargá | Solicitá |
---|---|---|---|---|
No detectada | 2005 | SpringerLink |
Información
Tipo de recurso:
libros
ISBN impreso
978-0-387-20880-0
ISBN electrónico
978-0-387-27649-6
Editor responsable
Springer Nature
País de edición
Reino Unido
Fecha de publicación
2005
Información sobre derechos de publicación
© Springer Science+Business Media, Inc. 2005
Cobertura temática
Tabla de contenidos
Tensors
Stuart S. Antman
All of the standard inferences in RSM as presented in previous chapters are based on point estimators which have sampling, or experimental, variability. Assuming a classical or frequentist point of view, every quantity computed based on experimental data is subject to sampling variability and is therefore a random quantity itself. As Draper [48] pointed out, one should not expect precise conclusions when using mathematical optimization techniques based on data subject to large errors. This comment applies to every technique previously discussed, namely, the steepest ascent/descent direction, eigenvalues of the quadratic matrix and point estimators of the stationary or optimal points in quadratic (second order) optimization for both canonical and ridge analysis. It also applies to more sophisticated mathematical programming techniques. In the RSM literature, there has been an over-emphasis on using different types of such mathematical techniques which neglect the main statistical issue that arises from random data: if the experiment is repeated and new models fitted, the parameters (or even the response model form) may change, and this will necessarily result in a different optimal solution.
Pp. 399-416
3-Dimensional Continuum Mechanics
Stuart S. Antman
All of the standard inferences in RSM as presented in previous chapters are based on point estimators which have sampling, or experimental, variability. Assuming a classical or frequentist point of view, every quantity computed based on experimental data is subject to sampling variability and is therefore a random quantity itself. As Draper [48] pointed out, one should not expect precise conclusions when using mathematical optimization techniques based on data subject to large errors. This comment applies to every technique previously discussed, namely, the steepest ascent/descent direction, eigenvalues of the quadratic matrix and point estimators of the stationary or optimal points in quadratic (second order) optimization for both canonical and ridge analysis. It also applies to more sophisticated mathematical programming techniques. In the RSM literature, there has been an over-emphasis on using different types of such mathematical techniques which neglect the main statistical issue that arises from random data: if the experiment is repeated and new models fitted, the parameters (or even the response model form) may change, and this will necessarily result in a different optimal solution.
Pp. 417-490
3-Dimensional Theory of Nonlinear Elasticity
Stuart S. Antman
All of the standard inferences in RSM as presented in previous chapters are based on point estimators which have sampling, or experimental, variability. Assuming a classical or frequentist point of view, every quantity computed based on experimental data is subject to sampling variability and is therefore a random quantity itself. As Draper [48] pointed out, one should not expect precise conclusions when using mathematical optimization techniques based on data subject to large errors. This comment applies to every technique previously discussed, namely, the steepest ascent/descent direction, eigenvalues of the quadratic matrix and point estimators of the stationary or optimal points in quadratic (second order) optimization for both canonical and ridge analysis. It also applies to more sophisticated mathematical programming techniques. In the RSM literature, there has been an over-emphasis on using different types of such mathematical techniques which neglect the main statistical issue that arises from random data: if the experiment is repeated and new models fitted, the parameters (or even the response model form) may change, and this will necessarily result in a different optimal solution.
Pp. 491-512
Problems in Nonlinear Elasticity
Stuart S. Antman
All of the standard inferences in RSM as presented in previous chapters are based on point estimators which have sampling, or experimental, variability. Assuming a classical or frequentist point of view, every quantity computed based on experimental data is subject to sampling variability and is therefore a random quantity itself. As Draper [48] pointed out, one should not expect precise conclusions when using mathematical optimization techniques based on data subject to large errors. This comment applies to every technique previously discussed, namely, the steepest ascent/descent direction, eigenvalues of the quadratic matrix and point estimators of the stationary or optimal points in quadratic (second order) optimization for both canonical and ridge analysis. It also applies to more sophisticated mathematical programming techniques. In the RSM literature, there has been an over-emphasis on using different types of such mathematical techniques which neglect the main statistical issue that arises from random data: if the experiment is repeated and new models fitted, the parameters (or even the response model form) may change, and this will necessarily result in a different optimal solution.
Pp. 513-584
Large-Strain Plasticity
Stuart S. Antman
All of the standard inferences in RSM as presented in previous chapters are based on point estimators which have sampling, or experimental, variability. Assuming a classical or frequentist point of view, every quantity computed based on experimental data is subject to sampling variability and is therefore a random quantity itself. As Draper [48] pointed out, one should not expect precise conclusions when using mathematical optimization techniques based on data subject to large errors. This comment applies to every technique previously discussed, namely, the steepest ascent/descent direction, eigenvalues of the quadratic matrix and point estimators of the stationary or optimal points in quadratic (second order) optimization for both canonical and ridge analysis. It also applies to more sophisticated mathematical programming techniques. In the RSM literature, there has been an over-emphasis on using different types of such mathematical techniques which neglect the main statistical issue that arises from random data: if the experiment is repeated and new models fitted, the parameters (or even the response model form) may change, and this will necessarily result in a different optimal solution.
Pp. 585-601
General Theories of Rods
Stuart S. Antman
All of the standard inferences in RSM as presented in previous chapters are based on point estimators which have sampling, or experimental, variability. Assuming a classical or frequentist point of view, every quantity computed based on experimental data is subject to sampling variability and is therefore a random quantity itself. As Draper [48] pointed out, one should not expect precise conclusions when using mathematical optimization techniques based on data subject to large errors. This comment applies to every technique previously discussed, namely, the steepest ascent/descent direction, eigenvalues of the quadratic matrix and point estimators of the stationary or optimal points in quadratic (second order) optimization for both canonical and ridge analysis. It also applies to more sophisticated mathematical programming techniques. In the RSM literature, there has been an over-emphasis on using different types of such mathematical techniques which neglect the main statistical issue that arises from random data: if the experiment is repeated and new models fitted, the parameters (or even the response model form) may change, and this will necessarily result in a different optimal solution.
Pp. 603-658
General Theories of Shells
Stuart S. Antman
All of the standard inferences in RSM as presented in previous chapters are based on point estimators which have sampling, or experimental, variability. Assuming a classical or frequentist point of view, every quantity computed based on experimental data is subject to sampling variability and is therefore a random quantity itself. As Draper [48] pointed out, one should not expect precise conclusions when using mathematical optimization techniques based on data subject to large errors. This comment applies to every technique previously discussed, namely, the steepest ascent/descent direction, eigenvalues of the quadratic matrix and point estimators of the stationary or optimal points in quadratic (second order) optimization for both canonical and ridge analysis. It also applies to more sophisticated mathematical programming techniques. In the RSM literature, there has been an over-emphasis on using different types of such mathematical techniques which neglect the main statistical issue that arises from random data: if the experiment is repeated and new models fitted, the parameters (or even the response model form) may change, and this will necessarily result in a different optimal solution.
Pp. 659-708
Dynamical Problems
Stuart S. Antman
All of the standard inferences in RSM as presented in previous chapters are based on point estimators which have sampling, or experimental, variability. Assuming a classical or frequentist point of view, every quantity computed based on experimental data is subject to sampling variability and is therefore a random quantity itself. As Draper [48] pointed out, one should not expect precise conclusions when using mathematical optimization techniques based on data subject to large errors. This comment applies to every technique previously discussed, namely, the steepest ascent/descent direction, eigenvalues of the quadratic matrix and point estimators of the stationary or optimal points in quadratic (second order) optimization for both canonical and ridge analysis. It also applies to more sophisticated mathematical programming techniques. In the RSM literature, there has been an over-emphasis on using different types of such mathematical techniques which neglect the main statistical issue that arises from random data: if the experiment is repeated and new models fitted, the parameters (or even the response model form) may change, and this will necessarily result in a different optimal solution.
Pp. 709-750
Appendix. Topics in Linear Analysis
Stuart S. Antman
All of the standard inferences in RSM as presented in previous chapters are based on point estimators which have sampling, or experimental, variability. Assuming a classical or frequentist point of view, every quantity computed based on experimental data is subject to sampling variability and is therefore a random quantity itself. As Draper [48] pointed out, one should not expect precise conclusions when using mathematical optimization techniques based on data subject to large errors. This comment applies to every technique previously discussed, namely, the steepest ascent/descent direction, eigenvalues of the quadratic matrix and point estimators of the stationary or optimal points in quadratic (second order) optimization for both canonical and ridge analysis. It also applies to more sophisticated mathematical programming techniques. In the RSM literature, there has been an over-emphasis on using different types of such mathematical techniques which neglect the main statistical issue that arises from random data: if the experiment is repeated and new models fitted, the parameters (or even the response model form) may change, and this will necessarily result in a different optimal solution.
Pp. 751-759
Appendix. Local Nonlinear Analysis
Stuart S. Antman
All of the standard inferences in RSM as presented in previous chapters are based on point estimators which have sampling, or experimental, variability. Assuming a classical or frequentist point of view, every quantity computed based on experimental data is subject to sampling variability and is therefore a random quantity itself. As Draper [48] pointed out, one should not expect precise conclusions when using mathematical optimization techniques based on data subject to large errors. This comment applies to every technique previously discussed, namely, the steepest ascent/descent direction, eigenvalues of the quadratic matrix and point estimators of the stationary or optimal points in quadratic (second order) optimization for both canonical and ridge analysis. It also applies to more sophisticated mathematical programming techniques. In the RSM literature, there has been an over-emphasis on using different types of such mathematical techniques which neglect the main statistical issue that arises from random data: if the experiment is repeated and new models fitted, the parameters (or even the response model form) may change, and this will necessarily result in a different optimal solution.
Pp. 761-767