New Developments in Statistical Information Theory Based on...

New Developments in Statistical Information Theory Based on Entropy and Divergence Measures

Leandro Pardo
Bạn thích cuốn sách này tới mức nào?
Chất lượng của file scan thế nào?
Xin download sách để đánh giá chất lượng sách
Chất lượng của file tải xuống thế nào?
This book presents new and original research in Statistical Information Theory, based on minimum divergence estimators and test statistics, from a theoretical and applied point of view, for different statistical problems with special emphasis on efficiency and robustness. Divergence statistics, based on maximum likelihood estimators, as well as Wald’s statistics, likelihood ratio statistics and Rao’s score statistics, share several optimum asymptotic properties, but are highly non-robust in cases of model misspecification under the presence of outlying observations. It is well-known that a small deviation from the underlying assumptions on the model can have drastic effect on the performance of these classical tests. Specifically, this book presents a robust version of the classical Wald statistical test, for testing simple and composite null hypotheses for general parametric models, based on minimum divergence estimators.
Năm:
2019
Nhà xuát bản:
MDPI
Ngôn ngữ:
english
Trang:
346
ISBN 10:
3038979376
ISBN 13:
9783038979371
File:
PDF, 5.73 MB
IPFS:
CID , CID Blake2b
english, 2019
Đọc online
Hoàn thành chuyển đổi thành trong
Chuyển đổi thành không thành công

Từ khóa thường sử dụng nhất