|
|
|
|
LEADER |
00000cam a2200000 a 4500 |
001 |
EBSCO_ocn704992965 |
003 |
OCoLC |
005 |
20231017213018.0 |
006 |
m o d |
007 |
cr cnu---unuuu |
008 |
110303s2011 enka ob 001 0 eng d |
040 |
|
|
|a N$T
|b eng
|e pn
|c N$T
|d E7B
|d OCLCQ
|d FXR
|d YDXCP
|d OCLCO
|d OCLCQ
|d CAMBR
|d OCLCF
|d OL$
|d OCLCQ
|d OTZ
|d COD
|d OCLCQ
|d OCLCO
|d OCLCA
|d COO
|d UUM
|d COCUF
|d CNNOR
|d STF
|d USU
|d INT
|d LOA
|d VT2
|d CNCEN
|d OCLCQ
|d WYU
|d LVT
|d S8J
|d S9I
|d OCLCQ
|d AGLDB
|d D6H
|d AU@
|d OCLCO
|d OCLCQ
|d A6Q
|d OCLCA
|d UKCRE
|d OCLCQ
|d OCLCO
|d UKAHL
|d QGK
|d OCLCQ
|d OCLCO
|d OCLCQ
|
015 |
|
|
|a GBB0A7521
|2 bnb
|
016 |
7 |
|
|a 101558557
|2 DNLM
|
016 |
7 |
|
|a 015644135
|2 Uk
|
019 |
|
|
|a 801658322
|a 817925247
|a 853659688
|a 994898677
|a 1014321280
|a 1042908465
|a 1043677657
|a 1058912026
|a 1059246746
|a 1066491823
|a 1076746252
|a 1081286710
|a 1084344547
|a 1096217353
|a 1105766366
|a 1148143551
|a 1170365116
|a 1171315620
|a 1228531029
|a 1259215619
|
020 |
|
|
|a 9780511993121
|q (electronic bk.)
|
020 |
|
|
|a 0511993129
|q (electronic bk.)
|
020 |
|
|
|a 9780511989308
|q (electronic bk.)
|
020 |
|
|
|a 051198930X
|q (electronic bk.)
|
020 |
|
|
|a 9780511975820
|q (electronic bk.)
|
020 |
|
|
|a 0511975821
|q (electronic bk.)
|
020 |
|
|
|a 1107218802
|
020 |
|
|
|a 9781107218802
|
020 |
|
|
|a 1282978349
|
020 |
|
|
|a 9781282978348
|
020 |
|
|
|a 9786612978340
|
020 |
|
|
|a 6612978341
|
020 |
|
|
|a 0511992092
|
020 |
|
|
|a 9780511992094
|
020 |
|
|
|a 0511987528
|
020 |
|
|
|a 9780511987526
|
020 |
|
|
|a 0511991118
|
020 |
|
|
|a 9780511991110
|
020 |
|
|
|z 9780521875806
|
020 |
|
|
|z 0521875803
|
020 |
|
|
|z 9780521699099
|
020 |
|
|
|z 0521699096
|
035 |
|
|
|a (OCoLC)704992965
|z (OCoLC)801658322
|z (OCoLC)817925247
|z (OCoLC)853659688
|z (OCoLC)994898677
|z (OCoLC)1014321280
|z (OCoLC)1042908465
|z (OCoLC)1043677657
|z (OCoLC)1058912026
|z (OCoLC)1059246746
|z (OCoLC)1066491823
|z (OCoLC)1076746252
|z (OCoLC)1081286710
|z (OCoLC)1084344547
|z (OCoLC)1096217353
|z (OCoLC)1105766366
|z (OCoLC)1148143551
|z (OCoLC)1170365116
|z (OCoLC)1171315620
|z (OCoLC)1228531029
|z (OCoLC)1259215619
|
050 |
|
4 |
|a QH324.2
|b .M35 2011eb
|
060 |
|
4 |
|a 2011 D-913
|
060 |
|
4 |
|a WA 950
|
072 |
|
7 |
|a MED
|x 076000
|2 bisacsh
|
072 |
|
7 |
|a MED
|x 030000
|2 bisacsh
|
072 |
|
7 |
|a MED
|x 078000
|2 bisacsh
|
082 |
0 |
4 |
|a 614.285
|2 22
|
049 |
|
|
|a UAMI
|
100 |
1 |
|
|a Malley, James D.
|
245 |
1 |
0 |
|a Statistical learning for biomedical data /
|c James D. Malley, Karen G. Malley, Sinisa Pajevic.
|
260 |
|
|
|a Cambridge ;
|a New York :
|b Cambridge University Press,
|c 2011.
|
300 |
|
|
|a 1 online resource (xii, 285 pages) :
|b illustrations
|
336 |
|
|
|a text
|b txt
|2 rdacontent
|
337 |
|
|
|a computer
|b c
|2 rdamedia
|
338 |
|
|
|a online resource
|b cr
|2 rdacarrier
|
347 |
|
|
|a data file
|
490 |
1 |
|
|a Practical guides to biostatistics and epidemiology
|
520 |
|
|
|a This highly motivating introduction to statistical learning machines explains underlying principles in nontechnical language, using many examples and figures.
|
504 |
|
|
|a Includes bibliographical references and index.
|
588 |
0 |
|
|a Print version record.
|
505 |
0 |
|
|a pt. 1. Introduction -- pt. 2. A machine toolkit -- pt. 3. Analysis fundamentals -- pt. 4. Machine strategies.
|
505 |
0 |
0 |
|g Part I.
|t Introduction --
|g 1.
|t Prologue --
|g 1.1.
|t Machines that learn -- some recent history --
|g 1.2.
|t Twenty canonical questions --
|g 1.3.
|t Outline of the book --
|g 1.4.
|t A comment about example datasets --
|g 1.5.
|t Software --
|g 2.
|t The landscape of learning machines --
|g 2.1.
|t Introduction --
|g 2.2.
|t Types of data for learning machines --
|g 2.3.
|t Will that be supervised or unsupervised? --
|g 2.4.
|t An unsupervised example --
|g 2.5.
|t More lack of supervision -- where are the parents? --
|g 2.6.
|t Engines, complex and primitive --
|g 2.7.
|t Model richness means what, exactly? --
|g 2.8.
|t Membership or probability of membership? --
|g 2.9.
|t A taxonomy of machines? --
|g 2.10.
|t A note of caution -- one of many --
|g 2.11.
|t Highlights from the theory --
|g 3.
|t A mangle of machines --
|g 3.1.
|t Introduction --
|g 3.2.
|t Linear regression --
|g 3.3.
|t Logistic regression --
|g 3.4.
|t Linear discriminant --
|g 3.5.
|t Bayes classifiers -- regular and naïve --
|g 3.6.
|t Logic regression --
|g 3.7.
|t k-Nearest neighbors --
|g 3.8.
|t Support vector machines --
|g 3.9.
|t Neural networks --
|g 3.10.
|t Boosting --
|g 3.11.
|t Evolutionary and genetic algorithms --
|g 4.
|t Three examples and several machines --
|g 4.1.
|t Introduction --
|g 4.2.
|t Simulated cholesterol data --
|g 4.3.
|t Lupus data --
|g 4.4.
|t Stroke data --
|g 4.5.
|t Biomedical means unbalanced --
|g 4.6.
|t Measures of machine performance --
|g 4.7.
|t Linear analysis of cholesterol data --
|g 4.8.
|t Nonlinear analysis of cholesterol data --
|g 4.9.
|t Analysis of the lupus data --
|g 4.10.
|t Analysis of the stroke data --
|g 4.11.
|t Further analysis of the lupus and stroke data --
|g Part II.
|t A machine toolkit --
|g 5.
|t Logistic regression --
|g 5.1.
|t Introduction --
|g 5.2.
|t Inside and around the model --
|g 5.3.
|t Interpreting the coefficients --
|g 5.4.
|t Using logistic regression as a decision rule --
|g 5.5.
|t Logistic regression applied to the cholesterol data --
|g 5.6.
|t A cautionary note --
|g 5.7.
|t Another cautionary note --
|g 5.8.
|t Probability estimates and decision rules --
|g 5.9.
|t Evaluating the goodness-of-fit of a logistic regression model --
|g 5.10.
|t Calibrating a logistic regression --
|g 5.11.
|t Beyond calibration --
|g 5.12.
|t Logistic regression and reference models --
|g 6.
|t A single decision tree --
|g 6.1.
|t Introduction --
|g 6.2.
|t Dropping down trees --
|g 6.3.
|t Growing a tree --
|g 6.4.
|t Selecting features, making splits --
|g 6.5.
|t Good split, bad split --
|g 6.6.
|t Finding good features for making splits --
|g 6.7.
|t Misreading trees --
|g 6.8.
|t Stopping and pruning rules --
|g 6.9.
|t Using functions of the features --
|g 6.10.
|t Unstable trees? --
|g 6.11.
|t Variable importance -- growing on trees? --
|g 6.12.
|t Permuting for importance --
|g 6.13.
|t The continuing mystery of trees --
|g 7.
|t Random Forests -- trees everywhere --
|g 7.1.
|t Random Forests in less than five minutes --
|g 7.2.
|t Random treks through the data --
|g 7.3.
|t Random treks through the features --
|g 7.4.
|t Walking through the forest --
|g 7.5.
|t Weighted and unweighted voting --
|g 7.6.
|t Finding subsets in the data using proximities --
|g 7.7.
|t Applying Random Forests to the Stroke data --
|g 7.8.
|t Random Forests in the universe of machines --
|g Part III.
|t Analysis fundamentals --
|g 8.
|t Merely two variables --
|g 8.1.
|t Introduction --
|g 8.2.
|t Understanding correlations --
|g 8.3.
|t Hazards of correlations --
|g 8.4.
|t Correlations big and small --
|g 9.
|t More than two variables --
|g 9.1.
|t Introduction --
|g 9.2.
|t Tiny problems, large consequences --
|g 9.3.
|t Mathematics to the rescue? --
|g 9.4.
|t Good models need not be unique --
|g 9.5.
|t Contexts and coefficients --
|g 9.6.
|t Interpreting and testing coefficients in models --
|g 9.7.
|t Merging models, pooling lists, ranking features --
|g 10.
|t Resampling methods --
|g 10.1.
|t Introduction --
|g 10.2.
|t The bootstrap --
|g 10.3.
|t When the bootstrap works --
|g 10.4.
|t When the bootstrap doesn't work --
|g 10.5.
|t Resampling from a single group in different ways --
|g 10.6.
|t Resampling from groups with unequal sizes --
|g 10.7.
|t Resampling from small datasets --
|g 10.8.
|t Permutation methods --
|g 10.9.
|t Still more on permutation methods --
|g 11.
|t Error analysis and model validation --
|g 11.1.
|t Introduction --
|g 11.2.
|t Errors? What errors? --
|g 11.3.
|t Unbalanced data, unbalanced errors --
|g 11.4.
|t Error analysis for a single machine --
|g 11.5.
|t Cross-validation error estimation --
|g 11.6.
|t Cross-validation or cross-training? --
|g 11.7.
|t The leave-one-out method --
|g 11.8.
|t The out-of-bag method --
|g 11.9.
|t Intervals for error estimates for a single machine --
|g 11.10.
|t Tossing random coins into the abyss --
|g 11.11.
|t Error estimates for unbalanced data --
|g 11.12.
|t Confidence intervals for comparing error values --
|g 11.13.
|t Other measures of machine accuracy --
|g 11.14.
|t Benchmarking and winning the lottery --
|g 11.15.
|t Error analysis for predicting continuous outcomes --
|g Part IV.
|t Machine strategies --
|g 12.
|t Ensemble methods -- let's take a vote --
|g 12.1.
|t Pools of machines --
|g 12.2.
|t Weak correlation with outcome can be good enough --
|g 12.3.
|t Model averaging --
|g 13.
|t Summary and conclusions --
|g 13.1.
|t Where have we been? --
|g 13.2.
|t So many machines --
|g 13.3.
|t Binary decision or probability estimate? --
|g 13.4.
|t Survival machines? Risk machines? --
|g 13.5.
|t And where are we going?
|
520 |
|
|
|a "This book is for anyone who has biomedical data and needs to identify variables that predict an outcome, for two-group outcomes such as tumor/not-tumor, survival/death, or response from treatment. Statistical learning machines are ideally suited to these types of prediction problems, especially if the variables being studied may not meet the assumptions of traditional techniques. Learning machines come from the world of probability and computer science but are not yet widely used in biomedical research."--Publisher's website
|
546 |
|
|
|a English.
|
590 |
|
|
|a eBooks on EBSCOhost
|b EBSCO eBook Subscription Academic Collection - Worldwide
|
650 |
|
0 |
|a Medical statistics
|x Data processing.
|
650 |
|
0 |
|a Biometry
|x Data processing.
|
650 |
|
2 |
|a Data Interpretation, Statistical
|
650 |
|
2 |
|a Models, Statistical
|
650 |
|
6 |
|a Biométrie
|x Informatique.
|
650 |
|
7 |
|a MEDICAL
|x Preventive Medicine.
|2 bisacsh
|
650 |
|
7 |
|a MEDICAL
|x Forensic Medicine.
|2 bisacsh
|
650 |
|
7 |
|a MEDICAL
|x Public Health.
|2 bisacsh
|
650 |
|
7 |
|a Biometry
|x Data processing.
|2 fast
|0 (OCoLC)fst00832613
|
650 |
|
7 |
|a Medical statistics
|x Data processing.
|2 fast
|0 (OCoLC)fst01014674
|
700 |
1 |
|
|a Malley, Karen G.
|
700 |
1 |
|
|a Pajevic, Sinisa.
|
776 |
0 |
8 |
|i Print version:
|a Malley, James D.
|t Statistical learning for biomedical data.
|d Cambridge : Cambridge University Press, 2011
|z 9780521875806
|w (DLC) 2011377705
|w (OCoLC)663441381
|
830 |
|
0 |
|a Practical guides to biostatistics and epidemiology.
|
856 |
4 |
0 |
|u https://ebsco.uam.elogim.com/login.aspx?direct=true&scope=site&db=nlebk&AN=352491
|z Texto completo
|
938 |
|
|
|a Askews and Holts Library Services
|b ASKH
|n AH22948792
|
938 |
|
|
|a ebrary
|b EBRY
|n ebr10442834
|
938 |
|
|
|a EBSCOhost
|b EBSC
|n 352491
|
938 |
|
|
|a YBP Library Services
|b YANK
|n 3604836
|
938 |
|
|
|a YBP Library Services
|b YANK
|n 3610811
|
938 |
|
|
|a YBP Library Services
|b YANK
|n 3642193
|
938 |
|
|
|a YBP Library Services
|b YANK
|n 9248024
|
994 |
|
|
|a 92
|b IZTAP
|