LEADER 00000cam 2200973Ii 4500 001 ocn869091101 003 OCoLC 005 20160511074812.2 006 m o d 007 cr |n||||||||| 008 140125t20142014njua ob 001 0 eng d 020 9781400848911|q(electronic bk.) 020 1400848911|q(electronic bk.) 020 0691151687 020 9780691151687 020 9781306373845|q(MyiLibrary) 020 1306373840|q(MyiLibrary) 035 (OCoLC)869091101|z(OCoLC)873810539 040 EBLCP|beng|erda|epn|cEBLCP|dOCLCQ|dN$T|dE7B|dJSTOR|dGZM |dCOO|dUMI|dYDXCP|dDEBSZ|dOCLCO|dDEBBG|dOCLCQ|dOCLCO |dOCLCF|dCOH|dCUS|dCDS 049 MAIN 050 4 QA76.9.D343 .I384 2014 050 4 QB149|b.I949 2014 082 04 006.312 100 1 Ivezić, Željko. 245 10 Statistics, data mining, and machine learning in astronomy :|ba practical Python guide for the analysis of survey data /|cŽeljko Ivezić, Andrew J. Connolly, Jacob T. VanderPlas, and Alexander Gray. 264 1 Princeton, N.J. :|bPrinceton University Press,|c2014. 264 4 |c©2014 300 1 online resource (x, 540 pages) :|billustrations. 336 text|btxt|2rdacontent 337 computer|bc|2rdamedia 338 online resource|bcr|2rdacarrier 490 1 Princeton series in modern observational astronomy 505 0 I. Introduction -- 1. About the Book and Supporting Material -- 1.1. What do Data Mining, Machine Learning, and Knowledge Discovery mean? -- 1.2. What is this book about? -- 1.3. An incomplete survey of the relevant literature -- 1.4. Introduction to the Python Language and the Git Code Management Tool -- 1.5. Description of surveys and data sets used in examples -- 1.6. Plotting and visualizing the data in this book -- 1.7. How to efficiently use this book -- References -- 2. Fast Computation on Massive Data Sets -- 2.1. Data types and Data Management systems -- 2.2. Analysis of algorithmic efficiency -- 2.3. Seven types of computational Problem[s] -- 2.4. Seven strategies for speeding things up -- 2.5. Case studies: Speedup strategies in practice -- References. 505 8 II. Statistical Frameworks and Exploratory Data Analysis - - 3. Probability and Statistical Distributions -- 3.1. Brief overview of probability and random variables -- 3.2. Descriptive statistics -- 3.3. Common Univariate Distribution Functions -- 3.4. The Central Limit Theorem - - 3.5. Bivariate and Multivariate Distribution Functions - - 3.6. Correlation coefficients -- 3.7. Random number generation for arbitrary distributions -- References -- 4. Classical Statistical Inference -- 4.1. Classical vs. Bayesian Statistical Inference -- 4.2. Maximum Likelihood Estimation (MLE) -- 4.3. The goodness of Fit and Model Selection -- 4.4. ML Applied to Gaussian Mixtures: The Expectation Maximization Algorithm -- 4.5. Confidence estimates: the bootstrap and the jackknife -- 4.6. Hypothesis testing -- 4.7. Comparison of distributions -- 4.8. Nonparametric modeling and histograms -- 4.9. Selection effects and Luminosity Function Estimation -- 4.10. Summary -- References -- 5 Bayesian Statistical Inference -- 5.1. Introduction to the Bayesian method -- 5.2. Bayesian priors -- 5.3. Bayesian parameter uncertainty quantification -- 5.4. Bayesian model selection -- 5.5. Nonuniform priors: Eddington, Malmquist, and Lutz-Kelker biases -- 5.6. Simple examples of Bayesian analysis: Parameter estimation -- 5.7. Simple examples of Bayesian analysis: Model selection -- 5.8. Numerical methods for complex problems (MCMC) -- 5.9. Summary of pros and cons for classical and Bayesian methods -- References. 505 8 III. Data Mining and Machine Learning -- 6 Searching for Structure in Point Data -- 6.1. Nonparametric density estimation -- 6.2. Nearest-neighbor density estimation -- 6.3. Parametric density estimation -- 6.4. Finding clusters in data -- 6.5. Correlation functions -- 6.6. Which density estimation and clustering algorithms should I use? -- References -- 7 Dimensionality and its reduction -- 7.1. The curse of dimensionality -- 7.2. The data sets used in this chapter -- 7.3. Principal component analysis -- 7.4. Nonnegative matrix factorization -- 7.5. Manifold learning -- 7.6. Independent component analysis and projection pursuit -- 7.7. Which dimensionality reduction technique should I use? -- References -- 8 Regression and model fitting -- 8.1. Formulation of the regression problem -- 8.2. Regression for linear models -- 8.3. Regularization and penalizing the likelihood -- 8.4. Principal component regression -- 8.5. Kernel regression - - 8.6. Locally linear regression -- 8.7. Nonlinear regression -- 8.8. Uncertainties in the data -- 8.9. Regression that is robust to outliers -- 8.10. Gaussian process regression -- 8.11. Overfitting, underfitting, and cross-validation -- 8.12. Which regression method should I use? -- References. 505 8 III. Data Mining and Machine Learning (continued) -- 9 Classification -- 9.1. Data sets used in this chapter -- 9.2. Assigning categories: Classification -- 9.3. Generative classification -- 9.4. K-nearest-neighbor classifier -- 9.5. Discriminative classification -- 9.6. Support vector machines -- 9.7. Decision trees -- 9.8. Evaluating classifiers: ROC Curves -- 9.9. Which classifier should I use? -- References -- 10 Time Series Analysis -- 10.1. Main concepts for Time Series Analysis - - 10.2. Modeling toolkit for Time Series Analysis -- 10.3. Analysis of Periodic Time Series -- 10.4. Temporally localized signals -- 10.5. Analysis of Stochastic Processes -- 10.6. Which method should I use for Time Series Analysis? -- References. 505 8 IV. Appendices -- A An Introduction to Scientific Computing with Python -- A.1. A brief history of Python -- A.2. The ScyPy universe -- A.3. Getting started with Python -- A.4. IPython: The basics of interactive computing -- A.5. Introduction to NumPy -- A.6. Visualization with Matplotlib -- A.7. Overview of useful NumPy/SciPy modules -- A.8. Efficient coding with Python and NumPy -- A.9. Wrapping existing code in Python -- A.10. Other resources -- B AstroML: Machine Learning for Astronomy -- B.1. Introduction -- B.2. Dependencies -- B.3. Tools included in AstroML v0.1 -- C Astronomical Flux Measurements and Magnitudes -- C.1. The definition of the specific flux -- C.2. Wavelength window function for astronomical measurements -- C.3. The astronomical magnitude systems -- D SQL Query for Downloading SDSS Data -- E Approximating the Fourier Transform with the FFT -- References. 506 1 Unlimited number of concurrent users.|5UkHlHU 650 0 Statistical astronomy. 650 0 Astronomy|xData processing. 650 0 Python (Computer program language) 700 1 Connolly, Andrew|q(Andrew J.) 700 1 Vanderplas, Jacob T. 700 1 Gray, Alexander|q(Alexander G.) 830 0 Princeton series in modern observational astronomy. 856 40 |uhttps://www.jstor.org/stable/10.2307/j.ctt4cgbdj|zGo to ebook 936 JSTOR-D-2016/17