This document presents a bootstrap approach to reduce errors in estimating parameters for nonlinear regression models. It proposes using bootstrapping along with the Gauss-Newton method for parameter estimation. The method is tested on an exponential regression model. Results show the bootstrap algorithm yielded a better (reduced) error sum of squares compared to the analytical method, providing greater confidence in the estimated parameters.
This document discusses predictive mean matching (PMM) imputation in survey sampling. It begins with an outline and overview of the basic setup, assumptions, and PMM imputation method. It then presents three main theorems: 1) the asymptotic normality of the PMM estimator when the regression parameter β* is known, 2) the asymptotic normality when β* is estimated, and 3) the asymptotic properties of nearest neighbor imputation. The document also discusses variance estimation for the PMM estimator using replication methods like the bootstrap or jackknife. In summary, it provides a theoretical analysis of the asymptotic properties of PMM imputation and approaches for estimating the variance.
This document discusses nonstationary covariance modeling. It begins by introducing concepts of covariance, correlation, and how correlation affects estimation and prediction. It then discusses properties of random fields like second-order stationarity and intrinsic stationarity. The difference between these two is explained. Common parametric models for isotropic covariance and variogram functions are presented, including spherical, exponential, Gaussian, rational quadratic, and Matérn models. Parameters and properties of these models like range, smoothness, and valid dimensionality are described. Examples of each model type are shown graphically.
The document discusses the benefits of exercise for mental health. Regular physical activity can help reduce anxiety and depression and improve mood and cognitive functioning. Exercise causes chemical changes in the brain that may help protect against mental illness and improve symptoms.
This document summarizes several statistical methods for handling non-ignorable nonresponse in data, including maximum likelihood estimation, partial likelihood approaches, generalized method of moments, and exponential tilting methods. It discusses full likelihood-based maximum likelihood estimation using the observed data likelihood and EM algorithm. Partial likelihood approaches like conditional likelihood and pseudo likelihood are presented as alternatives that use a subset of the observed data.
The document discusses estimation of multi-Granger network causal models from time series data. It proposes a joint modeling approach to estimate vector autoregressive (VAR) models for multiple time series datasets simultaneously. The key steps are:
1. Estimate the inverse covariance matrices for each dataset using a factor model approach.
2. Use the estimated inverse covariance matrices in a generalized fused lasso optimization to jointly estimate the VAR coefficient matrices for each dataset.
Simulation results show the joint modeling approach improves estimation of the VAR coefficients and reduces forecasting error compared to estimating the models separately, especially when the number of time points is small. The factor modeling approach also provides a better estimate of the inverse covariance than using the empirical estimate.
The document outlines rules for graphing and linear regression. It discusses key aspects of graphing such as properly labeling axes, selecting appropriate scales, and drawing lines smoothly through data points. It then covers linear regression, defining the linear equation and describing how to calculate the slope and y-intercept by minimizing the sum of the deviations between observed and calculated data points. Formulas for determining the slope and y-intercept using linear regression are provided.
Some sampling techniques for big data analysisJae-kwang Kim
This document describes different sampling techniques for big data analysis, including reservoir sampling and its variants. It provides an example to illustrate simple random sampling and calculates the expected value and variance of sampling errors. It then discusses probability sampling and its advantages over non-probability sampling. The document also introduces survey sampling and challenges in the era of big data, as well as how sampling techniques can still be useful for handling big data. It outlines reservoir sampling and two methods to improve it: balanced reservoir sampling and stratified reservoir sampling. A simulation study is described to compare the performance of these reservoir sampling methods.
This document discusses predictive mean matching (PMM) imputation in survey sampling. It begins with an outline and overview of the basic setup, assumptions, and PMM imputation method. It then presents three main theorems: 1) the asymptotic normality of the PMM estimator when the regression parameter β* is known, 2) the asymptotic normality when β* is estimated, and 3) the asymptotic properties of nearest neighbor imputation. The document also discusses variance estimation for the PMM estimator using replication methods like the bootstrap or jackknife. In summary, it provides a theoretical analysis of the asymptotic properties of PMM imputation and approaches for estimating the variance.
This document discusses nonstationary covariance modeling. It begins by introducing concepts of covariance, correlation, and how correlation affects estimation and prediction. It then discusses properties of random fields like second-order stationarity and intrinsic stationarity. The difference between these two is explained. Common parametric models for isotropic covariance and variogram functions are presented, including spherical, exponential, Gaussian, rational quadratic, and Matérn models. Parameters and properties of these models like range, smoothness, and valid dimensionality are described. Examples of each model type are shown graphically.
The document discusses the benefits of exercise for mental health. Regular physical activity can help reduce anxiety and depression and improve mood and cognitive functioning. Exercise causes chemical changes in the brain that may help protect against mental illness and improve symptoms.
This document summarizes several statistical methods for handling non-ignorable nonresponse in data, including maximum likelihood estimation, partial likelihood approaches, generalized method of moments, and exponential tilting methods. It discusses full likelihood-based maximum likelihood estimation using the observed data likelihood and EM algorithm. Partial likelihood approaches like conditional likelihood and pseudo likelihood are presented as alternatives that use a subset of the observed data.
The document discusses estimation of multi-Granger network causal models from time series data. It proposes a joint modeling approach to estimate vector autoregressive (VAR) models for multiple time series datasets simultaneously. The key steps are:
1. Estimate the inverse covariance matrices for each dataset using a factor model approach.
2. Use the estimated inverse covariance matrices in a generalized fused lasso optimization to jointly estimate the VAR coefficient matrices for each dataset.
Simulation results show the joint modeling approach improves estimation of the VAR coefficients and reduces forecasting error compared to estimating the models separately, especially when the number of time points is small. The factor modeling approach also provides a better estimate of the inverse covariance than using the empirical estimate.
The document outlines rules for graphing and linear regression. It discusses key aspects of graphing such as properly labeling axes, selecting appropriate scales, and drawing lines smoothly through data points. It then covers linear regression, defining the linear equation and describing how to calculate the slope and y-intercept by minimizing the sum of the deviations between observed and calculated data points. Formulas for determining the slope and y-intercept using linear regression are provided.
Some sampling techniques for big data analysisJae-kwang Kim
This document describes different sampling techniques for big data analysis, including reservoir sampling and its variants. It provides an example to illustrate simple random sampling and calculates the expected value and variance of sampling errors. It then discusses probability sampling and its advantages over non-probability sampling. The document also introduces survey sampling and challenges in the era of big data, as well as how sampling techniques can still be useful for handling big data. It outlines reservoir sampling and two methods to improve it: balanced reservoir sampling and stratified reservoir sampling. A simulation study is described to compare the performance of these reservoir sampling methods.
International journal of engineering and mathematical modelling vol2 no1_2015_1IJEMM
Our efforts are mostly concentrated on improving the convergence rate of the numerical procedures both from the viewpoint of cost-efficiency and accuracy by handling the parametrization of the shape to be optimized. We employ nested parameterization supports of either shape, or shape deformation, and the classical process of degree elevation resulting in exact geometrical data transfer from coarse to fine representations. The algorithms mimick classical multigrid strategies and are found very effective in terms of convergence acceleration. In this paper, we analyse and demonstrate the efficiency of the two-level correction algorithm which is the basic block of a more general miltilevel strategy.
BlUP and BLUE- REML of linear mixed modelKyusonLim
This document discusses linear mixed models and the estimation methods BLUP and BLUE. It provides an introduction to random and fixed effects, as well as the mixed model equations used to derive BLUP and BLUE simultaneously. BLUP provides the best linear unbiased predictions of random effects, while BLUE gives the best linear unbiased estimates of fixed effects. The document also provides an example using the orthodontic growth data set to demonstrate fitting a linear mixed model and estimate variance components with REML.
This document summarizes key concepts from Chapter 2 of a book on statistical methods for handling incomplete data. It introduces the likelihood-based approach and defines key terms like the likelihood function, maximum likelihood estimator, Fisher information, and missing at random. The chapter also provides examples of observed likelihood functions for censored regression and survival analysis models with missing data.
Regularization and variable selection via elastic netKyusonLim
The document summarizes the Elastic Net regularization method for variable selection in datasets with more predictors than observations (p > n). It describes how the Elastic Net overcomes limitations of LASSO and Ridge Regression by performing automatic variable selection, continuous shrinkage, and selecting groups of correlated predictors. The Naive Elastic Net formulation is presented, along with how it relates to LASSO and Ridge penalties. Computational details of the Elastic Net, including the LARS-EN algorithm and simulations, are discussed.
IJCER (www.ijceronline.com) International Journal of computational Engineerin...ijceronline
This document discusses period-doubling bifurcations and the route to chaos in an area-preserving discrete dynamical system. It highlights two objectives: 1) Evaluating period-doubling bifurcations using computer programs as the system parameter p is varied, obtaining the Feigenbaum universal constant and accumulation point beyond which chaos occurs; and 2) Confirming periodic behaviors by plotting time series graphs. The document provides background on Feigenbaum universality and describes the specific area-preserving map studied. It outlines the numerical method used to obtain periodic points via Newton's recurrence formula and describes applying this to the map's Jacobian matrix.
This document discusses statistical analysis of experimental data. It outlines concepts like accuracy, precision, error, and data rejection. Accuracy refers to how close a measurement is to the true value, while precision refers to the agreement between multiple measurements. There are different types of errors like systematic and random errors. Various statistical metrics are presented for evaluating accuracy and precision, including mean, standard deviation, percent error and Gaussian distributions. Guidelines for identifying and rejecting outlier data points using the Q-test are also provided.
Bayesian Inferences for Two Parameter Weibull DistributionIOSR Journals
This document discusses Bayesian inference methods for estimating the parameters of a two-parameter Weibull distribution. It begins by introducing the Weibull distribution and defining its probability density function. Maximum likelihood estimation is derived for the scale and shape parameters. Approximate Bayesian methods are then explored, including the Lindley and Laplace approximations, to obtain expressions for the marginal posterior densities since closed-form solutions are not available. The results indicate that the posterior variances for the scale parameter obtained with the Laplace method are smaller than those from the Lindley approximation or asymptotic variances of the maximum likelihood estimates.
The document discusses Bayesian networks and causal discovery methods. It provides definitions and examples of key concepts in Bayesian networks including directed acyclic graphs (DAGs), Markov blankets, and the Markov condition. It also describes different approaches to learning Bayesian network structures, including constraint-based methods such as the PC algorithm and score-based methods like greedy hill climbing. Causal discovery from data aims to infer causal relationships between variables using techniques like conditional independence tests on Bayesian networks.
Kriging is an optimal interpolation technique that estimates values at unmeasured locations based on measured data from nearby locations. It assigns weights to surrounding data points based on their distances and spatial covariance, accounting for clustering of data points. Simple kriging assumes a known mean and estimates values as a weighted average of residuals from the mean at nearby locations, with weights chosen to minimize the estimation variance. The document provides an example of applying simple kriging to estimate porosity using six nearby data points.
This document discusses regression with frailty in survival analysis using the Cox proportional hazards model. It introduces survival analysis concepts like the hazard function and survival function. It then describes how to incorporate frailty, a random effect, into the Cox model to account for clustering in survival times. The Newton-Raphson method is used to estimate model parameters by maximizing the penalized partial likelihood. A simulation study applies this approach to data on infections in kidney patients.
This document proposes an approximate Bayesian inference method for estimating propensity scores under nonresponse. It involves treating the estimating equations as random variables and assigning a prior distribution to the transformed parameters. Samples are drawn from the posterior distribution of the parameters given the observed data to make inferences. The method is shown to be asymptotically consistent and confidence regions can be constructed from the posterior samples. Extensions are discussed to incorporate auxiliary variables and perform Bayesian model selection by assigning a spike-and-slab prior over the model parameters.
This document summarizes an exercise involving calculating the inverse of square matrices in three ways: analytically, using LU decomposition, and singular value decomposition. It finds that analytical calculation becomes impractically slow for matrices larger than order 13, while LU decomposition and SVD using GNU Scientific Library functions can calculate the inverse of matrices up to order 350 in under 20 seconds. SVD is found to be slightly more efficient than LU decomposition for higher order matrices. Accuracy is also compared when the input matrix is close to singular, finding SVD returns the most accurate inverse.
A proposed nth – order jackknife ridge estimator for linear regression designsAlexander Decker
This document proposes and evaluates an nth-order Jackknife Ridge estimator for linear regression to address issues of collinearity. Several existing estimators are introduced, including Ordinary Least Squares, Ridge, Jackknife Ridge, and second-order Jackknife Ridge. The document then proposes a general nth-order Jackknife Ridge estimator and develops it mathematically. Simulation results using five sample matrices show that higher-order Jackknife Ridge estimators have lower bias than lower-order estimators, with the bias decreasing by a factor of 10-4 with each increasing order. Higher-order estimators also have smaller variances than Ordinary Least Squares. The nth-order estimator approach provides a way to minimize bias and
The document summarizes an exercise on using random number generators and Monte Carlo techniques. It compares two methods - analytical and accept/reject - for generating a sinusoidal distribution of random numbers from a uniform distribution. Both methods produced very good fits to a sinusoidal distribution when a large number of random numbers were tested, with R^2 values close to 1. The accept/reject method may be more generally applicable since problems can arise with evaluating and inverting aspects of the analytical method. The exercise also used Monte Carlo techniques to model detector resolution in a nuclear physics experiment. Accounting for detector resolution made the calculated distribution less precise and less representative of the true distribution, showing the need for improved detector resolution.
This document summarizes an exercise on using computational methods to solve partial differential equations (PDEs). It first discusses using relaxation techniques to solve the Laplace equation for modeling electric fields around parallel plate capacitors. The Jacobi method was used to iteratively calculate potentials on a grid until convergence within a specified tolerance. Smaller tolerances and larger grids required more iterations to converge. Longer capacitor plates produced more uniform fields resembling the theoretical infinite case. The document demonstrates how computational methods can effectively solve important physical problems modeled by PDEs.
Estimation of Parameters and Missing Responses In Second Order Response Surfa...inventionjournals
This article is an attempt to explore missing responses in second order response design model
using Expected Maximization algorithm with and without imposing restrictions on the design matrix towards or
thogonality are derived and are implemented with suitable examples. The properties of estimated parameters
and estimated responses are also studied and findings are presented in detail at the end of the study.
The document discusses methods for performing spatial statistics on large datasets. Standard maximum likelihood estimation is computationally infeasible for datasets with tens of thousands of observations due to the need to compute and store large covariance matrices. The document outlines several approximation methods that can accommodate large datasets, including variogram fitting, pairwise likelihood approximations, independent block approximations, tapering of the covariance function, low-rank approximations using basis functions, and approximations based on stochastic partial differential equations. These methods allow inference for large spatial datasets by avoiding direct computation and storage of large covariance matrices.
IJCER (www.ijceronline.com) International Journal of computational Engineerin...ijceronline
This document summarizes research on controlling unstable periodic orbits that coexist with a strange attractor using periodic proportional pulses. The technique is demonstrated on the nonlinear dynamics model xn+1 = axn - bxn2, which exhibits chaos for certain parameter values. Control curves are plotted and used to stabilize periodic orbits of periods 1, 2, 3 and 4 by applying proportional pulses every q iterations with a kicking factor λ. The results show this technique can stabilize unstable periodic orbits embedded in the chaotic attractor.
Lesson 19: The Mean Value Theorem (Section 021 handout)Matthew Leingang
The Mean Value Theorem is the most important theorem in calculus. It is the first theorem which allows us to infer information about a function from information about its derivative. From the MVT we can derive tests for the monotonicity (increase or decrease) and concavity of a function.
This two page document contains timestamps from October 10th, with the first page noting a time of 10:01 AM and the second page noting a time of 2:37 PM. The document appears to be recording times of activity on those dates but provides no other context or details about the content.
International journal of engineering and mathematical modelling vol2 no1_2015_1IJEMM
Our efforts are mostly concentrated on improving the convergence rate of the numerical procedures both from the viewpoint of cost-efficiency and accuracy by handling the parametrization of the shape to be optimized. We employ nested parameterization supports of either shape, or shape deformation, and the classical process of degree elevation resulting in exact geometrical data transfer from coarse to fine representations. The algorithms mimick classical multigrid strategies and are found very effective in terms of convergence acceleration. In this paper, we analyse and demonstrate the efficiency of the two-level correction algorithm which is the basic block of a more general miltilevel strategy.
BlUP and BLUE- REML of linear mixed modelKyusonLim
This document discusses linear mixed models and the estimation methods BLUP and BLUE. It provides an introduction to random and fixed effects, as well as the mixed model equations used to derive BLUP and BLUE simultaneously. BLUP provides the best linear unbiased predictions of random effects, while BLUE gives the best linear unbiased estimates of fixed effects. The document also provides an example using the orthodontic growth data set to demonstrate fitting a linear mixed model and estimate variance components with REML.
This document summarizes key concepts from Chapter 2 of a book on statistical methods for handling incomplete data. It introduces the likelihood-based approach and defines key terms like the likelihood function, maximum likelihood estimator, Fisher information, and missing at random. The chapter also provides examples of observed likelihood functions for censored regression and survival analysis models with missing data.
Regularization and variable selection via elastic netKyusonLim
The document summarizes the Elastic Net regularization method for variable selection in datasets with more predictors than observations (p > n). It describes how the Elastic Net overcomes limitations of LASSO and Ridge Regression by performing automatic variable selection, continuous shrinkage, and selecting groups of correlated predictors. The Naive Elastic Net formulation is presented, along with how it relates to LASSO and Ridge penalties. Computational details of the Elastic Net, including the LARS-EN algorithm and simulations, are discussed.
IJCER (www.ijceronline.com) International Journal of computational Engineerin...ijceronline
This document discusses period-doubling bifurcations and the route to chaos in an area-preserving discrete dynamical system. It highlights two objectives: 1) Evaluating period-doubling bifurcations using computer programs as the system parameter p is varied, obtaining the Feigenbaum universal constant and accumulation point beyond which chaos occurs; and 2) Confirming periodic behaviors by plotting time series graphs. The document provides background on Feigenbaum universality and describes the specific area-preserving map studied. It outlines the numerical method used to obtain periodic points via Newton's recurrence formula and describes applying this to the map's Jacobian matrix.
This document discusses statistical analysis of experimental data. It outlines concepts like accuracy, precision, error, and data rejection. Accuracy refers to how close a measurement is to the true value, while precision refers to the agreement between multiple measurements. There are different types of errors like systematic and random errors. Various statistical metrics are presented for evaluating accuracy and precision, including mean, standard deviation, percent error and Gaussian distributions. Guidelines for identifying and rejecting outlier data points using the Q-test are also provided.
Bayesian Inferences for Two Parameter Weibull DistributionIOSR Journals
This document discusses Bayesian inference methods for estimating the parameters of a two-parameter Weibull distribution. It begins by introducing the Weibull distribution and defining its probability density function. Maximum likelihood estimation is derived for the scale and shape parameters. Approximate Bayesian methods are then explored, including the Lindley and Laplace approximations, to obtain expressions for the marginal posterior densities since closed-form solutions are not available. The results indicate that the posterior variances for the scale parameter obtained with the Laplace method are smaller than those from the Lindley approximation or asymptotic variances of the maximum likelihood estimates.
The document discusses Bayesian networks and causal discovery methods. It provides definitions and examples of key concepts in Bayesian networks including directed acyclic graphs (DAGs), Markov blankets, and the Markov condition. It also describes different approaches to learning Bayesian network structures, including constraint-based methods such as the PC algorithm and score-based methods like greedy hill climbing. Causal discovery from data aims to infer causal relationships between variables using techniques like conditional independence tests on Bayesian networks.
Kriging is an optimal interpolation technique that estimates values at unmeasured locations based on measured data from nearby locations. It assigns weights to surrounding data points based on their distances and spatial covariance, accounting for clustering of data points. Simple kriging assumes a known mean and estimates values as a weighted average of residuals from the mean at nearby locations, with weights chosen to minimize the estimation variance. The document provides an example of applying simple kriging to estimate porosity using six nearby data points.
This document discusses regression with frailty in survival analysis using the Cox proportional hazards model. It introduces survival analysis concepts like the hazard function and survival function. It then describes how to incorporate frailty, a random effect, into the Cox model to account for clustering in survival times. The Newton-Raphson method is used to estimate model parameters by maximizing the penalized partial likelihood. A simulation study applies this approach to data on infections in kidney patients.
This document proposes an approximate Bayesian inference method for estimating propensity scores under nonresponse. It involves treating the estimating equations as random variables and assigning a prior distribution to the transformed parameters. Samples are drawn from the posterior distribution of the parameters given the observed data to make inferences. The method is shown to be asymptotically consistent and confidence regions can be constructed from the posterior samples. Extensions are discussed to incorporate auxiliary variables and perform Bayesian model selection by assigning a spike-and-slab prior over the model parameters.
This document summarizes an exercise involving calculating the inverse of square matrices in three ways: analytically, using LU decomposition, and singular value decomposition. It finds that analytical calculation becomes impractically slow for matrices larger than order 13, while LU decomposition and SVD using GNU Scientific Library functions can calculate the inverse of matrices up to order 350 in under 20 seconds. SVD is found to be slightly more efficient than LU decomposition for higher order matrices. Accuracy is also compared when the input matrix is close to singular, finding SVD returns the most accurate inverse.
A proposed nth – order jackknife ridge estimator for linear regression designsAlexander Decker
This document proposes and evaluates an nth-order Jackknife Ridge estimator for linear regression to address issues of collinearity. Several existing estimators are introduced, including Ordinary Least Squares, Ridge, Jackknife Ridge, and second-order Jackknife Ridge. The document then proposes a general nth-order Jackknife Ridge estimator and develops it mathematically. Simulation results using five sample matrices show that higher-order Jackknife Ridge estimators have lower bias than lower-order estimators, with the bias decreasing by a factor of 10-4 with each increasing order. Higher-order estimators also have smaller variances than Ordinary Least Squares. The nth-order estimator approach provides a way to minimize bias and
The document summarizes an exercise on using random number generators and Monte Carlo techniques. It compares two methods - analytical and accept/reject - for generating a sinusoidal distribution of random numbers from a uniform distribution. Both methods produced very good fits to a sinusoidal distribution when a large number of random numbers were tested, with R^2 values close to 1. The accept/reject method may be more generally applicable since problems can arise with evaluating and inverting aspects of the analytical method. The exercise also used Monte Carlo techniques to model detector resolution in a nuclear physics experiment. Accounting for detector resolution made the calculated distribution less precise and less representative of the true distribution, showing the need for improved detector resolution.
This document summarizes an exercise on using computational methods to solve partial differential equations (PDEs). It first discusses using relaxation techniques to solve the Laplace equation for modeling electric fields around parallel plate capacitors. The Jacobi method was used to iteratively calculate potentials on a grid until convergence within a specified tolerance. Smaller tolerances and larger grids required more iterations to converge. Longer capacitor plates produced more uniform fields resembling the theoretical infinite case. The document demonstrates how computational methods can effectively solve important physical problems modeled by PDEs.
Estimation of Parameters and Missing Responses In Second Order Response Surfa...inventionjournals
This article is an attempt to explore missing responses in second order response design model
using Expected Maximization algorithm with and without imposing restrictions on the design matrix towards or
thogonality are derived and are implemented with suitable examples. The properties of estimated parameters
and estimated responses are also studied and findings are presented in detail at the end of the study.
The document discusses methods for performing spatial statistics on large datasets. Standard maximum likelihood estimation is computationally infeasible for datasets with tens of thousands of observations due to the need to compute and store large covariance matrices. The document outlines several approximation methods that can accommodate large datasets, including variogram fitting, pairwise likelihood approximations, independent block approximations, tapering of the covariance function, low-rank approximations using basis functions, and approximations based on stochastic partial differential equations. These methods allow inference for large spatial datasets by avoiding direct computation and storage of large covariance matrices.
IJCER (www.ijceronline.com) International Journal of computational Engineerin...ijceronline
This document summarizes research on controlling unstable periodic orbits that coexist with a strange attractor using periodic proportional pulses. The technique is demonstrated on the nonlinear dynamics model xn+1 = axn - bxn2, which exhibits chaos for certain parameter values. Control curves are plotted and used to stabilize periodic orbits of periods 1, 2, 3 and 4 by applying proportional pulses every q iterations with a kicking factor λ. The results show this technique can stabilize unstable periodic orbits embedded in the chaotic attractor.
Lesson 19: The Mean Value Theorem (Section 021 handout)Matthew Leingang
The Mean Value Theorem is the most important theorem in calculus. It is the first theorem which allows us to infer information about a function from information about its derivative. From the MVT we can derive tests for the monotonicity (increase or decrease) and concavity of a function.
This two page document contains timestamps from October 10th, with the first page noting a time of 10:01 AM and the second page noting a time of 2:37 PM. The document appears to be recording times of activity on those dates but provides no other context or details about the content.
Dokumen tersebut merangkum hasil analisis regresi tak linier untuk memprediksi kecepatan rata-rata pelari berusia di atas 70 tahun berdasarkan data jarak dan kecepatan larinya. Persamaan regresi yang didapat adalah v̂ = 4,60s-0,06 dengan nilai R2 sebesar 0,81, yang menunjukkan hubungan antara jarak dan kecepatan sebesar 81%. Prediksi kecepatan rata-rata pelari berusia di atas 70 t
Ringkasan dokumen tersebut adalah:
1. Dokumen tersebut membahas tentang analisis regresi linier sederhana antara satu variabel independen dengan satu variabel dependen.
2. Persamaan umum regresi linier adalah Y = a + bX, dimana a adalah konstanta dan b adalah koefisien regresi.
3. Dokumen tersebut juga menjelaskan cara menghitung nilai a dan b, serta uji signifikansi dan linearitas model regresi menggunakan analisis
This document discusses nonlinear regression analysis using R. It provides an example of using the nls function to fit nonlinear curves to data. Specifically, it generates random y-data using an exponential decay function of t, plots the data, and performs nonlinear regression to estimate the parameters of the underlying exponential decay model. It also discusses fitting a Gompertz growth curve model to height data using nls. The output is analyzed to test if parameter estimates are statistically significant. Finally, it briefly introduces self-starting nonlinear regression models in R that do not require initial parameter values.
Panduan Lengkap Analisis Statistika dengan Aplikasi SPSSMuliadin Forester
Dokumen tersebut memberikan panduan lengkap mengenai analisis statistika menggunakan perangkat lunak SPSS (Statistical Package for Social Science), mencakup uji asumsi klasik seperti uji normalitas, multikolinearitas, heteroskedastisitas, autokorelasi, dan linearitas; serta analisis regresi linear berganda dan tabel statistik. Panduan ini disadur dari beberapa situs web dan disederhanakan untuk kemudahan pemahaman.
Auto Regressive Process (1) with Change Point: Bayesian ApprochIJRESJOURNAL
Abstract : Here we consider first order autoregressive process with changing autoregressive coefficient at some point of time m. This is called change point inference problem. For Bayes estimation of m and autoregressive coefficient we used MHRW (Metropolis Hasting Random Walk) algorithm and Gibbs sampling. The effects of prior information on the Bayes estimates are also studied.
This document presents a nonparametric approach to multiple regression that uses ranks instead of raw values for both the dependent and independent variables. The key points are:
1. It develops a nonparametric multiple regression model using the ranks of observations on the dependent variable and ranks of observations on the independent variables.
2. The method of least squares is applied to the rank-based model to obtain estimates of the regression coefficients.
3. Prediction equations are presented that allow predicting dependent variable ranks based on independent variable ranks.
This document provides an overview of linear regression analysis and correlation. It defines key terms like predictor, predictand, R-squared value, and significance tests. It explains how to perform a least squares linear fit to determine the regression coefficients that minimize the error between predicted and actual values. It also discusses testing the significance of regression coefficients and how to determine if adding more predictors improves the regression model. Finally, it introduces concepts like multiple regression, Fourier analysis, and the minimum correlation needed to benefit from adding additional predictors.
Regression analysis is used to model relationships between variables. Simple linear regression involves modeling the relationship between a single independent variable and dependent variable. The regression equation estimates the dependent variable (y) as a linear function of the independent variable (x). The parameters β0 and β1 are estimated using the method of least squares. The coefficient of determination (r2) measures how well the regression line fits the data. Additional tests like the t-test, confidence intervals, and F-test are used to test if the independent variable significantly predicts the dependent variable. While these tests can indicate a statistically significant relationship, they do not prove causation.
This document discusses algorithms for computing the Kolmogorov-Smirnov distribution, which is used to measure goodness of fit between empirical and theoretical distributions. It describes existing algorithms that are either fast but unreliable or precise but slow. The authors propose a new algorithm that uses different approximations depending on sample size and test statistic value, to provide fast and reliable computation of both the distribution and its complement. Their C program implements this approach using multiple methods, including an exact but slow recursion formula and faster but less precise approximations for large samples.
International Journal of Mathematics and Statistics Invention (IJMSI) is an international journal intended for professionals and researchers in all fields of computer science and electronics. IJMSI publishes research articles and reviews within the whole field Mathematics and Statistics, new teaching methods, assessment, validation and the impact of new technologies and it will continue to provide information on the latest trends and developments in this ever-expanding subject. The publications of papers are selected through double peer reviewed to ensure originality, relevance, and readability. The articles published in our journal can be accessed online.
Numerical treatments to nonlocal Fredholm –Volterra integral equation with co...iosrjce
IOSR Journal of Mathematics(IOSR-JM) is a double blind peer reviewed International Journal that provides rapid publication (within a month) of articles in all areas of mathemetics and its applications. The journal welcomes publications of high quality papers on theoretical developments and practical applications in mathematics. Original research papers, state-of-the-art reviews, and high quality technical notes are invited for publications.
This document discusses regression analysis techniques. Regression analysis is used to model the relationship between a dependent variable (Y) and one or more independent variables (X1, X2, etc). Simple linear regression involves one independent variable, while multiple linear regression involves two or more independent variables. The key assumptions of linear regression are outlined. Methods for estimating regression coefficients using least squares and testing the significance of regression coefficients and the overall regression model are also described. An example application involving modeling personal pollutant exposure (Y) based on hours outdoors (X1) and home pollutant levels (X2) is provided.
Use of the correlation coefficient as a measure of effectiveness of a scoring...Wajih Alaiyan
The document discusses using the correlation coefficient to measure the effectiveness of machine scoring systems compared to human scoring. It provides three applications of using the correlation coefficient: 1) Assigning machine scores that match the expected value of the human score ranking, 2) Assigning machine scores that match the expected human score within clusters of essays, 3) Using instrumental variables to estimate machine scores in a way that maximizes the correlation with human scores. The analysis shows that maximizing the correlation coefficient provides a justified way to measure scoring system effectiveness.
1. The document discusses statistical estimation and properties of estimators such as bias, variance, consistency, and asymptotic normality.
2. Key concepts covered include unbiasedness, mean squared error, relative efficiency, sufficiency, and properties of estimators like consistency, asymptotic unbiasedness, and best asymptotic normality.
3. Examples are provided to illustrate theoretical estimators for parameters like the variance of a distribution or coefficients in a linear regression model.
First post of a Data Science blog about Linear Regression using Matlab.
For more information, please visit:
http://datascienceinsights.blogspot.com.br/
https://github.com/tadeuferreirajr/MachineLearning
This document discusses applying Newton's method and continuation methods to reduce optimal control problems to boundary value problems. It begins by introducing Newton's method and the implicit function theorem, which defines a curve of solutions. It then discusses the continuous analogue of Newton's method, which transforms the equations into an initial value problem that is integrated. Finally, it shows how an optimal control problem can be reduced to a boundary value problem by applying the Pontryagin maximum principle, yielding equations that can be solved using continuation methods.
This document presents an application of the differential transform method to solve nonlinear differential equations. The differential transform method can decompose nonlinear terms in a differential equation into a power series, allowing iterative calculation of the solution coefficients. Three examples are provided to demonstrate solving first-order nonlinear differential equations using this method. The power series solutions converged rapidly to the exact solutions. The differential transform method is concluded to be a useful technique for solving both linear and nonlinear differential equations.
This document presents an application of the differential transform method to solve nonlinear differential equations. The differential transform method can decompose nonlinear terms in a differential equation into a power series, allowing iterative calculation of the solution coefficients. Three examples are provided to demonstrate solving first-order nonlinear differential equations using this method. The power series solutions converged rapidly to the exact solutions. The differential transform method is concluded to be a useful technique for solving both linear and nonlinear differential equations.
This document provides an overview of limits and continuity in mathematics. It begins by introducing functions of discrete and continuous variables, and defines the concept of a limit for sequences. A limit describes the long-term behavior of a function as the input approaches some value. The document then discusses finding the limit of various sequences, such as sequences that approach 0. It also introduces subsequences and defines the algebra of limits of sequences, such as how the limit behaves under operations like addition, subtraction, multiplication and division. The overall goal is to understand the concept of limit and apply it to determine if functions are continuous or not.
The document discusses simple linear regression and correlation methods. It defines deterministic and probabilistic models for describing the relationship between two variables. A simple linear regression model assumes a population regression line with intercept a and slope b, where observations may deviate from the line by some random error e. Key assumptions of the model are that e has a normal distribution with mean 0 and constant variance across values of x, and errors are independent. The slope b estimates the average change in y per unit change in x.
In this paper, we propose four algorithms for L1 norm computation of regression parameters, where two of them are more efficient for simple and multiple regression models. However, we start with restricted simple linear regression and corresponding derivation and computation of the weighted median problem. In this respect, a computing function is coded. With discussion on the m parameters model, we continue to expand the algorithm to include unrestricted simple linear regression, and two crude and efficient algorithms are proposed. The procedures are then generalized to the m parameters model by presenting two new algorithms, where the algorithm 4 is selected as more efficient. Various properties of these algorithms are discussed.
Bayesian Estimation for Missing Values in Latin Square Designinventionjournals
International Journal of Mathematics and Statistics Invention (IJMSI) is an international journal intended for professionals and researchers in all fields of computer science and electronics. IJMSI publishes research articles and reviews within the whole field Mathematics and Statistics, new teaching methods, assessment, validation and the impact of new technologies and it will continue to provide information on the latest trends and developments in this ever-expanding subject. The publications of papers are selected through double peer reviewed to ensure originality, relevance, and readability. The articles published in our journal can be accessed online.
This document presents a numerical approach to solving ordinary differential equations. It discusses solving boundary value problems by converting them to initial value problems using the shooting method. It describes solving the initial value problems using the Runge-Kutta method. It provides examples of applying these methods to solve various second order differential equations. The conclusion discusses solving linear and nonlinear differential equations with numerical methods. The future scope section proposes extensions like solving other boundary condition types, performing error and stability analysis, and applying the methods to nonlinear problems.
Similar to A Bootstrap Approach to Error-Reduction of Nonlinear Regression Parameters Estimation (20)
This document provides a technical review of secure banking using RSA and AES encryption methodologies. It discusses how RSA and AES are commonly used encryption standards for secure data transmission between ATMs and bank servers. The document first provides background on ATM security measures and risks of attacks. It then reviews related work analyzing encryption techniques. The document proposes using a one-time password in addition to a PIN for ATM authentication. It concludes that implementing encryption standards like RSA and AES can make transactions more secure and build trust in online banking.
This document analyzes the performance of various modulation schemes for achieving energy efficient communication over fading channels in wireless sensor networks. It finds that for long transmission distances, low-order modulations like BPSK are optimal due to their lower SNR requirements. However, as transmission distance decreases, higher-order modulations like 16-QAM and 64-QAM become more optimal since they can transmit more bits per symbol, outweighing their higher SNR needs. Simulations show lifetime extensions up to 550% are possible in short-range networks by using higher-order modulations instead of just BPSK. The optimal modulation depends on transmission distance and balancing the energy used by electronic components versus power amplifiers.
This document provides a review of mobility management techniques in vehicular ad hoc networks (VANETs). It discusses three modes of communication in VANETs: vehicle-to-infrastructure (V2I), vehicle-to-vehicle (V2V), and hybrid vehicle (HV) communication. For each communication mode, different mobility management schemes are required due to their unique characteristics. The document also discusses mobility management challenges in VANETs and outlines some open research issues in improving mobility management for seamless communication in these dynamic networks.
This document provides a review of different techniques for segmenting brain MRI images to detect tumors. It compares the K-means and Fuzzy C-means clustering algorithms. K-means is an exclusive clustering algorithm that groups data points into distinct clusters, while Fuzzy C-means is an overlapping clustering algorithm that allows data points to belong to multiple clusters. The document finds that Fuzzy C-means requires more time for brain tumor detection compared to other methods like hierarchical clustering or K-means. It also reviews related work applying these clustering algorithms to segment brain MRI images.
1) The document simulates and compares the performance of AODV and DSDV routing protocols in a mobile ad hoc network under three conditions: when users are fixed, when users move towards the base station, and when users move away from the base station.
2) The results show that both protocols have higher packet delivery and lower packet loss when users are either fixed or moving towards the base station, since signal strength is better in those scenarios. Performance degrades when users move away from the base station due to weaker signals.
3) AODV generally has better performance than DSDV, with higher throughput and packet delivery rates observed across the different user mobility conditions.
This document describes the design and implementation of 4-bit QPSK and 256-bit QAM modulation techniques using MATLAB. It compares the two techniques based on SNR, BER, and efficiency. The key steps of implementing each technique in MATLAB are outlined, including generating random bits, modulation, adding noise, and measuring BER. Simulation results show scatter plots and eye diagrams of the modulated signals. A table compares the results, showing that 256-bit QAM provides better performance than 4-bit QPSK. The document concludes that QAM modulation is more effective for digital transmission systems.
The document proposes a hybrid technique using Anisotropic Scale Invariant Feature Transform (A-SIFT) and Robust Ensemble Support Vector Machine (RESVM) to accurately identify faces in images. A-SIFT improves upon traditional SIFT by applying anisotropic scaling to extract richer directional keypoints. Keypoints are processed with RESVM and hypothesis testing to increase accuracy above 95% by repeatedly reprocessing images until the threshold is met. The technique was tested on similar and different facial images and achieved better results than SIFT in retrieval time and reduced keypoints.
This document studies the effects of dielectric superstrate thickness on microstrip patch antenna parameters. Three types of probes-fed patch antennas (rectangular, circular, and square) were designed to operate at 2.4 GHz using Arlondiclad 880 substrate. The antennas were tested with and without an Arlondiclad 880 superstrate of varying thicknesses. It was found that adding a superstrate slightly degraded performance by lowering the resonant frequency and increasing return loss and VSWR, while decreasing bandwidth and gain. Specifically, increasing the superstrate thickness or dielectric constant resulted in greater changes to the antenna parameters.
This document describes a wireless environment monitoring system that utilizes soil energy as a sustainable power source for wireless sensors. The system uses a microbial fuel cell to generate electricity from the microbial activity in soil. Two microbial fuel cells were created using different soil types and various additives to produce different current and voltage outputs. An electronic circuit was designed on a printed circuit board with components like a microcontroller and ZigBee transceiver. Sensors for temperature and humidity were connected to the circuit to monitor the environment wirelessly. The system provides a low-cost way to power remote sensors without needing battery replacement and avoids the high costs of wiring a power source.
1) The document proposes a model for a frequency tunable inverted-F antenna that uses ferrite material.
2) The resonant frequency of the antenna can be significantly shifted from 2.41GHz to 3.15GHz, a 31% shift, by increasing the static magnetic field placed on the ferrite material.
3) Altering the permeability of the ferrite allows tuning of the antenna's resonant frequency without changing the physical dimensions, providing flexibility to operate over a wide frequency range.
This document summarizes a research paper that presents a speech enhancement method using stationary wavelet transform. The method first classifies speech into voiced, unvoiced, and silence regions based on short-time energy. It then applies different thresholding techniques to the wavelet coefficients of each region - modified hard thresholding for voiced speech, semi-soft thresholding for unvoiced speech, and setting coefficients to zero for silence. Experimental results using speech from the TIMIT database corrupted with white Gaussian noise at various SNR levels show improved performance over other popular denoising methods.
This document reviews the design of an energy-optimized wireless sensor node that encrypts data for transmission. It discusses how sensing schemes that group nodes into clusters and transmit aggregated data can reduce energy consumption compared to individual node transmissions. The proposed node design calculates the minimum transmission power needed based on received signal strength and uses a periodic sleep/wake cycle to optimize energy when not sensing or transmitting. It aims to encrypt data at both the node and network level to further optimize energy usage for wireless communication.
This document discusses group consumption modes. It analyzes factors that impact group consumption, including external environmental factors like technological developments enabling new forms of online and offline interactions, as well as internal motivational factors at both the group and individual level. The document then proposes that group consumption modes can be divided into four types based on two dimensions: vertical (group relationship intensity) and horizontal (consumption action period). These four types are instrument-oriented, information-oriented, enjoyment-oriented, and relationship-oriented consumption modes. Finally, the document notes that consumption modes are dynamic and can evolve over time.
The document summarizes a study of different microstrip patch antenna configurations with slotted ground planes. Three antenna designs were proposed and their performance evaluated through simulation: a conventional square patch, an elliptical patch, and a star-shaped patch. All antennas were mounted on an FR4 substrate. The effects of adding different slot patterns to the ground plane on resonance frequency, bandwidth, gain and efficiency were analyzed parametrically. Key findings were that reshaping the patch and adding slots increased bandwidth and shifted resonance frequency. The elliptical and star patches in particular performed better than the conventional design. Three antenna configurations were selected for fabrication and measurement based on the simulations: a conventional patch with a slot under the patch, an elliptical patch with slots
1) The document describes a study conducted to improve call drop rates in a GSM network through RF optimization.
2) Drive testing was performed before and after optimization using TEMS software to record network parameters like RxLevel, RxQuality, and events.
3) Analysis found call drops were occurring due to issues like handover failures between sectors, interference from adjacent channels, and overshooting due to antenna tilt.
4) Corrective actions taken included defining neighbors between sectors, adjusting frequencies to reduce interference, and lowering the mechanical tilt of an antenna.
5) Post-optimization drive testing showed improvements in RxLevel, RxQuality, and a reduction in dropped calls.
This document describes the design of an intelligent autonomous wheeled robot that uses RF transmission for communication. The robot has two modes - automatic mode where it can make its own decisions, and user control mode where a user can control it remotely. It is designed using a microcontroller and can perform tasks like object recognition using computer vision and color detection in MATLAB, as well as wall painting using pneumatic systems. The robot's movement is controlled by DC motors and it uses sensors like ultrasonic sensors and gas sensors to navigate autonomously. RF transmission allows communication between the robot and a remote control unit. The overall aim is to develop a low-cost robotic system for industrial applications like material handling.
This document reviews cryptography techniques to secure the Ad-hoc On-Demand Distance Vector (AODV) routing protocol in mobile ad-hoc networks. It discusses various types of attacks on AODV like impersonation, denial of service, eavesdropping, black hole attacks, wormhole attacks, and Sybil attacks. It then proposes using the RC6 cryptography algorithm to secure AODV by encrypting data packets and detecting and removing malicious nodes launching black hole attacks. Simulation results show that after applying RC6, the packet delivery ratio and throughput of AODV increase while delay decreases, improving the security and performance of the network under attack.
The document describes a proposed modification to the conventional Booth multiplier that aims to increase its speed by applying concepts from Vedic mathematics. Specifically, it utilizes the Urdhva Tiryakbhyam formula to generate all partial products concurrently rather than sequentially. The proposed 8x8 bit multiplier was coded in VHDL, simulated, and found to have a path delay 44.35% lower than a conventional Booth multiplier, demonstrating its potential for higher speed.
This document discusses image deblurring techniques. It begins by introducing image restoration and focusing on image deblurring. It then discusses challenges with image deblurring being an ill-posed problem. It reviews existing approaches to screen image deconvolution including estimating point spread functions and iteratively estimating blur kernels and sharp images. The document also discusses handling spatially variant blur and summarizes the relationship between the proposed method and previous work for different blur types. It proposes using color filters in the aperture to exploit parallax cues for segmentation and blur estimation. Finally, it proposes moving the image sensor circularly during exposure to prevent high frequency attenuation from motion blur.
This document describes modeling an adaptive controller for an aircraft roll control system using PID, fuzzy-PID, and genetic algorithm. It begins by introducing the aircraft roll control system and motivation for developing an adaptive controller to minimize errors from noisy analog sensor signals. It then provides the mathematical model of aircraft roll dynamics and describes modeling the real-time flight control system in MATLAB/Simulink. The document evaluates PID, fuzzy-PID, and PID-GA (genetic algorithm) controllers for aircraft roll control and finds that the PID-GA controller delivers the best performance.
Current Ms word generated power point presentation covers major details about the micronuclei test. It's significance and assays to conduct it. It is used to detect the micronuclei formation inside the cells of nearly every multicellular organism. It's formation takes place during chromosomal sepration at metaphase.
The binding of cosmological structures by massless topological defectsSérgio Sacani
Assuming spherical symmetry and weak field, it is shown that if one solves the Poisson equation or the Einstein field
equations sourced by a topological defect, i.e. a singularity of a very specific form, the result is a localized gravitational
field capable of driving flat rotation (i.e. Keplerian circular orbits at a constant speed for all radii) of test masses on a thin
spherical shell without any underlying mass. Moreover, a large-scale structure which exploits this solution by assembling
concentrically a number of such topological defects can establish a flat stellar or galactic rotation curve, and can also deflect
light in the same manner as an equipotential (isothermal) sphere. Thus, the need for dark matter or modified gravity theory is
mitigated, at least in part.
The use of Nauplii and metanauplii artemia in aquaculture (brine shrimp).pptxMAGOTI ERNEST
Although Artemia has been known to man for centuries, its use as a food for the culture of larval organisms apparently began only in the 1930s, when several investigators found that it made an excellent food for newly hatched fish larvae (Litvinenko et al., 2023). As aquaculture developed in the 1960s and ‘70s, the use of Artemia also became more widespread, due both to its convenience and to its nutritional value for larval organisms (Arenas-Pardo et al., 2024). The fact that Artemia dormant cysts can be stored for long periods in cans, and then used as an off-the-shelf food requiring only 24 h of incubation makes them the most convenient, least labor-intensive, live food available for aquaculture (Sorgeloos & Roubach, 2021). The nutritional value of Artemia, especially for marine organisms, is not constant, but varies both geographically and temporally. During the last decade, however, both the causes of Artemia nutritional variability and methods to improve poorquality Artemia have been identified (Loufi et al., 2024).
Brine shrimp (Artemia spp.) are used in marine aquaculture worldwide. Annually, more than 2,000 metric tons of dry cysts are used for cultivation of fish, crustacean, and shellfish larva. Brine shrimp are important to aquaculture because newly hatched brine shrimp nauplii (larvae) provide a food source for many fish fry (Mozanzadeh et al., 2021). Culture and harvesting of brine shrimp eggs represents another aspect of the aquaculture industry. Nauplii and metanauplii of Artemia, commonly known as brine shrimp, play a crucial role in aquaculture due to their nutritional value and suitability as live feed for many aquatic species, particularly in larval stages (Sorgeloos & Roubach, 2021).
ESPP presentation to EU Waste Water Network, 4th June 2024 “EU policies driving nutrient removal and recycling
and the revised UWWTD (Urban Waste Water Treatment Directive)”
The technology uses reclaimed CO₂ as the dyeing medium in a closed loop process. When pressurized, CO₂ becomes supercritical (SC-CO₂). In this state CO₂ has a very high solvent power, allowing the dye to dissolve easily.
The debris of the ‘last major merger’ is dynamically youngSérgio Sacani
The Milky Way’s (MW) inner stellar halo contains an [Fe/H]-rich component with highly eccentric orbits, often referred to as the
‘last major merger.’ Hypotheses for the origin of this component include Gaia-Sausage/Enceladus (GSE), where the progenitor
collided with the MW proto-disc 8–11 Gyr ago, and the Virgo Radial Merger (VRM), where the progenitor collided with the
MW disc within the last 3 Gyr. These two scenarios make different predictions about observable structure in local phase space,
because the morphology of debris depends on how long it has had to phase mix. The recently identified phase-space folds in Gaia
DR3 have positive caustic velocities, making them fundamentally different than the phase-mixed chevrons found in simulations
at late times. Roughly 20 per cent of the stars in the prograde local stellar halo are associated with the observed caustics. Based
on a simple phase-mixing model, the observed number of caustics are consistent with a merger that occurred 1–2 Gyr ago.
We also compare the observed phase-space distribution to FIRE-2 Latte simulations of GSE-like mergers, using a quantitative
measurement of phase mixing (2D causticality). The observed local phase-space distribution best matches the simulated data
1–2 Gyr after collision, and certainly not later than 3 Gyr. This is further evidence that the progenitor of the ‘last major merger’
did not collide with the MW proto-disc at early times, as is thought for the GSE, but instead collided with the MW disc within
the last few Gyr, consistent with the body of work surrounding the VRM.
EWOCS-I: The catalog of X-ray sources in Westerlund 1 from the Extended Weste...Sérgio Sacani
Context. With a mass exceeding several 104 M⊙ and a rich and dense population of massive stars, supermassive young star clusters
represent the most massive star-forming environment that is dominated by the feedback from massive stars and gravitational interactions
among stars.
Aims. In this paper we present the Extended Westerlund 1 and 2 Open Clusters Survey (EWOCS) project, which aims to investigate
the influence of the starburst environment on the formation of stars and planets, and on the evolution of both low and high mass stars.
The primary targets of this project are Westerlund 1 and 2, the closest supermassive star clusters to the Sun.
Methods. The project is based primarily on recent observations conducted with the Chandra and JWST observatories. Specifically,
the Chandra survey of Westerlund 1 consists of 36 new ACIS-I observations, nearly co-pointed, for a total exposure time of 1 Msec.
Additionally, we included 8 archival Chandra/ACIS-S observations. This paper presents the resulting catalog of X-ray sources within
and around Westerlund 1. Sources were detected by combining various existing methods, and photon extraction and source validation
were carried out using the ACIS-Extract software.
Results. The EWOCS X-ray catalog comprises 5963 validated sources out of the 9420 initially provided to ACIS-Extract, reaching a
photon flux threshold of approximately 2 × 10−8 photons cm−2
s
−1
. The X-ray sources exhibit a highly concentrated spatial distribution,
with 1075 sources located within the central 1 arcmin. We have successfully detected X-ray emissions from 126 out of the 166 known
massive stars of the cluster, and we have collected over 71 000 photons from the magnetar CXO J164710.20-455217.
Basics of crystallography, crystal systems, classes and different forms
A Bootstrap Approach to Error-Reduction of Nonlinear Regression Parameters Estimation
1. IOSR Journal of Mathematics (IOSR-JM)
e-ISSN: 2278-5728, p-ISSN: 2319-765X. Volume 11, Issue 1 Ver. III (Jan - Feb. 2015), PP 45-51
www.iosrjournals.org
DOI: 10.9790/5728-11134551 www.iosrjournals.org 45 |Page
A Bootstrap Approach to Error-Reduction of Nonlinear
Regression Parameters Estimation
1
Obiora-Ilouno H. O. And 2
Mbegbu J. I
1
Department of Statistics Nnamdi Azikiwe University, Awka, Nigeria
2
Department of Mathematics University of Benin, Benin City Edo State, Nigeria.
Abstract: In this paper we proposed bootstrap algorithm for the estimation of parameters of the non-linear
regression analysis by cooperating the Gauss- Newton method. We used bootstrapping to provide estimates of
exponential regression coefficients. The computational difficulties that would have encountered in using the
proposed method have been resolved by developing a computer program in R which was used to implement the
algorithm. From the result obtained the bootstrap algorithm yielded a better reduced error sum of squares
SS than the analytical method. With these results, we have a greater confidence in the result obtained.
Keywords: bootstrap, algorithm, exponential, regression, parameters, non-linear, Gauss-Newton.
I. Introduction
Bootstrap resampling as computer based methods is a tools for constructing inferential procedures in
statistical data analysis. The bootstrap approach introduced by Efron (1979) was used initially for estimation of
certain statistics such as the mean and the variability of estimated characteristics of the probability distribution
of a set of observations by providing confidence intervals for the population parameters (Efron, 2003; Efron and
Tibshirani (1998). It is a technique in which sample of size n are obtained randomly with replacement from the
original sample each with 1
n chances of being selected. This is basically to analyze the population by replacing
the unknown distribution function F by the empirical distribution function Fˆ obtained from the sample. (See
Davison and Hinkley, 1997, 2003; Efron, 2003; Bickel and Freedman, 2007; Hall, 2003; Casella, 2003).
Exponential regression is a non-linear regression model often used to measure the growth of a variable,
such as population, GDP etc. An exponential relationship between X and Y exists whenever the dependent
variable Y changes by a constant percentage, as the variable X also changes.
II. Materials And Method
Let jkXXXfY ,,,,,,, 2121 (2.1)
be a non linear regression model with s being the parameters, sX are the predictor variables and the error
term 2
,0~ N independently identically distributed and are uncorrelated.
Equation (2.1) is assumed to be intrinsically nonlinear. Suppose n is number of observations on the Y and sX
, then
niXXXfY ijikiii ,,2,1;,,,,,,, 2121 (2.2)
The n-equation can be written compactly in a matrix notation as
,XfY (2.3)
where
njknnn
k
k
n XXX
XXX
XXX
X
y
y
y
Y
2
1
2
1
21
22212
11211
2
1
,,,
and 0E
The error sum of squares for the nonlinear model is defined as
2
1
,
n
i
ii XfYSQ (2.4)
2. A Bootstrap Approach To Error-Reduction Of Nonlinear Regression Parameters Estimation
DOI: 10.9790/5728-11134551 www.iosrjournals.org 46 |Page
Denoting the least square estimates of ˆby , these minimize the . The least square estimates of
are obtained by differentiating (2.4) with respect to , equate to zero and solve for , this results in J normal
equations:
n
i
i
p
ii JpniXfXfY
1 ˆ
,,2,1,,,2,1;0,ˆ,
(2.5)
In estimating the parameters of nonlinear regression model, we use the Gauss-Newton method based on
Taylor’s series to approximate equation (2.3). Now, considering the function which is the
deterministic component of
niXY iii ,,2,1, (2.6)
Let 0
be the initial approximate value of . Adopting Taylor’s series expansion of about 0
, we
have the linear approximation
0
,,, 00
iii XfXfXf (2.7)
Substituting expressions (2.7) in (2.6) we obtain
JpniXfXfY
J
p
ippi
i
ii ,,2,1,,,2,1,,
1
00
0
Equation (2.8) may be viewed as a linear approximation in a neighborhood of the starting value
Let 00
, ii Xff
00
ppp
JpandniforXfZ i
p
pi ,,2,1,,2,1,,
0
0
(2.8)
Hence, equation (2.8) becomes
J
p
ippiii niZfY
1
000
,,2,1, (2.9)
J
p
ippiii niZfY
1
000
,,2,1, (2.10)
In a matrix form, we have
njJnnn
J
J
nn ZZZ
ZZZ
ZZZ
fY
fY
fY
2
1
0
0
2
0
1
00
2
0
1
0
1
0
22
0
12
0
1
0
21
0
11
0
0
22
0
11
(2.11)
Compactly, equation (2.11) becomes
000
ZfY (2.12)
( )S
ˆ
( , )f X
( , )if X
0
3. A Bootstrap Approach To Error-Reduction Of Nonlinear Regression Parameters Estimation
DOI: 10.9790/5728-11134551 www.iosrjournals.org 47 |Page
where
0 0
11 1
0 0
'' '0 0 0 0 0 0 0 012 2
1 1 2 2 1 1
0 0
1
, , , , ,
J
J
n n J n
n Jn
Z Z
Z Z
Y f Y f Y f Y f Z
Z Z
We obtain the Sum of squares error SS
000000
ZfYZfYSS
0ˆ22 00000
0
ZZZfY
SS
00000 ˆZZZfY
(2.13)
Hence,
100000ˆ
ZZZfY (2.14)
Therefore, the least square estimates of
0
is
001000ˆ fYZZZ
(2.15)
Thus,
00
2
0
1
0 ˆ,,ˆ,ˆˆ
J minimizes the error sum of squares,
n
i
J
p
ppiii ZfYS
1
2
1
000* ˆ (2.16)
Now, the estimates of parameters p of non linear regression (2.1) are
Jpppp ,,2,1;ˆ 001
(2.17)
Iteratively, equation (2.17) reduces to
rrr
rrr
ˆ
ˆ
ˆ
1
11
001
Thus
11 ' 'r r r r r r
Z Z Z Y f
(2.18)
where 1ˆr r r r r
Z Z Z Y f
are the least squares estimates of obtained at the
th
r 1 iterations. The iterative process continues until
r
rr 1
where
5
10
is the error tolerance (See Smith (1998) and Nduka (1999)).
*
S is evaluated after each iteration to see if a reduction in its value has actually been achieved. At the end of
the th
r 1 iteration, we have
n
i
r
p
p
j
r
pi
r
ii
r
ZfYS
1
2
1
* ˆ (2.19)
4. A Bootstrap Approach To Error-Reduction Of Nonlinear Regression Parameters Estimation
DOI: 10.9790/5728-11134551 www.iosrjournals.org 48 |Page
and iteration is stopped if convergence is achieved. The final estimates of the parameters at the end of the
th
r 1 iteration are:
11
2
1
1
1
,,,
r
j
rrr
(2.20)
The Bootstrap Algorithm based on the Resampling Observations for the Estimation of Non- linear
Regression Parameters
jiii ZYW , being the original sample of size n for the resampling, sWi are drawn independently and
identically from a distribution of F.
ni yyyY ,,, 21 is the column vector of the response variables,
jnjjji zzzZ ,,, 21 is the matrix of dimension kn for the predictor variables, where
niandkj ,,2,1,,2,1 .
Let the 11 k vector
jiii ZYW , denote the values associated with th
i nwww ,,, 21
observation sets.
Step 1: Draw bootstrap sample )()(
2
)(
1 ,,, b
n
bb
www of size n with replacement from the observation
giving 1
n probability of each iW value being sampled from the population and label the elements of
each vector
)()()(
, b
ji
b
i
b
i XYW
where. niandkj ,,2,1,,2,1 . From these form the vector for the response variable
)()(
2
)(
1
)(
,,, b
n
bbb
i yyyY and the matrix of the predictor variables
)()()()(
,,, n
jn
b
jz
b
ji
b
ji zzzZ .
Step2: Calculate the least square estimates for nonlinear regression coefficient from the bootstrap sample;
fYZZZ
10ˆ .
Step3: Compute
001 ˆˆˆ using the Gauss-Newton method, the
1ˆ value is treated as the initial value
in the first approximated linear model.
Step4: We return to the second step and again compute sˆ . At each iteration, new sˆ represent increments
that are added to the estimates from the previous iteration according to step 3 and eventually find
2ˆ ,
which is
rrr
toup ˆˆˆˆ 1112
.
Step 5: Stopping Rule; the iteration process continues until
r
rr 1
, where
5
10
, for the
values of 110 ,,, p from the first bootstrap sample
)1(ˆ b
.
Step6: Repeat steps 1 to 5 for Br ,,2,1 , where B is the number of repetition.
Step 7: Obtain the probability distributions )(ˆ b
F of bootstrap estimates
)()21()1( ˆ.,ˆ,ˆ bBbb
and use
)(ˆ b
F to estimate regression coefficients, variances. The bootstrap estimates of regression
coefficient is the mean of the distribution )(ˆ b
F , (see Topuz and Sahinler,2007; Obiora-Ilouno and
Mbegbu, 2012).
br
B
r
br
b
B
ˆ
ˆ
ˆ 1
(2.21)
The bootstrap standard deviation from )(ˆ b
F distribution is
5. A Bootstrap Approach To Error-Reduction Of Nonlinear Regression Parameters Estimation
DOI: 10.9790/5728-11134551 www.iosrjournals.org 49 |Page
Br
B
s
B
b
bdrbdr
b
,,2,1;
ˆˆˆˆ
ˆ
1
1
(2.22)
The Computer Program In R For Bootstrapping Non-Linear Regression:
#x is the vector of independent variable
#theta is the vector of parameters of the model
#This function calculates the matrix of partial derivatives
F=function(x,theta)
{
output=matrix(0,ncol=2,nrow=length(x))
for(i in 1:length(x)) output[i,]=c(exp(theta[2]*x[i]),theta[1]*x[i]*exp(theta[2]*x[i]))
output
}
#This function calculates the regression coefficients using the Gauss-Newton Method
gaussnewton=function(y,x,initial,tol)
{
theta=initial
count=0
eps=y-(theta[1]*exp(theta[2]*x))
SS=sum(eps**2)
diff=1
while(tol<diff)
{
S=SS
ff=F(x,theta)
theta=c(theta+solve(t(ff)%*%ff)%*%t(ff)%*%eps)
eps=y-(theta[1]*exp(theta[2]*x))
SS=sum(eps**2)
diff=abs(SS-S)
count=count+1
if(count==100) break
pp=c(theta,SS)
#at each iteration
}
pp
}
#This part of the code does the bootstrap
boot=function(data,p,b,initial)
{
n=length(data[,1])
z=matrix(0,ncol=p,nrow=n)
output=matrix(0,ncol=p+1,nrow=b)
for (i in 1:b)
{
u=sample(n,n,replace=T)
for (j in 1:n) z[j,]=data[u[j],]
y=z[,1]
x=z[,2:p]
logreg=gaussnewton(y,x,initial,0.00001)
coef=logreg
output[i,]=c(coef)
}
output
}
6. A Bootstrap Approach To Error-Reduction Of Nonlinear Regression Parameters Estimation
DOI: 10.9790/5728-11134551 www.iosrjournals.org 50 |Page
#Then to run the code we use the following
y <- c(data)
x <- c(data)
data=cbind(y,x)
initial=c(initial)
expo=boot(data,p,B,initial)
#Run the following to view the bootstrap results
theta_0=mean(expo[,1])
theta_0
theta_1=mean(expo[,2])
theta_1
SSE=mean(expo[,3])
SSE
Problem[ GUJARATI AND PORTER, 2009]
The data below relates to the management fees that a leading mutual fund pays to its investment advisors to
manage its assets. The fees paid depend on the net asset value of the fund.
Develop a regression model for the management fees to the advisors.
Fee % 0.520 0.508 0.484 0.46 0.4398 0.4238 0.4115 0.402 0.3944 0.388 0.3825 0.3738
Asset 0.5 5.0 10.0 15 20 25 30 35 40 45 55 60
III. Results And Discussion
From the data, the higher the net asset values of the fund, the lower are the advisory fees. (see Gujarati and
Porter ,2009).
The Analytical Result of R Program (Without Bootstrapping)
y <- c(0.520,0.508,0.484,0.46,0.4398,0.4238,0.4115,0.402,0.3944,0.388,0.3825,0.3738)
x <- c(0.5,5,10,15,20,25,30,35,40,45,55,60)
run=gaussnewton(y,x,c(0.5028,-0.002),0.00001)
[1] 0.505805756 -0.005487479 0.001934099
[1] 0.508724700 -0.005949598 0.001699222
[1] 0.508897316 -0.005964811 0.001699048
The estimated model from the result is:
0.00596ˆ 0.50890iY X
where Y and X are the fee and assets respectively.
The Result of R Program Using Bootstrapping Algorithm
> # run the code use the following for the bootstrap algorithm
> y <- c(0.520,0.508,0.484,0.46,0.4398,0.4238,0.4115,0.402,0.3944,0.388,0.3825,0.3738)
6050403020100
0.52
0.50
0.48
0.46
0.44
0.42
0.40
0.38
0.36
Asset
fee%
Relationship of advisory fees to fund assets
7. A Bootstrap Approach To Error-Reduction Of Nonlinear Regression Parameters Estimation
DOI: 10.9790/5728-11134551 www.iosrjournals.org 51 |Page
> x <- c(0.5,5,10,15,20,25,30,35,40,45,55,60)
> data=cbind(y,x)
> initial=c(0.5028,-0.002)
> expo=boot(data,2,1000,initial)
>
> #Run the following to view the bootstrap results
> theta_0=mean(expo[,1])
> theta_0
[1] 0.5075696
> theta_1=mean(expo[,2])
> theta_1
[1] -0.005964939
> SSE=mean(expo[,3])
> SSE
[1] 0.001328289
IV. Discussion
are the revised parameter estimates at the end of the last
iteration. The least squares criterion measure SS for the starting values has been reduced in the first iteration
and also further reduced in second, third iterations respectively. The third iteration led to no change in either the
estimates of the coefficient or the least squares SS criterion measure. Hence, convergence is achieved, and
the iteration end.
The fitted regression function is, XY 00586.0exp50889.0ˆ
The results of the analytical and the bootstrap computations are shown in Table 2
Table 2: Analytical and Bootstrap Results
0 1 SS
Analytical 0.50889 -0.00596 0.001699
Bootstrap 0.50757 -0.00596 0.001328
The fitted regression function for both the analytical bootstrapping computation are
ˆ 0.50889exp 0.00596Y X and XY 00596.0exp50757.0ˆ respectively.
V. Conclusion
From the result of our analysis the bootstrap approach yields approximately the same inference as the
analytical method. Also the bootstrap algorithm yielded a better reduced error sum of squares SS than the
analytical method (see Table 2). With these results, we have a greater confidence in the result obtained by
bootstrap approach.
References
[1]. Casella, G., 2003, Introduction to the Silver Anniversary of the Bootstrap, Statistical Science, Vol.18, No.2, pp. 133-134.
[2]. Efron, B., (1979). Bootstrap Methods: Another look at the Jackknife. The Annuals of Statistics, 7, 1-26.
[3]. Efron, B., 2003, Second Thought on Bootstrapping, Statistical Science, Vol. 18, No. 2, 135-140.
[4]. Efron, B. and Tibshirani, R.J., 1998, An Introduction to the Bootstrap, Library of Congress Cataloging in Publication Data, CRC
Press, Boca Raton, Florida, 60-82.
[5]. Freedman, . A., 1981, Bootstrapping Regression Models, Annals of Statistical Statistics, Vol. 9, 2, 158-167.
[6]. Nduka, E. C., 1999, Principles of Applied Statistics 1, Regression and Correlation Analyses, Crystal Publishers, Imo State, Nigeria.
[7]. Obiora-Ilouno H.O. and Mbegbu J.I (2012), Jackknife Algorithm for the Estimation of Logistic Regression Parameters, Journal of
Natural Sciences Research, IISTE UK, Vol.2, No.4, 74-82.
[8]. Sahinler, S. and Topuz, D., 2007, Bootstrap and Jackknife Resampling Algorithm for Estimation of Regression Parameters, Journal
of Applied Quantitative Method, Vol. 2, No. 2, pp. 188-199.
[9]. Smith, H. and Draper, N.R., 1998, Applied Regression Analysis, Third Edition, Wiley- Interscience, New York.
[10]. Venables, W. N and Smith, 2007, An Introduction to R, A Programming Environment for Data Analysis and Graphics, Version
2.6.1, pp. 1-100.
0 10.50889 and 0.00596