About 510,000 results
Open links in new tab
  1. Regression with multiple dependent variables? - Cross Validated

    Nov 14, 2010 · Is it possible to have a (multiple) regression equation with two or more dependent variables? Sure, you could run two separate regression equations, one for each DV, but that …

  2. How should outliers be dealt with in linear regression analysis ...

    What statistical tests or rules of thumb can be used as a basis for excluding outliers in linear regression analysis? Are there any special considerations for multilinear regression?

  3. regression - When is R squared negative? - Cross Validated

    Also, for OLS regression, R^2 is the squared correlation between the predicted and the observed values. Hence, it must be non-negative. For simple OLS regression with one predictor, this is …

  4. regression - Difference between forecast and prediction ... - Cross ...

    I was wondering what difference and relation are between forecast and prediction? Especially in time series and regression? For example, am I correct that: In time series, forecasting seems …

  5. correlation - What is the difference between linear regression on y ...

    The Pearson correlation coefficient of x and y is the same, whether you compute pearson(x, y) or pearson(y, x). This suggests that doing a linear regression of y given x or x given y should be …

  6. regression - Trying to understand the fitted vs residual plot?

    Dec 23, 2016 · A good residual vs fitted plot has three characteristics: The residuals "bounce randomly" around the 0 line. This suggests that the assumption that the relationship is linear is …

  7. regression - Interpreting the residuals vs. fitted values plot for ...

    Consider the following figure from Faraway's Linear Models with R (2005, p. 59). The first plot seems to indicate that the residuals and the fitted values are uncorrelated, as they should be in a

  8. When conducting multiple regression, when should you center …

    Jun 5, 2012 · In some literature, I have read that a regression with multiple explanatory variables, if in different units, needed to be standardized. (Standardizing consists in subtracting the mean …

  9. regression - Converting standardized betas back to original …

    I have a problem where I need to standardize the variables run the (ridge regression) to calculate the ridge estimates of the betas. I then need to convert these back to the original variables scale.

  10. regression - Maximum likelihood method vs. least squares method …

    What is the main difference between maximum likelihood estimation (MLE) vs. least squares estimaton (LSE) ? Why can't we use MLE for predicting $y$ values in linear ...