Logistic Regression Explanation

157 Views Asked by At

I have two questions regarding logistic regression.

1) I understand that the results of a logistic regression model yield a table stating coefficients together with a p-statistic for each variable . This $p$-statistic states whether the variable at hand is important at predicting the outcome variable. For example, given the null hypothesis that the independent variable at hand is not important at predicting the outcome variable, then if $p< 0.05$ we conclude that the variable is important indeed. My question: why do we have $p$-statistics for logistic regression and not for linear regression? Wouldn't a coefficient close to $0$ be enough to conclude that the independent variable at hand is not an important predictor?

2) I'm having a bit of trouble understanding the relationship between the logistic regression equation $p = 1/(1+e^{-y})$ and its logit counterpart $\log(p/(1-p)) = MX + B$. I understand that these two equations are equivalent, correct? Is it safe to explain their relationship (and relevance to our model) by saying that we are most interested in the outcome of $p = 1/(1+e^{-y})$ given that this outcome gives us the probability that each observation produces the desired outcome, while the logit equation is only important to solve for coefficients?

1

There are 1 best solutions below

4
On

1 Certainly you get p-values for regression coefficients ( $\beta$ ) in simple linear regression. They also test the hypothesis $\beta = 0$. This extends to multiple linear regression where there are multiple regression coefficients. For logistic regression, how accurate and reliable the p-values for the regression coefficients are is a different question.

2 In this case

$$ \log(p/(1-p))=MX+B \\ \Rightarrow p/(1-p)=e^{MX+B} \\ \Rightarrow p=e^{MX+B}/(1+e^{MX+B})\\ \Rightarrow p = 1/(1+e^{-(MX+B)}) $$ In other words, $y=MX+B$ in your notation.