In the previous post, we learnt about a PValue, a prerequisite for learning Multiple Linear Regression.
Business Problem: In this series, we will take a look at a dataset of 50 startup companies.
A venture capitalist has hired you as a data scientist and wants you to help him select which type of company he should invest so that he can make the most profit. You need to review spending on R&D, Admin cost, marketing cost and location to make the decision
How to get the dataset?
What is Multiple Linear Regression?
In the second post on linear regression, our equation was simple and straightforward
Y = mX + C
where Y
was the dependent variable, X
was the dependent variable and c
was the Yintercept when X = 0
. The reason the equation was simple because the Y
was dependent on one variable only. If there were multiple variables affecting the value of Y
, then what should it be called?  :). You know the answer to that question.
Multiple Linear Regression!!!
The equation would also be something as simple as that
Y = b_{0}X_{0} + b_{1}X_{1}+b_{2}X_{2}+ …. + C
Steps to solve the problem?
 Step 1: Data preprocessing and analysis
Take a closer look at dataset. You will notice that all the Independent Variables
, except State
, is numerical. The variable State
is either California or New York. From our first post, you would know that this type of data is called categorical data. We should always convert categorical data into numerical data to avoid bias and find if there is collinearity between Profit
and State
. Collinearity is just a fancy way of asking  “Is there some relation between Profit
and State
?.
When you convert a categorical data to numeric data, the new column is called Dummy Variable. So as we learned from our first post, we should convert it to a sparse matrix.
Question
*Do you need two columns to represent New York and California states?
The answer is No.
It is easy to derive from the above screenshot that if New York is 1 then California by default would be 0 and viceversa. So this actually works like a switch which can have only 2 states 0 or 1.
Important tip: You should never use all of your dummy variables in your Regression column. They should always be 1 less than the number of values.
If we drop one dummy variable then are we not making this a biased equation
Without dropping column my equation would look like this
Y = b_{0}X_{0} + b_{1}X_{1}+ b_{2}X_{2} + b_{3}X_{3} + b_{4}X_{4} + b_{5}X_{5} + C
After Dropping one dummy variable
Y = b_{0}X_{0} + b_{1}X_{1}+ b_{2}X_{2} + b_{3}X_{3} + b_{4}X_{4} + C
If we dropped the variable then it may appear that when California
is the state then the value of b_{4}X_{4} will be 0, hence we lose one variable. In reality, the regression algorithm marks that the first dummy variable which is represented by 0
is set as default. So, the regression equation will never be biased. The regression equation is going to use constant C
to adjust the value of California
.
But what is wrong with using all the dummy variable?
When you use both the values, your equation would be something
Y = b_{0}X_{0} + b_{1}X_{1}+ b_{2}X_{2} + b_{3}X_{3} + b_{4}X_{4} + b_{5}X_{5} + C
If we do this then we will introduce multicollinearity, where the algorithm will not be able to able to distinguish the effect on Price
. This is because D_{2} is always equal to 1  D_{1}. The algorithm will then try to predict the effect of D2
over D1
and would think that there is a relation between Independent
variable as well.
 Step 2: Understanding all possible methods to build a model using Multiple Linear Regression models
There are 5 methods to build a model
 All in one
 Backward Elimination
 Forward Selection
 BiDirectional Elimination
 Score comparison
All In: All in means that you use all the variables when you know for sure that all the independent variables have a definite effect on the dependent variable. An example is that if doctors told you for sure that to live past 80 yrs of age, you should eat good food and exercise daily. In other words, you have domain expert telling you that all the variables directly affect the dependent variable.
Backward elimination: In backward elimination, you take all the variables and create the algorithm. Select a significance level, then consider the predictor with Highest Pvalue and if
PValue > Significance level
then eliminate the variable from the equation, else keep it.
Forward Selection: In forward selection, you start with linear regression using every single variable. You will end up with
n
simple linear equation. Next, you chose the one with the lowest PValue. This is your starting equation. Y = m_{0}X_{0}. Next, you pick one variable again and create an equation with two variables and out ofn1
possibilities again chose the one with the lowest Pvalue. The process continues until we don’t have any variable that is lower than our selected Significance level.
BiDirectional elimination: In bidirectional elimination, you chose 2 significance level. One to enter the equation and one to stay. You start with the forward selection using condition PValue < SL _{Enter} and then follow backward elimination using condition Pvalue < SL_{stay}. You stop and declare the final equation when no new variable can enter or exit the equation.
Score Comparison: In this model, you create all possible combination of the equation, compare the performance using say MSE(Mean Square Error) and use the one with the lowest MSE. That is an insane amount of possible equation. For Example  A model with 10 variables will have 1023 possible combination.
Note for the purpose of brevity and sanity, we will be using Backward Elimination
model to solve this problem. Also, because this model is fastest and we will still be able to see how the step by method works.
 Step 3: Data preprocessing in python
We will now start the calculation using the process that we learnt in the first post of this series. If you remember from Linear regression post




Next part of data processing is to find if there is any missing data. If there is missing data then we need to use Imputer to fix the missing data. To do that we will first check the data description, then try to get a count of missing values in the dataset


Series([], dtype: float64)
Currently, there are no missing values in the dataset hence the result of num_emptycolumns
.
Moving on to the next item in the data clean up is taking care of categorical values in the State
column.




Hey, what about all the talk about keeping n1 categorical items?
So you are right to notice that whatever we did will lead us directly to the dummy variable trap. I fell into one when I was trying to learn it. So how do we fix it? If you have been thinking about just crossing out one of the columns as we did in the pic few scrolls above ……. you are right!!! That’s exactly how we are going to fix it.


Now divide the data in training and test set


Step 4: Training the model
Now that we have cleaned up our data, we can train our data. We are not going to do any feature scaling here, because the library and class that we are going to use, will do that automatically for us. We will use the same LineaRegression
class that we used in the linear regression post.


The above code trained our data. Now we need to check how does our model score or how good is it at predicting the data?
To Predict we just need to call the predict function.


You can see that the predicted values ypred
is pretty close to y_test
. Some of the rows do have a significant difference and the numbers may look far apart but the others are pretty close. So how close are we? How confident are we that the trendline would fit closely. To answer this we need to look at our training and test score.


Train Score: 0.9501847627493607
Test Score: 0.9347068473282446
The above score tells us that our model has 95%
and 93.5%
accurate for training and test data. This not bad.
One last thing that I learnt was that if someone asked me how do I calculate the future values that may come up? To do that we need to get the coefficient and intercept of the regression equation. The regression equation as discussed earlier would look something like this 
Y = b_{0}X_{0} + b_{1}X_{1}+ b_{2}X_{2} + b_{3}X_{3} + b_{4}X_{4} + b_{5}X_{5} + C


[9.59284160e+02 6.99369053e+02 7.73467193e01 3.28845975e02
3.66100259e02]
42554.16761772438
If we take the first row from X_test
 1 0 66051.5 182646 118148
1*9.59284160e+02 + 06.99369053e+02 + 66051.5 * 7.73467193e01 + 1826463.28845975e02 + 118148 * 3.66100259e02 + 42554.16761772438
The output would be 103015.19329118208
, which is same as first row of y_pred
.


103015.19329118208
So that is the most basic way to do to solve a problem using Multiple Linear regression. Hope you enjoyed this series. Stay tuned for the next series where we will actually see the continuation of Multiple Linear regression
and learn the Backward Elimination process.