Data Analytics Program

 

Data Analytics Program:

Build a simple linear regression model for Salary Prediction from years of experience.(Download Salary dataset from kaggle). Find the accuracy of the model.

In [ ]:
In [61]:
In [62]:
Out[62]:
experiencesalary
01100
12200
23300
34405
45499
57700
68800
79900
In [63]:
Out[63]:
experiencesalary
0FalseFalse
1FalseFalse
2FalseFalse
3FalseFalse
4FalseFalse
5FalseFalse
6FalseFalse
7FalseFalse
In [64]:
Out[64]:
experiencesalary
count8.0000008.00000
mean4.875000488.00000
std2.900123289.79648
min1.000000100.00000
25%2.750000275.00000
50%4.500000452.00000
75%7.250000725.00000
max9.000000900.00000
In [65]:
Out[65]:
experiencesalary
01100
12200
23300
34405
45499
In [66]:
Out[66]:
experiencesalary
34405
45499
57700
68800
79900
In [67]:
Out[67]:
experiencesalary
34405
In [68]:
Out[68]:
experience    int64
salary        int64
dtype: object
In [69]:
Out[69]:
 experiencesalary
experience1.0000000.999980
salary0.9999801.000000
In [70]:
Out[70]:
0    100
1    200
2    300
3    405
4    499
5    700
6    800
7    900
Name: salary, dtype: int64
In [71]:
Out[71]:
experience
01
12
23
34
45
57
68
79
In [72]:
In [73]:
In [74]:
In [75]:
   experience
4           5
7           9
0           1
3           4
5           7
1           2
In [76]:
4    499
7    900
0    100
3    405
5    700
1    200
Name: salary, dtype: int64
In [77]:
(2, 1)
In [78]:
In [79]:
Out[79]:
LinearRegression()
In [80]:
Out[80]:
array([300.80147059, 800.39705882])
In [81]:
Out[81]:
Actualpredicted
2300300.801471
6800800.397059
In [82]:
[1000.23529412]
In [83]:
In [ ]:
In [ ]:
In [ ]:
In [ ]:
In [ ]:







Q2. Use the house price prediction dataset to build a multiple linear regression model for predicting purchases. Identify independent and target variable. Split the variables into training and testing sets and print them. Find the accuracy of the model.

In [1]:
In [88]:
Out[88]:
areabedroomsageprice
010003.020550000
110044.02156000
212005.02260000
31300NaN2370000
420006.03080000
530008.03490000
640009.04099000
In [89]:
Out[89]:
<bound method NDFrame.describe of    area  bedrooms  age   price
0  1000       3.0   20  550000
1  1004       4.0   21   56000
2  1200       5.0   22   60000
3  1300       NaN   23   70000
4  2000       6.0   30   80000
5  3000       8.0   34   90000
6  4000       9.0   40   99000>
In [90]:
Out[90]:
area          int64
bedrooms    float64
age           int64
price         int64
dtype: object
In [91]:
Out[91]:
areabedroomsageprice
010003.020550000
110044.02156000
212005.02260000
31300NaN2370000
420006.03080000
In [92]:
Out[92]:
areabedroomsageprice
0FalseFalseFalseFalse
1FalseFalseFalseFalse
2FalseFalseFalseFalse
3FalseTrueFalseFalse
4FalseFalseFalseFalse
5FalseFalseFalseFalse
6FalseFalseFalseFalse
In [93]:
Out[93]:
area        0
bedrooms    1
age         0
price       0
dtype: int64
In [94]:
Out[94]:
areabedroomsageprice
010003.00000020550000
110044.0000002156000
212005.0000002260000
313005.8333332370000
420006.0000003080000
530008.0000003490000
640009.0000004099000
In [95]:
In [96]:
Out[96]:
areabedroomsage
010003.00000020
110044.00000021
212005.00000022
313005.83333323
420006.00000030
530008.00000034
640009.00000040
In [97]:
Out[97]:
0    550000
1     56000
2     60000
3     70000
4     80000
5     90000
6     99000
Name: price, dtype: int64
In [98]:
In [99]:
Out[99]:
areabedroomsage
110044.00000021
420006.00000030
212005.00000022
010003.00000020
313005.83333323
In [100]:
Out[100]:
1     56000
4     80000
2     60000
0    550000
3     70000
Name: price, dtype: int64
In [101]:
(2, 3)
In [102]:
In [103]:
Out[103]:
LinearRegression()
In [104]:
Out[104]:
array([2578443.38316721, 1507365.03322259])
In [105]:
Out[105]:
Actualpredicted
6990002.578443e+06
5900001.507365e+06
In [ ]:
In [ ]:


Create ‘User’ Data set having 5 columns namely: User ID, Gender, Age, EstimatedSalary and Purchased. Build a logistic regression model that can predict whether on the given parameter a person will buy a car or not. Display confusion matrix.

In [20]:
In [21]:
Out[21]:
GenderAgeSalaryPurchased
0male4472000No
1male2748000Yes
2female3054000No
3male3861000No
4female4030000Yes
5female3558000Yes
6male3752000No
7male4879000Yes
8male5083000No
9female3767000Yes
In [22]:
Out[22]:
GenderAgeSalaryPurchased
014472000No
112748000Yes
203054000No
313861000No
404030000Yes
503558000Yes
613752000No
714879000Yes
815083000No
903767000Yes
In [23]:
Out[23]:
GenderAgeSalaryPurchased
0144720000
1127480001
2030540000
3138610000
4040300001
5035580001
6137520000
7148790001
8150830000
9037670001
In [24]:
Out[24]:
GenderAgeSalaryPurchased
0TrueTrueTrueTrue
1TrueTrueTrueTrue
2TrueTrueTrueTrue
3TrueTrueTrueTrue
4TrueTrueTrueTrue
5TrueTrueTrueTrue
6TrueTrueTrueTrue
7TrueTrueTrueTrue
8TrueTrueTrueTrue
9TrueTrueTrueTrue
In [25]:
In [26]:
Out[26]:
array([[    1,    44, 72000],
       [    1,    27, 48000],
       [    0,    30, 54000],
       [    1,    38, 61000],
       [    0,    40, 30000],
       [    0,    35, 58000],
       [    1,    37, 52000],
       [    1,    48, 79000],
       [    1,    50, 83000],
       [    0,    37, 67000]], dtype=int64)
In [27]:
Out[27]:
array([0, 1, 0, 0, 1, 1, 0, 1, 0, 1])
In [28]:
In [29]:
Out[29]:
array([[ 0.5       , -0.70226353, -1.26927717],
       [ 0.5       ,  1.56976553,  1.49447151],
       [ 0.5       , -0.49571543, -0.34802761],
       [ 0.5       ,  0.74357315,  0.77794407],
       [-2.        , -1.11535972, -0.6551108 ]])
In [30]:
Out[30]:
LogisticRegression(random_state=10)
In [31]:
Out[31]:
array([0, 0, 0, 1, 0])
In [39]:
Out[39]:
<AxesSubplot:xlabel='Predicted', ylabel='Actual'>
In [51]:
In [ ]:
In [ ]:
In [ ]:




Frequent itemset and Association rule mining

#dataset creation in python transactions = [['bread','milk'],['bread', 'diaper''milk','eggs'],['milk', 'diaper','beer','coke'], ['bread','milk','diaper','beer'], ['bread','milk','diaper','coke']]

In [4]:
Out[4]:
[['bread', 'milk'],
 ['bread', 'diapermilk', 'eggs'],
 ['milk', 'diaper', 'beer', 'coke'],
 ['bread', 'milk', 'diaper', 'beer'],
 ['bread', 'milk', 'diaper', 'coke']]
In [ ]:
In [ ]:
In [ ]:
In [ ]:
In [ ]:
In [ ]:
In [ ]:
In [ ]:
In [ ]:
In [ ]:
In [ ]:
In [ ]:





z












z


Comments

Popular posts from this blog

Full Stack Developement Practical Slips Programs

Practical slips programs : Machine Learning

Some slips for full stacks