Machine Learning Tutorial Python - 21: Ensemble Learning - Bagging

Machine Learning Tutorial Python - 21: Ensemble Learning - Bagging

codebasics

2 года назад

92,425 Просмотров

Ссылки и html тэги не поддерживаются


Комментарии:

@nastaran1010
@nastaran1010 - 27.01.2024 01:41

I hope you see my questions you never response to my questions. why you didi not fit "BaggingClassifier' with '(x_train,y_train)', in exercise?

Ответить
@bhanusri3732
@bhanusri3732 - 13.01.2024 08:44

Why during cross validation using original unscaled X instead of X scaled? Does it not affect accuracy?

Ответить
@Koome777
@Koome777 - 14.12.2023 17:14

My results of the exercise: svm standalone 0.8, after bagging 0.8, Decision Tree standalone 0.65, after bagging 0.79. Bagging helps improve accuracy and reduce overfitting, especially in models that have high variance. Mostly used for unstable models like Decision Trees

Ответить
@jasonwang-wg8wu
@jasonwang-wg8wu - 06.12.2023 23:27

this was nice and straightforward, and the quip about "copy and paste" was hilarious

Ответить
@abebebelew2056
@abebebelew2056 - 05.12.2023 17:38

It us very helpful video to do my research project.!!!!!

Ответить
@aniketmlk6
@aniketmlk6 - 18.10.2023 14:12

Thanks a lot for your awesome series!

Ответить
@ogochukwustanleyikegbo2420
@ogochukwustanleyikegbo2420 - 28.08.2023 17:04

I also learnt that bagging doesn't do so much in increasing the performance of the model apart from lowering the variance.

Ответить
@ogochukwustanleyikegbo2420
@ogochukwustanleyikegbo2420 - 28.08.2023 15:48

My results after completing the exercise
svm:
Standalone 0.82
Bagged model 0.87

Decision Trees:
Standalone 0.79
Bagged model 0.84

Ответить
@SohamPaul-xy9jw
@SohamPaul-xy9jw - 10.07.2023 07:43

My bagging model score came out to be : 0.8027, SVC : 0.8804, Decision Tree : 0.804

Ответить
@ankitjhajhria7443
@ankitjhajhria7443 - 21.06.2023 12:26

why are we fitting our model on X,y
then what is the use of x_train and y_train and no use of scaling also if we are trainning our model on original X and y ?

Ответить
@diptopodder1011
@diptopodder1011 - 10.06.2023 21:42

How to train multiple file and then provide them label for individual file and classify a file?

Ответить
@coder_62
@coder_62 - 05.06.2023 14:28

O my gad, my computer get fever for 1 month wkwkwwk. Btw thank you sir for your clear explanation.!!!

Ответить
@1itech-Learn
@1itech-Learn - 29.05.2023 09:40

Thank you so much sir for this ML playlist

Ответить
@siddheshmhatre2811
@siddheshmhatre2811 - 21.05.2023 09:59

One of the most underrated playlists for ML . I wish lots of student will join ❤

Ответить
@nikhilraj93
@nikhilraj93 - 19.05.2023 18:18

I tried clicking on soultion
Now I have fever

Ответить
@kotakarthik-fq5cp
@kotakarthik-fq5cp - 18.05.2023 14:20

bagging svc gave a far better result than bagging decision tree

Ответить
@devesh_upreti
@devesh_upreti - 13.03.2023 12:34

SVC score without bagging 0.87
DecisionTreeClassifier score without bagging 0.76

SVC score with bagging 0.867
DecisionTreeClassifier score with bagging 1.0

Drastic improvement in Decision Tree Classifier

Ответить
@nachiketgalande8125
@nachiketgalande8125 - 04.03.2023 19:00

Thankyou SIR! for this amazing playlist on machine learning

Ответить
@sahilgundu5338
@sahilgundu5338 - 22.02.2023 22:39

Im not able to see top row, all column headings in names in CSV file downloaded from kaggle - pima-indians-diabetes.csv

Am I doing any mistake while downloading?

Ответить
@kmishy
@kmishy - 02.12.2022 00:49

thanks sir

Ответить
@kushanmadusankha5227
@kushanmadusankha5227 - 23.11.2022 06:25

Awesome 🔥
Appreciate your effort bro

Ответить
@Freeak6
@Freeak6 - 11.11.2022 10:26

Shouldn't you use X_train in the cross-validation calls?

Ответить
@bit-colombo5595
@bit-colombo5595 - 12.10.2022 04:20

Hi sir can make a video on how to combine classifiers like decision tree, random forest ,naive bayes and svm and get a colleciive result, like a weighted output

Ответить
@atur42
@atur42 - 01.10.2022 10:26

good work really

Ответить
@nitinpednekar8872
@nitinpednekar8872 - 07.09.2022 06:52

Sir, I don't see any time series forecasting model videos, request to upload videos for the same

Ответить
@dataguy7013
@dataguy7013 - 01.07.2022 18:04

Best explanation, EVER!!

Ответить
@sanskaragrawal5074
@sanskaragrawal5074 - 04.06.2022 14:57

Excellent expanation sir.The whole series has been exceptional.
Had one query -'How can reduction the size of data set decreasee variance .Decreasing no of features might decrease it,but how decreasing no of training examples can decrease it

Ответить
@LamNguyen-jp5vh
@LamNguyen-jp5vh - 31.05.2022 05:07

Hi, can you explain further the difference between bagging and bagged trees. I don't really understand the explanation in the video. Thank you so much for your help! Your videos are amazing.

Ответить
@pranav2901
@pranav2901 - 12.05.2022 07:00

thank you very much for this video

Ответить
@pranav2901
@pranav2901 - 12.05.2022 07:00

when the boosting will be uploaded ?

Ответить
@omsaichand752
@omsaichand752 - 30.04.2022 20:36

Your tutorials are not properly structured and are not learner centric!

Ответить
@elahe4737
@elahe4737 - 28.04.2022 19:04

That was clearly describe what is the bagging method, I wish you had a video about Boosting as well

Ответить
@bea59kaiwalyakhairnar37
@bea59kaiwalyakhairnar37 - 02.04.2022 17:17

You have to do outlier detection because the max is much higher than that of 75% value

Ответить
@paulkornreich9806
@paulkornreich9806 - 02.03.2022 06:07

This exercise was a challenge. Thank you. By just taking pure z of the set, some outliers were missed. Basically, all the outliers were the 0s for blood pressure and cholesterol. With those eliminated, I got significantly higher scores than the solution. All bagged models gave a similar 86% accuracy. The biggest jump from non-bagged model to bagged model was the Decision Tree which went from 79% accuracy without bagging to 86% with bagging. Also, I did the exercise several months after this video (was posted - not sure when it was made), so the libraries (especially SVC) may have improved (in their defaults).

Ответить
@vikranttripathi2258
@vikranttripathi2258 - 26.02.2022 17:21

Thank you for this wonderful explanation. I have a query here. We scaled X but everywhere we use X in cross_val_score. Could you please explain why we scaled X?

Ответить
@ritvijmishra4727
@ritvijmishra4727 - 05.02.2022 10:43

Thank you so much sir for this ML playlist. Your explanations are simple, exact, and extremely easy to follow. The method that you use of first familiarizing us with theory, then with a practical example and then an exercise is really effective. Looking forward to more of such videos in your ML series. Thanks once again, sir.

Ответить
@snehasneha9290
@snehasneha9290 - 28.01.2022 15:25

@ 21: 50 in cross Val score you x and y why not x_train and y_train can anyone explain this

Ответить
@snehasneha9290
@snehasneha9290 - 28.01.2022 14:36

by using the df. describe() how can we decide is it necessary to do outlier removal or not please can anyone help me for my question

Ответить
@anmoldubey3375
@anmoldubey3375 - 22.01.2022 14:23

Thankyou for making such a clear video in bagging and RF.

I have one doubt in RF, whe RF does rows and feature sampling so in feature sampling, some of the DT might not get relevant features or
not even the features we might wanna use, so doest this affect accuracy and not let us get the result that we want.

Ps i know this is a lot of writing!!!!!

Ответить
@tejaswiganti8098
@tejaswiganti8098 - 18.01.2022 20:49

I dont know why i am getting output and mean as 1 while using DesicionTreeClassifier and RandomForestClassifier,
I have tried with different values but value is same and not getting the exact reason.
Can you guys tell me where i have made mistake:|

Ответить
@rajagopalk9760
@rajagopalk9760 - 17.01.2022 11:01

Good presentation and preparation; easy to understand. I wish to get a clarification that, why the term "resampling with replacement" is used instead of "sampling with replacement". Is there incidental or there is any specific reason? Thank you.

Ответить
@tarunmohapatra5734
@tarunmohapatra5734 - 09.01.2022 21:12

I am waiting for Boosting and Xgboost methods sir

Ответить
@PP-tc1zp
@PP-tc1zp - 24.12.2021 00:55

Thank you sir for a very good explanation. Those examples are very good to training write a code and cause strong motivation.

Ответить
@PP-tc1zp
@PP-tc1zp - 21.12.2021 12:17

Thank you for your courses
I have differen code to detsct otluiers. This code also works very good. It is more simple.
Best Regards
'''Q1 = df.quantile(0.25)
Q3 = df.quantile(0.75)
IQR = Q3 - Q1

outlier_condition = ((df < (Q1 - 1.5 * IQR)) | (df > (Q3 +1.5 * IQR)))
df3 = df[~outlier_condition.any(axis=1)]

df3.shape'''

Ответить
@troubution
@troubution - 18.12.2021 10:52

Hi, Dhaval I hope you are doing well, I have a query in this, at step 35 you have provided input as X and y to the model. What if you have provided input as X_scaled instead of X, i think accuracy might be different.

Ответить
@bravelionable
@bravelionable - 24.11.2021 01:45

You are the best! Thank you

Ответить
@60pluscrazy
@60pluscrazy - 07.11.2021 18:28

Random forest explanation is superb 👌

Ответить