Avi Chawla
Search documents
X @Avi Chawla
Avi Chawla· 2025-12-02 13:20
If you found it insightful, reshare it with your network.Find me → @_avichawlaEvery day, I share tutorials and insights on DS, ML, LLMs, and RAGs. https://t.co/gly3mHfP4WAvi Chawla (@_avichawla):Few people know this about L2 regularization:(Hint: it is NOT just a regularization technique)Most models intend to use L2 Regularization for just one thing:↳ Reduce overfitting.However, L2 regularization is a great remedy for multicollinearity.Multicollinearity arises https://t.co/T4Q35XSrYM ...
X @Avi Chawla
Avi Chawla· 2025-12-02 06:50
In fact, this is where “ridge regression” also gets its name from:Using an L2 penalty eliminates the RIDGE in the likelihood function of a linear model.Check this👇 https://t.co/h07l36upoQ ...
X @Avi Chawla
Avi Chawla· 2025-12-02 06:50
Few people know this about L2 regularization:(Hint: it is NOT just a regularization technique)Most models intend to use L2 Regularization for just one thing:↳ Reduce overfitting.However, L2 regularization is a great remedy for multicollinearity.Multicollinearity arises when:→ Two (or more) features are highly correlated, OR,→ Two (or more) features can predict another feature.To understand how L2 regularization addresses multicollinearity, consider a dataset with two features and a dependent variable (y):→ ...
X @Avi Chawla
Avi Chawla· 2025-12-01 20:25
技术洞察 - 展示了 KMeans 聚类算法的 @3blue1brown 风格动画 [1] - 行业关注机器学习算法的可视化呈现 [1]
X @Avi Chawla
Avi Chawla· 2025-12-01 06:37
A @3blue1brown style animation of KMeans clustering: https://t.co/Irb2yzUCR1 ...
X @Avi Chawla
Avi Chawla· 2025-11-30 12:18
If you found it insightful, reshare it with your network.Find me → @_avichawlaEvery day, I share tutorials and insights on DS, ML, LLMs, and RAGs. https://t.co/ZI7b5wIQ8kAvi Chawla (@_avichawla):Andrew Ng's team once made a big mistake in a research paper.And it happened due to randomly splitting the data.Here's exactly what happened (with solution): ...
X @Avi Chawla
Avi Chawla· 2025-11-30 06:47
A few days later, Andrew Ng's team updated the paper after using the same group shuffle split strategy to ensure the same patients did not end up in both the training and validation sets.👉 Over to you: Have you faced this issue before? https://t.co/GES6FESMZm ...
X @Avi Chawla
Avi Chawla· 2025-11-30 06:47
First, we import the GroupShuffleSplit from sklearn and instantiate the object.Next, the split() method of this object lets us perform group splitting. It returns a generator, and we can unpack it to get the following output:- The data points in groups “A” and “C” are together in the training set.- The data points in group “B” are together in the validation/test set.Check this 👇 ...
X @Avi Chawla
Avi Chawla· 2025-11-30 06:47
Andrew Ng's team once made a big mistake in a research paper.And it happened due to randomly splitting the data.Here's exactly what happened (with solution): ...
X @Avi Chawla
Avi Chawla· 2025-11-29 19:27
RT Avi Chawla (@_avichawla)Speed up your native Python code by over 50x!And it takes just 4 simple steps.Python’s default interpreter (CPython) is slow primarily because of its dynamicity.For instance, after defining a variable of a specific type, it can be changed to some other type.But these dynamic manipulations come at the cost of run-time and memory overheads.The Cython module converts your Python code into C.Steps to use the Cython module (refer to the image as you read):1) Load the Cython module: %lo ...