Logistic Regression, But Make It Tea: ML Basics Served Hot

Published: (December 20, 2025 at 10:45 AM EST)
3 min read
Source: Dev.to

Source: Dev.to

What Is Logistic Regression?

Logistic Regression is a simple machine learning algorithm used to predict yes/no outcomes.

Imagine running a small tea stall and wanting to predict whether a passer‑by will buy tea.
Features you might consider:

  • Time of day
  • Weather
  • Whether the person looks tired
  • Whether they’re rushing

Logistic regression converts these features into a probability between 0 and 1, e.g., “There’s a 70 % chance they will buy tea.”

1. Cost Function — Measuring How Wrong You Are

A cost function quantifies the difference between the model’s predictions and reality.

  • Lower cost → better model.

Tea Analogy

If you guess whether 100 people will buy tea:

  • Accurate guesses → low cost
  • Frequent wrong guesses → high cost

The model learns by trying to minimize this cost.

2. Logistic Loss (Log‑Loss / Binary Cross‑Entropy)

Because logistic regression predicts probabilities, we use logistic loss instead of simple error counting. It penalizes confident but incorrect predictions more heavily than uncertain ones.

Tea Analogy

  • Predict 90 % chance of purchase but the person doesn’t buy → large penalty
  • Predict 55 % chance and they don’t buy → smaller penalty

Logistic loss encourages realistic probability estimates.

3. Gradient Descent — How the Model Learns

Gradient Descent is an optimization method that iteratively adjusts model parameters to minimize the cost function.

Analogy

Picture standing on a foggy hill, taking small steps downhill by feeling the slope. Each step reduces the elevation (cost) until you reach the lowest point.

Tea Example

Finding the optimal tea price:

  • ₹20 → few buyers
  • ₹10 → many buyers
  • ₹8 → even more buyers
  • ₹6 → too low, profit drops

By making tiny adjustments, you discover the sweet spot. Gradient descent performs the same incremental updates for model weights.

4. The Problem of Overfitting — When the Model Becomes “Too Smart”

Overfitting occurs when a model memorizes training data instead of learning general patterns.

Tea Analogy

Among 100 customers, only one person wearing a red shirt bought tea.
An overfitted model would learn “Red shirt = tea buyer always,” which is incorrect—it has learned noise.

Symptoms

  • Excellent performance on training data
  • Poor performance on new, unseen data

5. How to Prevent Overfitting

Common strategies:

  • Use more data
  • Simplify the model
  • Apply regularization (especially important for logistic regression)

6. Regularization — Keeping the Model Grounded

Regularization adds a penalty term to the cost function, discouraging the model from assigning excessive importance to irrelevant features.

Tea Analogy

If you start tracking trivial details (shoe brand, phone color, bag weight, hair length), the model may overfit. Regularization tells it to ignore these noisy features and focus on meaningful ones like weather, time, and tiredness.

7. Regularized Logistic Regression — Smarter Cost Function

The regularized cost function is:

Total Cost = Logistic Loss + Regularization Penalty

Types of Regularization

  • L1 (Lasso): Can zero out useless feature weights.
  • L2 (Ridge): Shrinks feature weights smoothly.

Tea Example

Regularization penalizes the model if it tries to learn rules such as:

  • “Red shirts always buy tea”
  • “Black shoes rarely buy tea”

This keeps the model generalizable and stable.

Conclusion

Using the familiar setting of a tea stall, we’ve covered the core ideas behind logistic regression: cost function, logistic loss, gradient descent, overfitting, and regularization. These concepts form the foundation for most machine‑learning algorithms.

Back to Blog

Related posts

Read more »