Subscribe
Sign in
Home
AI/ML Course Library
Latest
Top
Discussions
What exactly is inductive bias in CNN?
When is it a good thing? When is it not?
Nov 14
•
Vizuara AI Labs
and
Sreedath Panat
1
What is a teacher student model in Deep Learning?
How does its loss function actually work?
Nov 13
•
Vizuara AI Labs
and
Sreedath Panat
8
Contrastive learning in VLMs
A simple idea that made text and vision alignment possible
Nov 12
•
Vizuara AI Labs
and
Sreedath Panat
1
1
Introduction to VLMs
Vision Language Models
Nov 10
•
Vizuara AI Labs
and
Sreedath Panat
7
Dissecting the Vision Transformer paper: In 3 hours and 40 minutes
Let us cultivate the habit of reading research papers
Nov 5
•
Vizuara AI Labs
and
Sreedath Panat
7
LIVE workshop: Build a NanoVLM from scratch
Happening on Saturday, November 8th
Nov 5
•
Vizuara AI Labs
and
Sreedath Panat
I just built a Vision Transformer from Scratch
Starting with random weights
Nov 3
•
Vizuara AI Labs
and
Sreedath Panat
1
Engineering CI/CD Pipelines for Machine Learning Systems
This article delves into the concept of CI/CD, explaining its fundamentals and highlighting its importance in building reliable, scalable, and automated…
Nov 2
•
Prathamesh Dinesh Joshi
and
Vizuara AI Labs
3
1
2/3rd of trainable parameters in GPT-3 belong to MLP. Not attention heads.
2-minute read
Nov 1
•
Vizuara AI Labs
and
Sreedath Panat
2
October 2025
If you have 96 attention heads, will you run 96 loops?
Deeply understanding multi-head attention with weight splits
Oct 30
•
Vizuara AI Labs
and
Sreedath Panat
3
1
Why do we really need more than one attention head?
Understanding Multi-Head Attention: The Heart of Transformers
Oct 28
•
Vizuara AI Labs
and
Sreedath Panat
4
1
Why do we need "masking" in attention?
Understanding causal or masked self attention
Oct 25
•
Vizuara AI Labs
and
Sreedath Panat
4
This site requires JavaScript to run correctly. Please
turn on JavaScript
or unblock scripts