The following Vibe Coding single page web applications are built using AI and run on most web
browsers and mobile
devices.
"Vibe coding" is a term coined by Andrej Karpathy in early 2025 to describe a conversational and intuitive
approach to programming using AI tools, particularly large language models (LLMs).
Introduction to stateless vector AI
The simplest AI imaginable with just one neuron
Learn the basics of AI and experiment with a neural network consisting of a single neuron,
optimize parameters and experience how AI algorithms develop in real time.
Perfect for anyone who wants to understand the fundamentals of AI,
from aspiring data scientists, tech enthusiasts to average joe.
Vibe coding
Learning
Optimization
Curve fitting a workhorse of modern AI
Most people associate human-like machines that can see, speak and reason with brain-like neural networks
that can think and reflect on a problem.
But Artificial Neural Networks (ANNs) are mostly stateless and primarily master one thing: they excel at
curve fitting, mapping input data to outputs through complex functions, enabling them to solve diverse
problems, including training large language models (LLMs).
Let’s now demystify this core concept of modern AI
Vibe coding
Learning
Approximation
Optimization
Gradient Decent vs. Simulated Annealing
This software demonstrates the solution to an optimization problem, using the example of the shortest
route to a hospital. Optimization problems play a central role in machine learning.
Vibe coding
Learning
Optimization
To be continued ...
Large Language Models
LLMs Part 1: Understand the meaning of words through AI
This AI embedding algorithm
visualizes the relationships and meanings of words in a 2D vector space,
where every word tells a story through its position and connections.
This deep comprehension of word meanings forms the basic building block for large language models
(LLMs),
enabling them to generate coherent, contextually relevant text.
Vibe coding
NLP
Learning
Search
LLMs Part 2: Understand token position through AI
Unlike traditional sequence models like Recurrent Neural Networks (RNNs), which process tokens
sequentially,
Transformers process all tokens in parallel and thus lack an inherent understanding of token order.
Positional encoding addresses this challenge by encoding positional information into the token
embeddings.
Vibe coding
NLP
LLMs Part 3: Understand long context through AI
The attention mechanism revolutionized NLP by enabling models to focus on important parts of the input
sequence,
allowing for contextually accurate responses even in long sequences.
Introduced in the 2017 paper "Attention Is All You Need" by Vaswani et al., self-attention has become a
cornerstone of
modern AI and the transformer architecture of LLMs.
Vibe coding
NLP
Learning
LLMs Part 4: Put the pieces together. The simplest LM imaginable
Let us now demystify the intricacies of Large Language Models (LLMs) and bring theory to life with
an interactive simulation.
Watch in real-time as this basic yet fascinating transformer model runs a live simulation,
piecing together language constructs with the elegance of simplicity.
Vibe coding
NLP
Learning
Approximation
To be continued ...
Covered AI topics
Vibe coding
"Vibe coding" is a term coined by Andrej Karpathy in early 2025 to describe a conversational and intuitive
approach to programming using AI tools, particularly large language models (LLMs).
Search AI systems excel at retrieving information quickly and
efficiently, whether from
databases or the web.
NLP Understanding the meaning and context of language
and data is crucial for tasks like
natural language processing, enabling machines to understand, interpret, and generate human language.
Compression Reducing the size of data while preserving
important information is key for
efficient storage and transmission.
Planning AI can develop strategies and plan actions to achieve
specific goals, particularly in
robotics and game playing.
Optimization AI algorithms often seek to find the best
solution to a problem, whether it's
minimizing costs or maximizing efficiency.
Approximation AI models approximate complex functions and
patterns in data, allowing for
predictions and decision-making.
Logical Reasoning AI can apply logical rules to draw
inferences and make decisions based on
available information.
Perception AI systems process sensory information, such as
visual, auditory, and tactile data,
to understand and interpret the world. This is crucial for tasks like image recognition and speech
processing.
Learning AI can adapt and improve from experience through
various learning methods, such as
supervised, unsupervised, and reinforcement learning. This capability is essential for tasks like pattern
recognition and predictive analytics.
Interaction AI involves communication with humans or other
systems through natural language
processing, dialogue systems, and user interfaces. This is key for applications like virtual assistants and
customer service bots.