Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Written by Prashant Basnet
<section class="bg-white dark:bg-gray-900 px-4 py-8 max-w-2xl mx-auto text-gray-800 dark:text-gray-200">
<h1 class="text-2xl sm:text-3xl font-signature italic font-semibold text-center mb-4">
π Welcome β Youβve Landed on My Signature Page
</h1>
<p class="text-base sm:text-lg mb-4">
Hey, Iβm <strong class="text-black dark:text-white">Prashant Basnet</strong> β software developmemt engineer at
<a href="https://unisala.com" class="text-indigo-600 dark:text-indigo-400 underline hover:no-underline" target="_blank" rel="noopener noreferrer">
Unisala.com
</a>.
</p>
<p class="text-base sm:text-lg mb-6">
Youβre viewing my <strong>Signature</strong>, a digital space where I share what Iβm learning, building, and reflecting on, all in one place.
</p>
<div class="border-l-4 border-indigo-400 dark:border-indigo-500 pl-4 italic mb-6 text-sm sm:text-base text-gray-700 dark:text-gray-400">
π Found this page via LinkedIn, my personal site, or a shared link?
<br />
This isnβt a traditional portfolio. Itβs my public digital notebook where I document useful ideas, experiments, and lessons Iβve learned as I build.
</div>
<h2 class="text-lg font-semibold mb-2">What Youβll Find Here:</h2>
<ul class="list-disc list-inside space-y-1 text-sm sm:text-base">
<li>βοΈ Thoughts on algorithms, systems, and software design</li>
<li>π§ Insights from building at Unisala</li>
<li>π Direct links to everything Iβve published on Unisala</li>
</ul>
</section>
What GPT-3 Means:
GPT means generative pre-trained transformer, 3 means 3rd generation.
GPT 3 is just one example of transformer. Something that transforms from one sequence into another. Language translation is just a great example.
Machine Learning Basics:
A machine learning is very flexible structure with tun-able parameters like bunch of knob & dials, then somehow we use many examples of what the output should look like give an input, tweaking and tuning the values of these parameters to mimic this behaviour.
A simplest form of machine learning is linear regression, where input and output are a single number. For example sq.ft of a house and a price, we want a line of best fit throughout the data, that can predict future house prices.
Deep Learning:
When you create a giant model with huge number of parameters without it either grossly overfitting the training data or being too complex & unmanageable to train.
Deep learning describes a class of models that in the last decade have proven to scale remarkably well. What unifies them is they all use the same training algorithm, called back-propagation.
Transformers as a Model Class:
Input Data Formatting:
Transformers' Operation:
The only way parameters/weights interact with the data being processed is through weighted sums. We also sprinkle some simple non linear function throughout.These weighted sums are packaged together as various component in a matrix vector product.
Parameters in GPT-3:
Those 175 billion parameters/weights in GPT-3 are organised into 27,938 matrices.
These 27,983 matrices falls in 8 different categories.
These weights are the actual brain. They are the things learned during training. And they determine how a model behaves.
In the next thread we will look into details of these categories.
#AI #MachineLearning #DeepLearning #GPT3 #ArtificialIntelligence #Transformers #NeuralNetworks #DataScience #NaturalLanguageProcessing #NLP #DeepLearningModels #TechExplained #AIModels #MLBasics #GenerativeAI #TechEducation #AIInnovation #DataScienceCommunity #AIResearch #LearnAI #AIExplained #TechKnowledge #MLTechniques
this images & notes are taken from the following video, all credit goes to 3Blue1Brown