- Himanshu Ramchandani
- Posts
- ๐ฑ ๐๐น๐ฎ๐ฟ๐บ๐ถ๐ป๐ด ๐๐ต๐ถ๐ป๐ด๐ ๐๐ผ๐ ๐บ๐๐๐ ๐ธ๐ป๐ผ๐ ๐ฎ๐ฏ๐ผ๐๐ ๐๐ฒ๐ปe๐ฟ๐ฎ๐๐ถ๐๐ฒ๐๐ ๐ฎ๐ ๐ฎ ๐๐ฒ๐ฎ๐ฑ๐ฒ๐ฟ๐
๐ฑ ๐๐น๐ฎ๐ฟ๐บ๐ถ๐ป๐ด ๐๐ต๐ถ๐ป๐ด๐ ๐๐ผ๐ ๐บ๐๐๐ ๐ธ๐ป๐ผ๐ ๐ฎ๐ฏ๐ผ๐๐ ๐๐ฒ๐ปe๐ฟ๐ฎ๐๐ถ๐๐ฒ๐๐ ๐ฎ๐ ๐ฎ ๐๐ฒ๐ฎ๐ฑ๐ฒ๐ฟ๐
Microsoft Copilot, GenAI for Leaders, ML Lead Jobs
Welcome Back, Hero!
Hope you are doing well.
Upgraded some things in the newsletter, hope you like it!
Todayโs Content โ
How to Leverage Data, Products & AI for Your Business ๐ข
๐ฑ ๐๐น๐ฎ๐ฟ๐บ๐ถ๐ป๐ด ๐๐ต๐ถ๐ป๐ด๐ ๐๐ผ๐ ๐บ๐๐๐ ๐ธ๐ป๐ผ๐ ๐ฎ๐ฏ๐ผ๐๐ ๐๐ฒ๐ปe๐ฟ๐ฎ๐๐ถ๐๐ฒ๐๐ ๐ฎ๐ ๐ฎ ๐๐ฒ๐ฎ๐ฑ๐ฒ๐ฟ๐1 Action Tip from Data Experts for Leaders ๐ฌโ Start Small
For Developers ๐งโ๐ป โ Microsoft Copilot
Career & Job in the AI field ๐ โ ML Lead
How to Leverage Data, Products & AI for Your Business ๐ข
๐ฑ ๐๐น๐ฎ๐ฟ๐บ๐ถ๐ป๐ด ๐๐ต๐ถ๐ป๐ด๐ ๐๐ผ๐ ๐บ๐๐๐ ๐ธ๐ป๐ผ๐ ๐ฎ๐ฏ๐ผ๐๐ ๐๐ฒ๐ปe๐ฟ๐ฎ๐๐ถ๐๐ฒ๐๐ ๐ฎ๐ ๐ฎ ๐๐ฒ๐ฎ๐ฑ๐ฒ๐ฟ๐
GenerativeAI talk by Himanshu Ramchandani
GenerativeAI Revolution
1 โ Behind NLP
โ The objective of NLP is to make computers understand what we are trying to day.
โ We have to make sure it understands 2 things - context and intent.
2 โ Context and Intent
Context is if I tell you - โYour T-shirt is killerโ
We as humans take this example as a compliment.
The โkillerโ word for the computer is a negative.
But the context is positive.
Suppose you tell a vegetable seller that you got the money from your mother, and she wants me to buy tomatoes from the market, otherwise, she will be angry.
In the whole part, the vegetable seller will understand that you need tomatoes.
The whole story behind it is not important to the seller.
Thats Intent!
Real-world examples of NLP โ
- spam detection
- social media sentiment analysis
- machine translation
- text summarization
- semantic search
3 โ GenerativeAI had 2 biggest milestones
1 - Artificial Neurons
In deep learning, millions of these are used in dense neural networks.
2 - Pre-trained models
We can reuse the model weights so that we don't have to train it again and again.
4 โ GPT-3 - letโs break it down
โ Generative
โ Generative means generating text (GPT3).
โ Generative Modeling is the branch of statistical modeling.
โ In order for the model to generate text similar to ours, we need a lot of data.
โ Pre-Trained
โ Training a model takes a lot of time as well as resources along with cost.
โ Using our trained model will allow us to fine-tune it for downstream tasks.
โ Understanding chess rules will help us compete with others in a match.
GPT3 is trained on 5 datasets
- Common Crawl
- WebText2
- Books1
- Books2
- Wikipedia.
โ Transformers
โ Transformers work on attention mechanisms.
Example - โLakshya is sitting on a chair. He will take the lunch nowโ
In this statement, we know that the word โHeโ is used for Lakshya. That's attention.
โ Transformer works on the sequence-to-sequence architecture.
โ A sequence of words in one language is translated into a sequence of words in another language.
Thatโs all, GPT
5 โ The versions
The version 3 has 175 billion text parameters.
GPT-2 had 1.5 billion parameters
GPT-1 had 117 million parameters
PS โ Reach out to me if you need help with GenAI for product development, Data Science, Data Engineering, and Infra on Cloud โ https://topmate.io/himanshuramchandani/447594
1 Action Tip for Data & AI Leaders ๐ฌ
Are you starting small, and starting now?
GenerativeAI is complex and resource-intensive.
Starting small has several advantages - low risk, quick learning, demonstrable ROI, and Iterative improvement.
Don't wait for the perfect moment, start experimenting with generative AI now and grow your capabilities over time
All the best!
Career & Job in the AI field๐
Cased, AI-powered tools for DevOps โ ML Lead (Remote or San Francisco)
Community & Connections๐
Join 20,000+ data professionals on LinkedIn โ Himanshu Ramchandani
Until next Week!
PS: if you want to build your own newsletter, โ Here
Reply