Open in app

Sign In

Write

Sign In

Mastodon
Chamanth mvs
Chamanth mvs

94 Followers

Home

Lists

About

Published in

AI Mind

·Aug 17

LLM and Fine-Tuning

Understanding more about Large Language models — Large Language Model (LLM) is kind of powerful neural network, which enable computers to understand and generate human language. LLM is based on Transformers architecture. These LLMs are trained on massive datasets and huge corpus of internet data, which showed impressive results like understanding complex nuanced language and generating language…

Large Language Models

12 min read

LLM and Fine-Tuning
LLM and Fine-Tuning
Large Language Models

12 min read


Published in

data-driven fiction

·Jul 30

A brief on GPT-2 and GPT-3 models

Summarizing OpenAI ’s GPT-2 and GPT-3 models — In one of the last article, I discussed about Decoder-only transformer model, which is GPT-1 model from OpenAI. Decoder-only Transformer model Understanding Large Language models with GPT-1medium.com GPT-2 model is designed by making minor changes in the architecture of GPT-1 model. Similarly, GPT-3 model is slightly an enhanced version of GPT-2 model.

Generative Ai

6 min read

A brief on GPT-2 and GPT-3 models
A brief on GPT-2 and GPT-3 models
Generative Ai

6 min read


Published in

data-driven fiction

·Jul 29

A step into Zero-Shot Learning

A conceptual understanding of Zero shot learning — Zero-Shot Learning is a subfield of transfer learning, a machine learning method where a model developed for a one task is reused as the starting point for a model on a different task. On a broader level, there are two types of transfer learning: homogenous transfer learning — where the…

Zero Shot Learning

6 min read

A step into Zero-Shot Learning
A step into Zero-Shot Learning
Zero Shot Learning

6 min read


Published in

AI Mind

·Jul 24

*args and **kwargs explained

A detailed view on positional and keyword arguments — To understand about *args and **kwargs, you should have an idea about Pass by reference, positional arguments, keyword arguments, packed values and unpacked values. These are the articles, which are recommended to read before following through this article. Parameter vs Arguments and Pass by reference Understanding of Pass by reference with module scope and function scopepub.aimind.so

Programming

12 min read

*args and **kwargs explained
*args and **kwargs explained
Programming

12 min read


Published in

AI Mind

·Jul 19

Parameter vs Arguments and Pass by reference

Understanding of Pass by reference with module scope and function scope — Function is a sequence of program instructions that performs a specific task, packaged as a unit — wikipedia A function is a block of code that performs specific task. It can be reused multiple times. When a function call is made, the program will pause the current program and execute…

Python

3 min read

Parameter vs Arguments and Pass by reference
Parameter vs Arguments and Pass by reference
Python

3 min read


Published in

AI Mind

·Jul 16

Positional arguments and keyword arguments

A detailed explanation on positional arguments and keyword arguments in python — Positional arguments The most common way of assigning arguments to parameters is through the order in which they are passed in (that is through their position). A function example_func() is defined with two parameters x and y def example_func(x, y): # body of the function return None When this function call is…

Python

5 min read

Positional arguments and keyword arguments
Positional arguments and keyword arguments
Python

5 min read


Published in

AI Mind

·Jun 28

Detailed view of BERT

Exploring Large Language models with underlying concepts — Bi-directional Encoder Representation from Transformers — is proposed by Google in the research paper — BERT-Pre-training of Deep Bidirectional Transformers for Language Understanding, which became revolutionary in the field of NLP (Natural Language Processing). The biggest challenge in NLP-based tasks is the lack of enough training data. Even though there…

Transformers

12 min read

Detailed view of BERT
Detailed view of BERT
Transformers

12 min read


Published in

data-driven fiction

·Jun 18

Decoder-only Transformer model

Understanding Large Language models with GPT-1 — Over past 6 months, everywhere around the internet — ChatGPT was the term buzzing around and it is in fact acknowledged that the next-decade will be the decade of generative models. There are many Generative models like Autoregressive models, Variational autoencoders, Normalizing flow models, Generative adversarial networks, etc.. ChatGPT could…

Transformers

6 min read

Decoder-only Transformer model
Decoder-only Transformer model
Transformers

6 min read


Published in

DataDrivenInvestor

·Jun 12

Self-Attention is not typical Attention model

Understanding Transformer model — Self-Attention is at the core of Transformer model and in fact, the entire LLM models mechanisms relies on the Transformer model. This is one of the trickiest model, I studied till date. Transformer model is an advanced sequence to sequence model, which consists of group of Encoder-Decoder models BUT, the…

Transformers

14 min read

Self-Attention is not typical Attention model
Self-Attention is not typical Attention model
Transformers

14 min read


Published in

Artificial Intelligence in Plain English

·May 27

Detailed explanation about Attention mechanism

Sequence-Sequence model with Attention mechanism — There are Encoder-Decoder — seq2seq models which are capable in solving many problems like Machine Translation, Image captioning and many more. Then, Why do we need enhanced sequence-to-sequence models like Attention model? At least once, everyone of you might have experienced Translation speech either in political meetings or in movies…

Attention Mechanism

7 min read

Detailed explanation about Attention mechanism
Detailed explanation about Attention mechanism
Attention Mechanism

7 min read

Chamanth mvs

Chamanth mvs

94 Followers

Data Science and ML practitioner | I share my learnings and thoughts here

Following
  • NYU Center for Data Science

    NYU Center for Data Science

  • Dariusz Gross #DATAsculptor

    Dariusz Gross #DATAsculptor

  • Alexandru-Ioan Plesoiu

    Alexandru-Ioan Plesoiu

  • Aman Kharwal

    Aman Kharwal

  • Zoheb Abai

    Zoheb Abai

See all (8)

Help

Status

Writers

Blog

Careers

Privacy

Terms

About

Text to speech

Teams