100% satisfaction guarantee Immediately available after payment Both online and in PDF No strings attached 4.2 TrustPilot
logo-home
Summary

Summary Large language models

Rating
-
Sold
-
Pages
2
Uploaded on
25-07-2023
Written in
2022/2023

Large language models introduction

Institution
Course








Whoops! We can’t load your doc right now. Try again or contact support.

Connected book

Written for

Course

Document information

Summarized whole book?
No
Which chapters are summarized?
Chapter i
Uploaded on
July 25, 2023
Number of pages
2
Written in
2022/2023
Type
Summary

Subjects

Content preview

LARGE LANGUAGE MODELS - Prompt engineering




Large language models refer to advanced neural network-based architectures that have been
trained on massive amounts of text data to process and understand human language. These
models have shown remarkable capabilities in various natural language processing (NLP) tasks,
such as language generation, translation, question-answering, sentiment analysis, and more.
They are typically characterized by having tens of billions of parameters, allowing them to
capture complex language patterns and generate coherent and contextually relevant responses.




Some of the well-known large language models include:




GPT-3 (Generative Pre-trained Transformer 3): Developed by OpenAI, GPT-3 is one of the most
famous and largest language models, with 175 billion parameters. It has demonstrated
impressive performance across a wide range of NLP tasks and can generate human-like text.




BERT (Bidirectional Encoder Representations from Transformers): Developed by Google, BERT
is another influential language model with 340 million parameters. It introduced the concept of
bidirectional training and context-based word embeddings, leading to significant improvements
in many NLP tasks.




T5 (Text-to-Text Transfer Transformer): Developed by Google, T5 is a large model that frames
all NLP tasks as a text-to-text problem. It has 11 billion parameters and has shown strong
performance in a multitude of NLP tasks.




XLNet: Developed by Google, XLNet is a generalized autoregressive pre-training method that
leverages both autoregressive and autoencoding objectives. It has 340 million parameters and
has achieved state-of-the-art results in various NLP benchmarks.




RoBERTa (A Robustly Optimized BERT Pretraining Approach): A variation of BERT developed by
Facebook AI, RoBERTa uses a larger batch size and more training data to achieve better
performance across multiple NLP tasks.
$7.99
Get access to the full document:

100% satisfaction guarantee
Immediately available after payment
Both online and in PDF
No strings attached

Get to know the seller
Seller avatar
ramakrishnatammineedi

Get to know the seller

Seller avatar
ramakrishnatammineedi Published
Follow You need to be logged in order to follow users or courses
Sold
0
Member since
2 year
Number of followers
0
Documents
5
Last sold
-

0.0

0 reviews

5
0
4
0
3
0
2
0
1
0

Recently viewed by you

Why students choose Stuvia

Created by fellow students, verified by reviews

Quality you can trust: written by students who passed their tests and reviewed by others who've used these notes.

Didn't get what you expected? Choose another document

No worries! You can instantly pick a different document that better fits what you're looking for.

Pay as you like, start learning right away

No subscription, no commitments. Pay the way you're used to via credit card and download your PDF document instantly.

Student with book image

“Bought, downloaded, and aced it. It really can be that simple.”

Alisha Student

Frequently asked questions