What is ChatGPT? What do G, P, and T stand for?
Bill Gates: ChatGPT is the most revolutionary technological advancement since 1980.
In this era of AI transformation, we can only join in and keep up.
This is my study note, I hope it will be helpful for you to understand ChatGPT.
1. What do the GPTs in ChatGPT stand for?
GPT, Generative Pre-trained Transformer, generative pre-trained transformation model.
What's the meaning?
Generative means that it can generate content spontaneously.
Pre-trained, pre-training, does not require you to get it and train again. It directly prepares a general language model for you.
Transformer, transformation model, is a very powerful model proposed by Google. It can help better deal with NLP related problems. It is a very good neural network structure.
2. Although Transformer was proposed by Google. But the most successful application is OpenAI’s ChatGPT.
Because ChatGPT stands on the shoulders of giants.
ChatGPT is the crystallization of all human society. Without the paving the way, ChatGPT would not have come out so smoothly.
There will be more very powerful applications in the future, which will be based on ChatGPT.
3. There is a very important concept in Transformer, the attention mechanism.
What is the attention mechanism?
It is to pick out the important information from the information you input, focus on this important information, and ignore the unimportant information. This will help you better understand what you are saying.
The attention mechanism can help the Transformer model focus on the most important part of the input information.
4. Machine learning methods are divided into supervised learning, unsupervised learning, and reinforcement learning.
Supervised learning: There is labeled data, direct feedback can be given, and results and the future can be predicted.
Unsupervised learning: There are no labels, goals, and no feedback, but it searches for the hidden meanings in the data by itself. Results
Reinforcement learning: It is a decision-making process with a series of reward mechanisms and punishment mechanisms to help machine learning do better
ChatGPT uses unsupervised learning and reinforcement learning.
5. ChatGPT can generate and create a large amount of content, but it actually relies on guessing probability.
For example, the weather is gray and my mood is very___
The AI trained with a large amount of data will predict that the word with the highest probability of appearing in this space is "depressed".
Then "depressed" will be filled in this blank, so the answer is:
The weather is gray and gray, and I feel very depressed
This feels incredible , but that’s the fact.
All NLP (natural language processing) tasks at this stage are not machines that can truly understand the human world.
He was just playing word games and solving probability puzzles again and again.
#6. In this word game of "guessing the probability", the large prediction model (LLM, Large Language Model) has evolved into the two most mainstream directions: BERT and GPT.
BERT was the most popular direction before, dominating almost all NLP fields.
And perform well in natural language understanding tasks (such as text classification, emotional tendency judgment, etc.).
The GPT direction is relatively weak, and the most well-known player is OpenAl.
In fact, before the release of GPT3.0, the direction of GPT has always been weaker than BERT (GPT3.0 is the predecessor of GPT3.5, the model behind ChatGPT).
7. What is the difference between BERT and GPT?
BERT is a two-way language model. It guesses the word in the middle before and after the connection, so it is two-way, just like cloze.
For example: I ___go home on the 20th
BERT guessed "I plan to go home on the 20th" and guessed the "plan" in the middle.
GPT is a one-way language model, guessing the next word, so it is one-way, like writing a composition.
For example: I plan to go home on the 20th___
GPT guessed "I plan to go home on the 20th", and guessed the word "go home" after it.
8. How to ask questions to GPT?
There are two ways: fine-tune and prompt.
fine-tune, parameter adjustment: The model parameters need to be updated to complete the generated content.
fine-tune is professional, with high threshold and small audience. However, it has high diversity and accuracy and is suitable for complex tasks. A game for a few players.
prompt, prompt word: No need to modify the model and parameters, just give some tips and examples to complete the task.
prompt is simpler, the threshold is low, and the audience is large. Suitable for simple tasks. All players are.
The content we enter in the ChatGPT input box is the prompt.
9. ChatGPT is generative AI.
AI is divided into two types in terms of content production methods: analytical AI and generative AI.
Analytical AI is mainly used for analysis and classification. How much data you feed it, what kind of content it can analyze, it is limited to the data itself.
Generative AI creates new content that does not exist in the data based on learning and summarizing the distribution of data. Can generate text, pictures, code, speech synthesis, video and 3D models.
ChatGPT is a generative AI that is best at text and writing code.
10. Finally, from the perspective of knowledge acquisition, ChatGPT is a new generation of “knowledge representation and invocation method”.
In the early days, knowledge was stored in a structured manner in databases. We get it through SQL.
Later, with the birth of the Internet, more unstructured knowledge such as texts, pictures, and videos were stored on the Internet. We get it through search engines.
Now, knowledge is stored in the large model in the form of parameters. We directly invoke this knowledge using natural language.
The above is the detailed content of What is ChatGPT? What do G, P, and T stand for?. For more information, please follow other related articles on the PHP Chinese website!

Hot AI Tools

Undresser.AI Undress
AI-powered app for creating realistic nude photos

AI Clothes Remover
Online AI tool for removing clothes from photos.

Undress AI Tool
Undress images for free

Clothoff.io
AI clothes remover

AI Hentai Generator
Generate AI Hentai for free.

Hot Article

Hot Tools

Notepad++7.3.1
Easy-to-use and free code editor

SublimeText3 Chinese version
Chinese version, very easy to use

Zend Studio 13.0.1
Powerful PHP integrated development environment

Dreamweaver CS6
Visual web development tools

SublimeText3 Mac version
God-level code editing software (SublimeText3)

Hot Topics



DALL-E 3 was officially introduced in September of 2023 as a vastly improved model than its predecessor. It is considered one of the best AI image generators to date, capable of creating images with intricate detail. However, at launch, it was exclus

DDREASE is a tool for recovering data from file or block devices such as hard drives, SSDs, RAM disks, CDs, DVDs and USB storage devices. It copies data from one block device to another, leaving corrupted data blocks behind and moving only good data blocks. ddreasue is a powerful recovery tool that is fully automated as it does not require any interference during recovery operations. Additionally, thanks to the ddasue map file, it can be stopped and resumed at any time. Other key features of DDREASE are as follows: It does not overwrite recovered data but fills the gaps in case of iterative recovery. However, it can be truncated if the tool is instructed to do so explicitly. Recover data from multiple files or blocks to a single

0.What does this article do? We propose DepthFM: a versatile and fast state-of-the-art generative monocular depth estimation model. In addition to traditional depth estimation tasks, DepthFM also demonstrates state-of-the-art capabilities in downstream tasks such as depth inpainting. DepthFM is efficient and can synthesize depth maps within a few inference steps. Let’s read about this work together ~ 1. Paper information title: DepthFM: FastMonocularDepthEstimationwithFlowMatching Author: MingGui, JohannesS.Fischer, UlrichPrestel, PingchuanMa, Dmytr

The performance of JAX, promoted by Google, has surpassed that of Pytorch and TensorFlow in recent benchmark tests, ranking first in 7 indicators. And the test was not done on the TPU with the best JAX performance. Although among developers, Pytorch is still more popular than Tensorflow. But in the future, perhaps more large models will be trained and run based on the JAX platform. Models Recently, the Keras team benchmarked three backends (TensorFlow, JAX, PyTorch) with the native PyTorch implementation and Keras2 with TensorFlow. First, they select a set of mainstream

Facing lag, slow mobile data connection on iPhone? Typically, the strength of cellular internet on your phone depends on several factors such as region, cellular network type, roaming type, etc. There are some things you can do to get a faster, more reliable cellular Internet connection. Fix 1 – Force Restart iPhone Sometimes, force restarting your device just resets a lot of things, including the cellular connection. Step 1 – Just press the volume up key once and release. Next, press the Volume Down key and release it again. Step 2 – The next part of the process is to hold the button on the right side. Let the iPhone finish restarting. Enable cellular data and check network speed. Check again Fix 2 – Change data mode While 5G offers better network speeds, it works better when the signal is weaker

I cry to death. The world is madly building big models. The data on the Internet is not enough. It is not enough at all. The training model looks like "The Hunger Games", and AI researchers around the world are worrying about how to feed these data voracious eaters. This problem is particularly prominent in multi-modal tasks. At a time when nothing could be done, a start-up team from the Department of Renmin University of China used its own new model to become the first in China to make "model-generated data feed itself" a reality. Moreover, it is a two-pronged approach on the understanding side and the generation side. Both sides can generate high-quality, multi-modal new data and provide data feedback to the model itself. What is a model? Awaker 1.0, a large multi-modal model that just appeared on the Zhongguancun Forum. Who is the team? Sophon engine. Founded by Gao Yizhao, a doctoral student at Renmin University’s Hillhouse School of Artificial Intelligence.

Recently, the military circle has been overwhelmed by the news: US military fighter jets can now complete fully automatic air combat using AI. Yes, just recently, the US military’s AI fighter jet was made public for the first time and the mystery was unveiled. The full name of this fighter is the Variable Stability Simulator Test Aircraft (VISTA). It was personally flown by the Secretary of the US Air Force to simulate a one-on-one air battle. On May 2, U.S. Air Force Secretary Frank Kendall took off in an X-62AVISTA at Edwards Air Force Base. Note that during the one-hour flight, all flight actions were completed autonomously by AI! Kendall said - "For the past few decades, we have been thinking about the unlimited potential of autonomous air-to-air combat, but it has always seemed out of reach." However now,

New SOTA for multimodal document understanding capabilities! Alibaba's mPLUG team released the latest open source work mPLUG-DocOwl1.5, which proposed a series of solutions to address the four major challenges of high-resolution image text recognition, general document structure understanding, instruction following, and introduction of external knowledge. Without further ado, let’s look at the effects first. One-click recognition and conversion of charts with complex structures into Markdown format: Charts of different styles are available: More detailed text recognition and positioning can also be easily handled: Detailed explanations of document understanding can also be given: You know, "Document Understanding" is currently An important scenario for the implementation of large language models. There are many products on the market to assist document reading. Some of them mainly use OCR systems for text recognition and cooperate with LLM for text processing.
