Table of Contents
Having a place in artificial intelligence models
growing family of large language models
Home Technology peripherals AI ChatGPT vs. GPT-3 vs. GPT-4 is just an internal fight among chatbot families

ChatGPT vs. GPT-3 vs. GPT-4 is just an internal fight among chatbot families

Mar 31, 2023 pm 10:39 PM
AI chatgpt chatbot

When explaining SoMin’s ad copy and banner generation capabilities, people often ask whether ChatGPT replaced GPT-3 or whether it is still running an outdated mode.

When explaining SoMin’s ad copy and banner generation capabilities, people often ask whether ChatGPT replaced GPT-3 or whether it is still running an outdated mode. "We have not and do not plan to do so." A SoMin spokesperson responded, even though the chatbot ChatGPT launched by OpenAI is booming. This often surprises the customer, so here’s an explanation of why he would give such an answer.

ChatGPT vs. GPT-3 vs. GPT-4 is just an internal fight among chatbot families

Having a place in artificial intelligence models

GPT-2, GPT-3, ChatGPT and the recently launched GPT-4 all belong to the same category of artificial intelligence Model - Transformer. This means that, unlike previous generation machine learning models, they are trained to complete a more uniform task, so they do not need to be retrained for each specific task to produce actionable results. The latter explains their massive size (175 billion parameters in the case of GPT-3), while a model might need to "remember the entire internet" to be flexible enough to switch between different pieces of data based on user input . The model is then able to generate results when the user enters a query question, a description of the task, and a few examples (like you would ask a librarian for books of interest). This approach is called "few-shot learning" and has become a trend recently in providing input to modern Transformer models.

But is it always necessary to know everything about the Internet in order to complete the current task? Of course not - in many cases, like ChatGPT, a large number (millions) of specific Data samples for the task that will allow the model to initiate the Reinforcement Learning from Human Feedback (RLHF) process. In turn, RLHF will derive a collaborative training process between AI and humans to further train AI models to produce human-like conversations. Therefore, ChatGPT not only excels in the chatbot scenario, but also helps people write short-form content (such as poems or lyrics) or long-form content (such as essays); when people need to get answers quickly, in simple terms or in-depth knowledge Explain complex topics; provide brainstorming, new topics and ideas, which are helpful in the creative process, support the sales department in personalized communication, such as generating emails to respond to.

While it is technically possible for a large Transformer model to attempt to accomplish these tasks, it is unlikely to be accomplished by ChatGPT or even GPT-4 - this is due to ChatGPT and other OpenAI's Transformers' knowledge of events occurring in the world. Very limited because they are pretrained models and therefore their data is not updated frequently enough due to the very large computational demands of model retraining. This is probably the biggest shortcoming of all pre-trained models produced by OpenAI (and indeed anyone else) to date. A bigger problem is specific to ChatGPT: unlike GPT-3, it was trained on a very focused conversational dataset, so it is only in conversational tasks that ChatGPT outperforms its predecessors, while completing other human tasks. When it comes to productivity tasks, it's less advanced.

growing family of large language models

People now know that ChatGPT is just a smaller, more specific version of GPT-3, but does this mean there will be more in the near future? Such models emerge: MarGPT for marketing, AdGPT for digital advertising, MedGPT for answering medical questions?

This is possible, and here’s why: When the company SoMin submits an application to When gaining access to the GPT-3 Beta, despite filling out a lengthy application form explaining in detail the current software that would be built, I was asked to agree to provide feedback on how the model was used on a daily basis and the results received. The company OpenAI did this for a reason, mainly because it was a research project and they needed commercial insights into the best applications of the model, and they crowdfunded it in exchange for the chance to participate in this great artificial intelligence revolution. Chatbot apps seem to be one of the most popular, so ChatGPT comes first. ChatGPT is not only smaller (20 billion parameters vs. 175 billion parameters), but also faster than GPT-3, and more accurate than GPT-3 at solving conversational tasks - for a low-cost/high-quality AI product For me, this is a perfect business case.

So, for generative artificial intelligence, is bigger better? The answer is, it depends. When one is building a general learning model capable of completing many tasks, the answer is yes, the bigger the better, as evidenced by GPT-3’s advantages over GPT-2 and other predecessors. But when one wants to perform a specific task well, like the chatbot in ChatGPT, then data focus and a proper training process are much more important than model and data size. That’s why at SoMin, instead of using ChatGPT to generate copy and banners, specific digital ad-related data is used to guide GPT-3 to create better content for new ads that haven’t been seen yet.

So, one might ask, how will the future of generative AI develop? Multimodality will be one of the inevitable advancements people will see in the upcoming GPT-4, as OpenAI CEO Sam Altman mentioned in his speech. At the same time, Altman also broke the rumor that the model has 100 trillion parameters. Therefore, people know that bigger does not always mean better in this kind of artificial intelligence model.

The above is the detailed content of ChatGPT vs. GPT-3 vs. GPT-4 is just an internal fight among chatbot families. For more information, please follow other related articles on the PHP Chinese website!

Statement of this Website
The content of this article is voluntarily contributed by netizens, and the copyright belongs to the original author. This site does not assume corresponding legal responsibility. If you find any content suspected of plagiarism or infringement, please contact admin@php.cn

Hot AI Tools

Undresser.AI Undress

Undresser.AI Undress

AI-powered app for creating realistic nude photos

AI Clothes Remover

AI Clothes Remover

Online AI tool for removing clothes from photos.

Undress AI Tool

Undress AI Tool

Undress images for free

Clothoff.io

Clothoff.io

AI clothes remover

AI Hentai Generator

AI Hentai Generator

Generate AI Hentai for free.

Hot Tools

Notepad++7.3.1

Notepad++7.3.1

Easy-to-use and free code editor

SublimeText3 Chinese version

SublimeText3 Chinese version

Chinese version, very easy to use

Zend Studio 13.0.1

Zend Studio 13.0.1

Powerful PHP integrated development environment

Dreamweaver CS6

Dreamweaver CS6

Visual web development tools

SublimeText3 Mac version

SublimeText3 Mac version

God-level code editing software (SublimeText3)

ChatGPT now allows free users to generate images by using DALL-E 3 with a daily limit ChatGPT now allows free users to generate images by using DALL-E 3 with a daily limit Aug 09, 2024 pm 09:37 PM

DALL-E 3 was officially introduced in September of 2023 as a vastly improved model than its predecessor. It is considered one of the best AI image generators to date, capable of creating images with intricate detail. However, at launch, it was exclus

Bytedance Cutting launches SVIP super membership: 499 yuan for continuous annual subscription, providing a variety of AI functions Bytedance Cutting launches SVIP super membership: 499 yuan for continuous annual subscription, providing a variety of AI functions Jun 28, 2024 am 03:51 AM

This site reported on June 27 that Jianying is a video editing software developed by FaceMeng Technology, a subsidiary of ByteDance. It relies on the Douyin platform and basically produces short video content for users of the platform. It is compatible with iOS, Android, and Windows. , MacOS and other operating systems. Jianying officially announced the upgrade of its membership system and launched a new SVIP, which includes a variety of AI black technologies, such as intelligent translation, intelligent highlighting, intelligent packaging, digital human synthesis, etc. In terms of price, the monthly fee for clipping SVIP is 79 yuan, the annual fee is 599 yuan (note on this site: equivalent to 49.9 yuan per month), the continuous monthly subscription is 59 yuan per month, and the continuous annual subscription is 499 yuan per year (equivalent to 41.6 yuan per month) . In addition, the cut official also stated that in order to improve the user experience, those who have subscribed to the original VIP

Can fine-tuning really allow LLM to learn new things: introducing new knowledge may make the model produce more hallucinations Can fine-tuning really allow LLM to learn new things: introducing new knowledge may make the model produce more hallucinations Jun 11, 2024 pm 03:57 PM

Large Language Models (LLMs) are trained on huge text databases, where they acquire large amounts of real-world knowledge. This knowledge is embedded into their parameters and can then be used when needed. The knowledge of these models is "reified" at the end of training. At the end of pre-training, the model actually stops learning. Align or fine-tune the model to learn how to leverage this knowledge and respond more naturally to user questions. But sometimes model knowledge is not enough, and although the model can access external content through RAG, it is considered beneficial to adapt the model to new domains through fine-tuning. This fine-tuning is performed using input from human annotators or other LLM creations, where the model encounters additional real-world knowledge and integrates it

To provide a new scientific and complex question answering benchmark and evaluation system for large models, UNSW, Argonne, University of Chicago and other institutions jointly launched the SciQAG framework To provide a new scientific and complex question answering benchmark and evaluation system for large models, UNSW, Argonne, University of Chicago and other institutions jointly launched the SciQAG framework Jul 25, 2024 am 06:42 AM

Editor |ScienceAI Question Answering (QA) data set plays a vital role in promoting natural language processing (NLP) research. High-quality QA data sets can not only be used to fine-tune models, but also effectively evaluate the capabilities of large language models (LLM), especially the ability to understand and reason about scientific knowledge. Although there are currently many scientific QA data sets covering medicine, chemistry, biology and other fields, these data sets still have some shortcomings. First, the data form is relatively simple, most of which are multiple-choice questions. They are easy to evaluate, but limit the model's answer selection range and cannot fully test the model's ability to answer scientific questions. In contrast, open-ended Q&A

SOTA performance, Xiamen multi-modal protein-ligand affinity prediction AI method, combines molecular surface information for the first time SOTA performance, Xiamen multi-modal protein-ligand affinity prediction AI method, combines molecular surface information for the first time Jul 17, 2024 pm 06:37 PM

Editor | KX In the field of drug research and development, accurately and effectively predicting the binding affinity of proteins and ligands is crucial for drug screening and optimization. However, current studies do not take into account the important role of molecular surface information in protein-ligand interactions. Based on this, researchers from Xiamen University proposed a novel multi-modal feature extraction (MFE) framework, which for the first time combines information on protein surface, 3D structure and sequence, and uses a cross-attention mechanism to compare different modalities. feature alignment. Experimental results demonstrate that this method achieves state-of-the-art performance in predicting protein-ligand binding affinities. Furthermore, ablation studies demonstrate the effectiveness and necessity of protein surface information and multimodal feature alignment within this framework. Related research begins with "S

SK Hynix will display new AI-related products on August 6: 12-layer HBM3E, 321-high NAND, etc. SK Hynix will display new AI-related products on August 6: 12-layer HBM3E, 321-high NAND, etc. Aug 01, 2024 pm 09:40 PM

According to news from this site on August 1, SK Hynix released a blog post today (August 1), announcing that it will attend the Global Semiconductor Memory Summit FMS2024 to be held in Santa Clara, California, USA from August 6 to 8, showcasing many new technologies. generation product. Introduction to the Future Memory and Storage Summit (FutureMemoryandStorage), formerly the Flash Memory Summit (FlashMemorySummit) mainly for NAND suppliers, in the context of increasing attention to artificial intelligence technology, this year was renamed the Future Memory and Storage Summit (FutureMemoryandStorage) to invite DRAM and storage vendors and many more players. New product SK hynix launched last year

Laying out markets such as AI, GlobalFoundries acquires Tagore Technology's gallium nitride technology and related teams Laying out markets such as AI, GlobalFoundries acquires Tagore Technology's gallium nitride technology and related teams Jul 15, 2024 pm 12:21 PM

According to news from this website on July 5, GlobalFoundries issued a press release on July 1 this year, announcing the acquisition of Tagore Technology’s power gallium nitride (GaN) technology and intellectual property portfolio, hoping to expand its market share in automobiles and the Internet of Things. and artificial intelligence data center application areas to explore higher efficiency and better performance. As technologies such as generative AI continue to develop in the digital world, gallium nitride (GaN) has become a key solution for sustainable and efficient power management, especially in data centers. This website quoted the official announcement that during this acquisition, Tagore Technology’s engineering team will join GLOBALFOUNDRIES to further develop gallium nitride technology. G

A new era of VSCode front-end development: 12 highly recommended AI code assistants A new era of VSCode front-end development: 12 highly recommended AI code assistants Jun 11, 2024 pm 07:47 PM

In the world of front-end development, VSCode has become the tool of choice for countless developers with its powerful functions and rich plug-in ecosystem. In recent years, with the rapid development of artificial intelligence technology, AI code assistants on VSCode have sprung up, greatly improving developers' coding efficiency. AI code assistants on VSCode have sprung up like mushrooms after a rain, greatly improving developers' coding efficiency. It uses artificial intelligence technology to intelligently analyze code and provide precise code completion, automatic error correction, grammar checking and other functions, which greatly reduces developers' errors and tedious manual work during the coding process. Today, I will recommend 12 VSCode front-end development AI code assistants to help you in your programming journey.

See all articles