Table of Contents
1. GPT: If nothing happens, it will become a blockbuster
2. Comprehensively passed the Turing test to lower the threshold for learning and commercial use
3. Different paths lead to the same goal: simulating the human brain may come faster
Reference link:
Home Technology peripherals AI The imminent release of GPT4 is comparable to the human brain, and many industry leaders cannot sit still!

The imminent release of GPT4 is comparable to the human brain, and many industry leaders cannot sit still!

Apr 14, 2023 pm 06:58 PM
parameter gpt4 human brain

​Author | Xu Jiecheng

Reviewer | Yun Zhao​

100 trillion, how big is this number? If you have the superpower of winning 5 million lottery tickets every day, then if you save all the money without eating or drinking, you still need to live about 5,500 years to save 100 trillion in wealth. However, the 100 trillion I want to talk to you about today is not behind the coveted units such as “RMB” and “Dollor”. The 100 trillion here refers to the number of parameters owned by OpenAI, an artificial intelligence research company co-founded by many Silicon Valley technology tycoons, that is about to release the fourth generation of generative pre-training Transformer-GPT-4.

In order to facilitate everyone to understand this data more intuitively, we can use the human brain to compare with GPT-4. Under normal circumstances, a normal human brain has about 80-100 billion neurons and about 100 trillion synapses. These neurons and synapses directly control almost all thoughts, judgments and behaviors of a person in a hundred years of life, and GPT-4 has as many parameters as synapses in the human brain. So, what is the potential of such a large-scale dense neural network? What surprises will the emergence of GPT-4 bring to us? Do we really have the ability to create a human brain?

The imminent release of GPT4 is comparable to the human brain, and many industry leaders cannot sit still!

Before exploring these exciting questions, we might as well first understand the development history of several "predecessors" of GPT-4.

1. GPT: If nothing happens, it will become a blockbuster

The first GPT series model GPT-1 was born in 2018, which is us The first year of the often said NLP pre-training model. As the first pre-training model based on Transformer, GPT-1 adopts two stages of pre-training FineTuning, using Transformer's decoder as the feature extractor. It has a total of 12 layers stacked with 110 million parameters. The pre-training stage uses "one-way Language model" as a training task.

In terms of performance, GPT-1 has a certain generalization ability and can be used in NLP tasks that have nothing to do with supervision tasks. Common tasks include:

  • Natural language reasoning: determine the relationship between two sentences (containment, contradiction, neutrality)
  • Question and answer and common sense reasoning: input article And several answers, the accuracy of the output answer
  • Semantic similarity recognition: Determine whether the semantics of two sentences are related
  • Classification: Determine which category the input text belongs to

Although GPT-1 has some effects on untuned tasks, its generalization ability is much lower than that of fine-tuned supervised tasks, so GPT-1 can only be regarded as a pretty good one. language understanding tools rather than conversational AI.

One year after the advent of GPT-1, GPT-2 also arrived as scheduled in 2019. Compared with its big brother GPT-1, GPT-2 did not make too many structural innovations and designs on the original network. It only used more network parameters and a larger data set: the largest model has a total of 48 layers, and the parameters The amount reaches 1.5 billion, and the learning target uses an unsupervised pre-training model to perform supervised tasks.

The imminent release of GPT4 is comparable to the human brain, and many industry leaders cannot sit still!

Source: Twitter

In terms of performance, OpenAI’s great strength seems to really bring some miracles. In addition to its understanding ability, GPT-2 has for the first time shown a strong talent in generation: reading summaries, chatting, continuing to write, making up stories, and even generating fake news, phishing emails or pretending to be others online is all a breeze. After "becoming bigger", GPT-2 did demonstrate a series of universal and powerful capabilities, and achieved the best performance at the time on multiple specific language modeling tasks. It’s no wonder that OpenAI said at the time that “GPT-2 was too dangerous to release.”

Since the success of GPT-2, OpenAI’s belief in the “Hercules” has become increasingly firm. GPT-3, released in 2020, continues to follow the development of micro-innovation and rapid expansion. ideas. Except that the Transformer in GPT-3 applies the Sparse structure, there is almost no difference in the structure of GPT-3 and GPT-2. In terms of "vigor", the GPT-3 model has reached 96 layers, and the training parameters have reached 175 billion (more than 10 times that of GPT-2).

The imminent release of GPT4 is comparable to the human brain, and many industry leaders cannot sit still!

GPT-3 has once again proved OpenAI’s vision. Due to GPT-3’s stronger performance and significantly more parameters, it contains more topic texts, which is obviously better than the previous generation GPT-2. As the largest dense neural network currently available, GPT-3 can convert web page descriptions into corresponding codes, imitate human narratives, create custom poems, generate game scripts, and even imitate deceased philosophers-predicting the true meaning of life. And GPT-3 does not require fine-tuning, it only requires a few samples of the output type (a small amount of learning) to deal with difficult grammar problems. It can be said that GPT-3 seems to have satisfied all our imaginations for language experts.

2. Comprehensively passed the Turing test to lower the threshold for learning and commercial use

Speaking of this, I believe everyone will have the same question - GPT-3 is already very Now that it is powerful, what else can we look forward to in GPT-4?

As we all know, the core way to test the intelligence of an AI system is the Turing test. When we are still unable to use scientific quantifiable standards to define the concept of human intelligence, The Turing test is currently one of the few feasible testing methods that can determine whether the other party has human intelligence. To use a proverb: If something looks like a duck, walks like a duck, and quacks like a duck, then it is a duck. Therefore, if the AI ​​system can successfully pass the Turing test, it means that the system has human thinking and may replace humans in some aspects. According to Korean IT media reports, since mid-November, the industry has reported that GPT-4 has fully passed the Turing test. South Korea's Vodier AI company executive Nam Se-dong said in a recent interview with South Korea's "Economic News": "Although the news that GPT-4 passed the Turing test has not been officially confirmed, the news should be quite credible."

Sam Altman, a technical person and current CEO of OpenAI, also seemed to confirm this information on his Twitter. Altman tweeted on Nov. 10, imitating a classic line from the "Star Wars" character Darth Vader: "Don't be proud of this technological panic you've created. The ability to pass the Turing Test is in 'original' There is nothing you can do in the face of force."

The imminent release of GPT4 is comparable to the human brain, and many industry leaders cannot sit still!

Source: Twitter

An executive from an AI startup company analyzed that “if GPT-4 really passed the Turing test perfectly, then its impact would be enough to cause a 'technological panic' in the AI ​​world, so Altman used the character of Darth Vader to announce this information."

If GPT-4 passes the Turing test without any restrictions, it will indeed make history. Although some AI models have previously claimed to pass the Turing test, they have never been unanimously recognized by the AI ​​industry. This is because the standards and rules of the Turing test are not clear, so many models cleverly exploit some "blind spots" in the test. The AI ​​model "Eugene" launched by the University of Reading in the UK in 2014 is a typical example. At that time, the University of Reading claimed to the judges that the model was a 13-year-old Ukrainian boy, so when the algorithm could not give a good answer At that time, the jury believed that this was because the test subject was a foreign child.

The imminent release of GPT4 is comparable to the human brain, and many industry leaders cannot sit still!

Source: Internet

Although the Turing test is not the absolute reference point for AI technology, But as the oldest and most well-known AI technology test to date, the Turing Test still has great symbolic significance. If GPT-4 really officially and definitely passes the Turing test, then it will most likely create the largest milestone in the AI ​​industry so far.

In addition, unlike GPT-3, GPT-4 will most likely be more than just a language model. Ilya Sutskever, chief scientist of OpenAI, once hinted at this in his multi-modal related article - "Text itself can express a lot of information about the world, but it is incomplete after all, because we also live in a visual world. .” Therefore, some experts in the industry believe that GPT-4 will be multi-modal and can accept audio, text, images and even video inputs, and predict that OpenAI’s Whisper audio data set will be used to create GPT-4. text data. This also means that GPT-4 will no longer have any limitations in receiving and processing external information.

The reason why the industry is paying attention to GPT-4 is probably because the actual commercial threshold of GPT-4 will be lower than that of traditional GPT-3. Enterprises that have previously been unable to use related technologies due to huge costs and infrastructure reasons will also be expected to use GPT-4. GPT-4 is currently in the final stage of listing and will be released between December this year and February next year. Alberto Garcia, an analyst at Cambridge AI Research, published a blog and predicted: "GPT-4 will focus more on optimizing data processing, so the learning cost of GPT-4 is expected to be lower than GPT-3. The learning cost per episode of GPT-4 It will probably be reduced from millions of dollars for GPT-3 to about $1 million."

3. Different paths lead to the same goal: simulating the human brain may come faster

If all the above information is true, then we can foresee at this moment that with the release of GPT-4, next year The field of deep learning research will usher in a new wave; a large number of chat service robots that are more advanced, more natural and whose identity is almost impossible to distinguish will likely appear in various industries; on this basis, there will also be more high-level chat robots. High-quality personalized AI services are born from different traditional businesses; we will also most likely realize barrier-free communication with cognitive intelligence for the first time.

Let’s go back to the problem of manufacturing or simulating human brains mentioned at the beginning. According to a study by MIT, although the neural network in GPT-3 does not attempt to directly imitate the human brain, the language processing method presented by GPT-3 has certain similarities with the solutions obtained during the evolution of the human brain. , when the same stimulus as the test human brain was input to the model, the model obtained the same type of activation as the human brain, and in more than 40 language model tests, GPT-3 made almost perfect inferences. The basics of these models The function is indeed similar to that of the language processing center of the human brain. In this regard, Daniel Yamins, assistant professor of psychology and computer science at Stanford University, also said: "The artificial intelligence network does not directly imitate the brain, but it ends up looking like the brain. This in a sense shows that there is a gap between artificial intelligence and nature. There seems to be some convergent evolution taking place.”

The imminent release of GPT4 is comparable to the human brain, and many industry leaders cannot sit still!

Source: Internet

It can be seen that although the GPT series models do not directly use blue The design idea of ​​​​the Brain Project is to simulate the brain structure, but the effect it presents seems to be closer to our expectations than the Blue Brain Project. Therefore, if this research direction is really feasible and GPT-4 can achieve a leap-forward breakthrough in some aspects based on GPT-3, then we will be one step closer to the goal of simulating some functions of the human brain. .

Finally, I would like to end with a quote from OpenAI CEO Sam Altman’s recent post on Twitter, which was also recognized by “Silicon Valley Iron Man” Elon Musk—— "General artificial intelligence will be established faster than most people think, and it will 'change' everything most people imagine over a long period of time."

The imminent release of GPT4 is comparable to the human brain, and many industry leaders cannot sit still!

Source: Twitter

https:/ /dzone.com/articles/what-can-you-do-with-the-openai-gpt-3-language-mod

https://analyticsindiamag.com/gpt-4-is-almost- here-and-it-looks-better-than-anything-else/

https://analyticsindiamag.com/openais-whisper-might-hold-the-key-to-gpt4/

The above is the detailed content of The imminent release of GPT4 is comparable to the human brain, and many industry leaders cannot sit still!. For more information, please follow other related articles on the PHP Chinese website!

Statement of this Website
The content of this article is voluntarily contributed by netizens, and the copyright belongs to the original author. This site does not assume corresponding legal responsibility. If you find any content suspected of plagiarism or infringement, please contact admin@php.cn

Hot AI Tools

Undresser.AI Undress

Undresser.AI Undress

AI-powered app for creating realistic nude photos

AI Clothes Remover

AI Clothes Remover

Online AI tool for removing clothes from photos.

Undress AI Tool

Undress AI Tool

Undress images for free

Clothoff.io

Clothoff.io

AI clothes remover

AI Hentai Generator

AI Hentai Generator

Generate AI Hentai for free.

Hot Article

R.E.P.O. Energy Crystals Explained and What They Do (Yellow Crystal)
1 months ago By 尊渡假赌尊渡假赌尊渡假赌
R.E.P.O. Best Graphic Settings
1 months ago By 尊渡假赌尊渡假赌尊渡假赌
Will R.E.P.O. Have Crossplay?
1 months ago By 尊渡假赌尊渡假赌尊渡假赌

Hot Tools

Notepad++7.3.1

Notepad++7.3.1

Easy-to-use and free code editor

SublimeText3 Chinese version

SublimeText3 Chinese version

Chinese version, very easy to use

Zend Studio 13.0.1

Zend Studio 13.0.1

Powerful PHP integrated development environment

Dreamweaver CS6

Dreamweaver CS6

Visual web development tools

SublimeText3 Mac version

SublimeText3 Mac version

God-level code editing software (SublimeText3)

New feature in PHP version 5.4: How to use callable type hint parameters to accept callable functions or methods New feature in PHP version 5.4: How to use callable type hint parameters to accept callable functions or methods Jul 29, 2023 pm 09:19 PM

New feature of PHP5.4 version: How to use callable type hint parameters to accept callable functions or methods Introduction: PHP5.4 version introduces a very convenient new feature - you can use callable type hint parameters to accept callable functions or methods . This new feature allows functions and methods to directly specify the corresponding callable parameters without additional checks and conversions. In this article, we will introduce the use of callable type hints and provide some code examples,

What do product parameters mean? What do product parameters mean? Jul 05, 2023 am 11:13 AM

Product parameters refer to the meaning of product attributes. For example, clothing parameters include brand, material, model, size, style, fabric, applicable group, color, etc.; food parameters include brand, weight, material, health license number, applicable group, color, etc.; home appliance parameters include brand, size, color , place of origin, applicable voltage, signal, interface and power, etc.

PHP Warning: Solution to in_array() expects parameter PHP Warning: Solution to in_array() expects parameter Jun 22, 2023 pm 11:52 PM

During the development process, we may encounter such an error message: PHPWarning: in_array()expectsparameter. This error message will appear when using the in_array() function. It may be caused by incorrect parameter passing of the function. Let’s take a look at the solution to this error message. First, you need to clarify the role of the in_array() function: check whether a value exists in the array. The prototype of this function is: in_a

i9-12900H parameter evaluation list i9-12900H parameter evaluation list Feb 23, 2024 am 09:25 AM

i9-12900H is a 14-core processor. The architecture and technology used are all new, and the threads are also very high. The overall work is excellent, and some parameters have been improved. It is particularly comprehensive and can bring users Excellent experience. i9-12900H parameter evaluation review: 1. i9-12900H is a 14-core processor, which adopts the q1 architecture and 24576kb process technology, and has been upgraded to 20 threads. 2. The maximum CPU frequency is 1.80! 5.00ghz, which mainly depends on the workload. 3. Compared with the price, it is very suitable. The price-performance ratio is very good, and it is very suitable for some partners who need normal use. i9-12900H parameter evaluation and performance running scores

C++ function parameter type safety check C++ function parameter type safety check Apr 19, 2024 pm 12:00 PM

C++ parameter type safety checking ensures that functions only accept values ​​of expected types through compile-time checks, run-time checks, and static assertions, preventing unexpected behavior and program crashes: Compile-time type checking: The compiler checks type compatibility. Runtime type checking: Use dynamic_cast to check type compatibility, and throw an exception if there is no match. Static assertion: Assert type conditions at compile time.

C++ program to find the value of the inverse hyperbolic sine function taking a given value as argument C++ program to find the value of the inverse hyperbolic sine function taking a given value as argument Sep 17, 2023 am 10:49 AM

Hyperbolic functions are defined using hyperbolas instead of circles and are equivalent to ordinary trigonometric functions. It returns the ratio parameter in the hyperbolic sine function from the supplied angle in radians. But do the opposite, or in other words. If we want to calculate an angle from a hyperbolic sine, we need an inverse hyperbolic trigonometric operation like the hyperbolic inverse sine operation. This course will demonstrate how to use the hyperbolic inverse sine (asinh) function in C++ to calculate angles using the hyperbolic sine value in radians. The hyperbolic arcsine operation follows the following formula -$$\mathrm{sinh^{-1}x\:=\:In(x\:+\:\sqrt{x^2\:+\:1})}, Where\:In\:is\:natural logarithm\:(log_e\:k)

Can't a language model with 10 billion parameters run? A Chinese doctor from MIT proposed SmoothQuant quantification, which reduced memory requirements by half and increased speed by 1.56 times! Can't a language model with 10 billion parameters run? A Chinese doctor from MIT proposed SmoothQuant quantification, which reduced memory requirements by half and increased speed by 1.56 times! Apr 13, 2023 am 09:31 AM

Although large-scale language models (LLM) have strong performance, the number of parameters can easily reach hundreds of billions, and the demand for computing equipment and memory is so large that ordinary companies cannot afford it. Quantization is a common compression operation that sacrifices some model performance in exchange for faster inference speed and less memory requirements by reducing the accuracy of model weights (such as 32 bit to 8 bit). But for LLMs with more than 100 billion parameters, existing compression methods cannot maintain the accuracy of the model, nor can they run efficiently on hardware. Recently, researchers from MIT and NVIDIA jointly proposed a general-purpose post-training quantization (GPQ).

Advanced usage of reference parameters and pointer parameters in C++ functions Advanced usage of reference parameters and pointer parameters in C++ functions Apr 21, 2024 am 09:39 AM

Reference parameters in C++ functions (essentially variable aliases, modifying the reference modifies the original variable) and pointer parameters (storing the memory address of the original variable, modifying the variable by dereferencing the pointer) have different usages when passing and modifying variables. Reference parameters are often used to modify original variables (especially large structures) to avoid copy overhead when passed to constructors or assignment operators. Pointer parameters are used to flexibly point to memory locations, implement dynamic data structures, or pass null pointers to represent optional parameters.

See all articles