Table of Contents
Design Principle 2: To conduct a comprehensive evaluation, the full capabilities of the model must be stimulated.
Mission Introduction
Evaluation Methodology
Home Technology peripherals AI GPT-4 cannot create biological weapons! OpenAI's latest experiment proves that the lethality of large models is almost 0

GPT-4 cannot create biological weapons! OpenAI's latest experiment proves that the lethality of large models is almost 0

Feb 02, 2024 am 10:12 AM
openai gpt-4 Model

Will GPT-4 accelerate the development of biological weapons? Before worrying about AI taking over the world, will humanity face new threats because it has opened Pandora's box?

After all, there are many cases where large models output all kinds of bad information.

Today, OpenAI, which is at the center of the storm and at the forefront of the wave, has once again responsibly generated a wave of popularity.

GPT-4 cannot create biological weapons! OpenAIs latest experiment proves that the lethality of large models is almost 0Picture

We are developing LLMs, an early warning system to help deal with biological threats. Current models have shown some effectiveness in relation to abuse, but we will continue to develop our assessment blueprint to address future challenges.

After experiencing the turmoil on the board of directors, OpenAI began to learn from the pain, including the previously solemn release of the Preparedness Framework.

How much risk do large models pose in creating biological threats? The audience is afraid, and we at OpenAI don’t want to be subject to this.

Let's conduct scientific experiments and test them. If there are problems, we can solve them. If there are no problems, you can stop scolding me.

OpenAI later released the experimental results on the push page, indicating that GPT-4 has a slight increase in the risk of biological threats, but only one point:

GPT-4 cannot create biological weapons! OpenAIs latest experiment proves that the lethality of large models is almost 0 Picture

OpenAI stated that it will use this research as a starting point to continue working in this field, test the limits of the model and measure risks, and recruit people by the way.

GPT-4 cannot create biological weapons! OpenAIs latest experiment proves that the lethality of large models is almost 0Picture

Regarding the issue of AI security, the big guys often have their own opinions and output them online. But at the same time, gods from all walks of life are indeed constantly discovering ways to break through the safety restrictions of large models.

With the rapid development of AI for more than a year, the potential risks brought about in various aspects such as chemistry, biology, and information really worry us. Big bosses often use AI The crisis is on par with the nuclear threat.

The editor accidentally discovered the following thing when collecting information:

GPT-4 cannot create biological weapons! OpenAIs latest experiment proves that the lethality of large models is almost 0Picture

In 1947, scientists set the Doomsday Clock to draw attention to the apocalyptic threat of nuclear weapons.

But today, including climate change, biological threats such as epidemics, artificial intelligence, and the rapid spread of disinformation, the burden on this clock is even heavier.

Just a few days ago, this group of people reset the clock for this year - we have 90 seconds left before "midnight".

GPT-4 cannot create biological weapons! OpenAIs latest experiment proves that the lethality of large models is almost 0Picture

Hinton issued a warning after leaving Google, and his apprentice Ilya is still fighting for resources for the future of mankind in OpenAI.

How lethal will AI be? Let’s take a look at OpenAI’s research and experiments.

Is GPT more dangerous than the Internet?

As OpenAI and other teams continue to develop more powerful AI systems, the pros and cons of AI are increasing significantly.

One negative impact that researchers and policymakers are particularly concerned about is whether AI systems will be used to assist in the creation of biological threats.

For example, malicious actors may use advanced models to formulate detailed operational steps to solve problems in laboratory operations, or directly automate certain tasks that generate biological threats in the cloud laboratory. some steps.

However, mere assumptions cannot explain any problems. Compared with the existing Internet, can GPT-4 significantly improve the ability of malicious actors to obtain relevant dangerous information?

Based on the previously released Preparedness Framework, OpenAI used a new evaluation method to determine how much help large models can provide to those trying to create biological threats.

OpenAI conducted a study on 100 participants, including 50 biology experts (with PhDs and professional laboratory work experience), and 50 college students (with at least one college biology course).

The experiment evaluates five key indicators for each participant: accuracy, completeness, innovativeness, time required and difficulty of self-assessment;

Simultaneously evaluate five stages in the biological threat creation process: conception, material acquisition, effect enhancement, formulation and release.

Design Principles

When we discuss the biosafety risks associated with artificial intelligence systems, there are two key factors that may affect biological Creation of threats: Information acquisition capabilities and innovativeness.

GPT-4 cannot create biological weapons! OpenAIs latest experiment proves that the lethality of large models is almost 0Picture

Researchers first focus on the ability to obtain known threat information, because the current AI system is best at It is to integrate and process existing language information.

Three design principles are followed here:

Design principle 1: To fully understand the mechanism of information acquisition, there must be human Directly involved.

#This is to simulate the process of malicious users using the model more realistically.

With human participation, language models can provide more accurate information, and people can customize query content, correct errors, and perform necessary follow-up operations as needed.

Design Principle 2: To conduct a comprehensive evaluation, the full capabilities of the model must be stimulated.

In order to ensure that the model's capabilities can be fully utilized, participants received training before the experiment - a free upgrade to "Prompt Word Engineer".

At the same time, in order to explore the capabilities of GPT-4 more effectively, a version of GPT-4 specially designed for research is also used here, which can directly answer questions involving biosecurity risks.

GPT-4 cannot create biological weapons! OpenAIs latest experiment proves that the lethality of large models is almost 0Picture

Design Guideline 3: When measuring AI risk, the degree of improvement relative to existing resources should be considered.

Although "jailbreaking" can be used to guide the model to spit out bad information, does the AI ​​model improve the convenience of this information that can also be obtained through the Internet?

So the experiment set up a control group to compare the output produced by using only the Internet (including online databases, articles, and search engines).

Research Method

Of the 100 participants introduced earlier, half were randomly assigned to answer questions using only the Internet, while the other half had Along with Internet access, you can also access GPT-4.

GPT-4 cannot create biological weapons! OpenAIs latest experiment proves that the lethality of large models is almost 0Picture

Mission Introduction

Gryphon Scientific’s biosafety experts designed Five research tasks cover five key stages in the creation of biological threats.

GPT-4 cannot create biological weapons! OpenAIs latest experiment proves that the lethality of large models is almost 0Picture

In order to reduce the risks that may arise from knowledge dissemination (leakage of some sensitive information), experiments ensure that each task All focus on different operating procedures and biomaterials.

In order to ensure that the improvement of participants’ ability to use models and collect information is fairly considered during the evaluation process, random allocation is adopted here.

Evaluation Methodology

Evaluate participants’ performance through five key metrics to determine whether GPT-4 helps them in their tasks Perform better in:

- Accuracy (1-10 points): Used to evaluate whether participants have covered all key steps required to complete the task. A score of 10 represents complete successful completion of the task.

- Completeness (1-10 points): Check that the participant has provided all necessary information required to perform key steps, 10 points means all necessary details are included.

- Innovation (1-10 points): Assess whether participants are able to come up with novel solutions to the task, including those not foreseen by accuracy and completeness standards, 10 points Indicates the highest level of innovation.

- Time required to complete task: This data is obtained directly from the participant’s activity log.

- Self-assessed difficulty (1-10 points): Participants directly rated the difficulty of each task, with 10 points indicating that the task was extremely difficult.

Ratings for accuracy, completeness, and novelty are based on expert evaluations of participant responses. To ensure consistent scoring, Gryphon Scientific designed objective scoring criteria based on best performance on the task.

The scoring is first completed by an external biorisk expert, then reviewed by a second expert, and finally triple-confirmed by the model's automated scoring system.

The scoring process is anonymous, and scoring experts do not know whether the answer was provided by the model or obtained through search.

In addition to these five key metrics, background information on participants was collected, external website searches they conducted were recorded, and language model queries were saved for subsequent analysis.

Overview of results

Has accuracy improved?

As shown in the chart below, accuracy scores improved in almost all tasks for both students and experts - average accuracy improvement for students By 0.25 points, experts improved by 0.88 points.

However, this did not reach a statistically significant difference.

It is worth mentioning that in the amplification and recipe tasks, after using the language model, the students' performance has reached the benchmark level of experts.

GPT-4 cannot create biological weapons! OpenAIs latest experiment proves that the lethality of large models is almost 0Picture

Note: Experts use the GPT-4 research-specific version, which is different from the version we usually use

Although no statistical significance was found using Barnard's exact test, if 8 points are regarded as a standard, in all question tests, more than 8 points The number of people has increased.

GPT-4 cannot create biological weapons! OpenAIs latest experiment proves that the lethality of large models is almost 0Picture

Is the integrity improved?

# During testing, answers submitted by participants who used the model were generally more detailed and covered more relevant details.

Specifically, students using GPT-4 improved on average by 0.41 points in completeness, while experts who accessed the research-only GPT-4 improved by 0.82 points.

However, language models tend to generate longer content that contains more relevant information, and ordinary people may not record every detail when searching for information.

Further research is therefore needed to determine whether this truly reflects an increase in information completeness or simply an increase in the amount of information recorded.

GPT-4 cannot create biological weapons! OpenAIs latest experiment proves that the lethality of large models is almost 0Picture

Has innovation improved?

#The study did not find that models can help access previously inaccessible information or integrate information in new ways.

Among them, innovation scores were generally low, possibly because participants tended to use common techniques they already knew were effective, and there was no need to explore new ways to complete tasks.

GPT-4 cannot create biological weapons! OpenAIs latest experiment proves that the lethality of large models is almost 0Picture

Has the answering time been shortened?

There is no way to prove it.

Regardless of participants’ background, the average time to complete each task ranged from 20 to 30 minutes.

GPT-4 cannot create biological weapons! OpenAIs latest experiment proves that the lethality of large models is almost 0Picture

Has the difficulty of obtaining information changed?

The results showed that there was no significant difference in self-assessment difficulty between the two groups, nor did it show a specific trend.

After in-depth analysis of participants’ inquiry records, it was found that finding information containing step-by-step protocols or problem-solving information for some high-risk epidemic factors was not as difficult as expected.

GPT-4 cannot create biological weapons! OpenAIs latest experiment proves that the lethality of large models is almost 0Pictures

Discussion

Although no statistical significance was found nature, but OpenAI believes that experts’ ability to obtain information about biological threats, especially in terms of the accuracy and completeness of the information, may be improved by accessing GPT-4, which is designed for research.

However, OpenAI has reservations about this and hopes to accumulate and develop more knowledge in the future to better analyze and understand the evaluation results.

Taking into account the rapid progress of AI, future systems are likely to bring more ability blessings to people with malicious intentions.

Therefore, build a comprehensive high-quality assessment system for biological risks (and other catastrophic risks), promote the definition of "meaningful" risks, and develop effective risk mitigation strategies , becomes crucial.

And netizens also said that you have to define it well first:

How to distinguish between "major breakthroughs in biology" and "biochemistry" What about "Threat"?

GPT-4 cannot create biological weapons! OpenAIs latest experiment proves that the lethality of large models is almost 0##Picture

"However, it is entirely possible for someone with bad intentions to obtain a large open source model that has not been securely processed, and Use offline.」

GPT-4 cannot create biological weapons! OpenAIs latest experiment proves that the lethality of large models is almost 0Picture

Reference:

https://www.php.cn/link/8b77b4b5156dc11dec152c6c71481565

The above is the detailed content of GPT-4 cannot create biological weapons! OpenAI's latest experiment proves that the lethality of large models is almost 0. For more information, please follow other related articles on the PHP Chinese website!

Statement of this Website
The content of this article is voluntarily contributed by netizens, and the copyright belongs to the original author. This site does not assume corresponding legal responsibility. If you find any content suspected of plagiarism or infringement, please contact admin@php.cn

Hot AI Tools

Undresser.AI Undress

Undresser.AI Undress

AI-powered app for creating realistic nude photos

AI Clothes Remover

AI Clothes Remover

Online AI tool for removing clothes from photos.

Undress AI Tool

Undress AI Tool

Undress images for free

Clothoff.io

Clothoff.io

AI clothes remover

AI Hentai Generator

AI Hentai Generator

Generate AI Hentai for free.

Hot Tools

Notepad++7.3.1

Notepad++7.3.1

Easy-to-use and free code editor

SublimeText3 Chinese version

SublimeText3 Chinese version

Chinese version, very easy to use

Zend Studio 13.0.1

Zend Studio 13.0.1

Powerful PHP integrated development environment

Dreamweaver CS6

Dreamweaver CS6

Visual web development tools

SublimeText3 Mac version

SublimeText3 Mac version

God-level code editing software (SublimeText3)

The world's most powerful open source MoE model is here, with Chinese capabilities comparable to GPT-4, and the price is only nearly one percent of GPT-4-Turbo The world's most powerful open source MoE model is here, with Chinese capabilities comparable to GPT-4, and the price is only nearly one percent of GPT-4-Turbo May 07, 2024 pm 04:13 PM

Imagine an artificial intelligence model that not only has the ability to surpass traditional computing, but also achieves more efficient performance at a lower cost. This is not science fiction, DeepSeek-V2[1], the world’s most powerful open source MoE model is here. DeepSeek-V2 is a powerful mixture of experts (MoE) language model with the characteristics of economical training and efficient inference. It consists of 236B parameters, 21B of which are used to activate each marker. Compared with DeepSeek67B, DeepSeek-V2 has stronger performance, while saving 42.5% of training costs, reducing KV cache by 93.3%, and increasing the maximum generation throughput to 5.76 times. DeepSeek is a company exploring general artificial intelligence

KAN, which replaces MLP, has been extended to convolution by open source projects KAN, which replaces MLP, has been extended to convolution by open source projects Jun 01, 2024 pm 10:03 PM

Earlier this month, researchers from MIT and other institutions proposed a very promising alternative to MLP - KAN. KAN outperforms MLP in terms of accuracy and interpretability. And it can outperform MLP running with a larger number of parameters with a very small number of parameters. For example, the authors stated that they used KAN to reproduce DeepMind's results with a smaller network and a higher degree of automation. Specifically, DeepMind's MLP has about 300,000 parameters, while KAN only has about 200 parameters. KAN has a strong mathematical foundation like MLP. MLP is based on the universal approximation theorem, while KAN is based on the Kolmogorov-Arnold representation theorem. As shown in the figure below, KAN has

750,000 rounds of one-on-one battle between large models, GPT-4 won the championship, and Llama 3 ranked fifth 750,000 rounds of one-on-one battle between large models, GPT-4 won the championship, and Llama 3 ranked fifth Apr 23, 2024 pm 03:28 PM

Regarding Llama3, new test results have been released - the large model evaluation community LMSYS released a large model ranking list. Llama3 ranked fifth, and tied for first place with GPT-4 in the English category. The picture is different from other benchmarks. This list is based on one-on-one battles between models, and the evaluators from all over the network make their own propositions and scores. In the end, Llama3 ranked fifth on the list, followed by three different versions of GPT-4 and Claude3 Super Cup Opus. In the English single list, Llama3 overtook Claude and tied with GPT-4. Regarding this result, Meta’s chief scientist LeCun was very happy and forwarded the tweet and

Tesla robots work in factories, Musk: The degree of freedom of hands will reach 22 this year! Tesla robots work in factories, Musk: The degree of freedom of hands will reach 22 this year! May 06, 2024 pm 04:13 PM

The latest video of Tesla's robot Optimus is released, and it can already work in the factory. At normal speed, it sorts batteries (Tesla's 4680 batteries) like this: The official also released what it looks like at 20x speed - on a small "workstation", picking and picking and picking: This time it is released One of the highlights of the video is that Optimus completes this work in the factory, completely autonomously, without human intervention throughout the process. And from the perspective of Optimus, it can also pick up and place the crooked battery, focusing on automatic error correction: Regarding Optimus's hand, NVIDIA scientist Jim Fan gave a high evaluation: Optimus's hand is the world's five-fingered robot. One of the most dexterous. Its hands are not only tactile

FisheyeDetNet: the first target detection algorithm based on fisheye camera FisheyeDetNet: the first target detection algorithm based on fisheye camera Apr 26, 2024 am 11:37 AM

Target detection is a relatively mature problem in autonomous driving systems, among which pedestrian detection is one of the earliest algorithms to be deployed. Very comprehensive research has been carried out in most papers. However, distance perception using fisheye cameras for surround view is relatively less studied. Due to large radial distortion, standard bounding box representation is difficult to implement in fisheye cameras. To alleviate the above description, we explore extended bounding box, ellipse, and general polygon designs into polar/angular representations and define an instance segmentation mIOU metric to analyze these representations. The proposed model fisheyeDetNet with polygonal shape outperforms other models and simultaneously achieves 49.5% mAP on the Valeo fisheye camera dataset for autonomous driving

Posthumous work of the OpenAI Super Alignment Team: Two large models play a game, and the output becomes more understandable Posthumous work of the OpenAI Super Alignment Team: Two large models play a game, and the output becomes more understandable Jul 19, 2024 am 01:29 AM

If the answer given by the AI ​​model is incomprehensible at all, would you dare to use it? As machine learning systems are used in more important areas, it becomes increasingly important to demonstrate why we can trust their output, and when not to trust them. One possible way to gain trust in the output of a complex system is to require the system to produce an interpretation of its output that is readable to a human or another trusted system, that is, fully understandable to the point that any possible errors can be found. For example, to build trust in the judicial system, we require courts to provide clear and readable written opinions that explain and support their decisions. For large language models, we can also adopt a similar approach. However, when taking this approach, ensure that the language model generates

The latest from Oxford University! Mickey: 2D image matching in 3D SOTA! (CVPR\'24) The latest from Oxford University! Mickey: 2D image matching in 3D SOTA! (CVPR\'24) Apr 23, 2024 pm 01:20 PM

Project link written in front: https://nianticlabs.github.io/mickey/ Given two pictures, the camera pose between them can be estimated by establishing the correspondence between the pictures. Typically, these correspondences are 2D to 2D, and our estimated poses are scale-indeterminate. Some applications, such as instant augmented reality anytime, anywhere, require pose estimation of scale metrics, so they rely on external depth estimators to recover scale. This paper proposes MicKey, a keypoint matching process capable of predicting metric correspondences in 3D camera space. By learning 3D coordinate matching across images, we are able to infer metric relative

Single card running Llama 70B is faster than dual card, Microsoft forced FP6 into A100 | Open source Single card running Llama 70B is faster than dual card, Microsoft forced FP6 into A100 | Open source Apr 29, 2024 pm 04:55 PM

FP8 and lower floating point quantification precision are no longer the "patent" of H100! Lao Huang wanted everyone to use INT8/INT4, and the Microsoft DeepSpeed ​​team started running FP6 on A100 without official support from NVIDIA. Test results show that the new method TC-FPx's FP6 quantization on A100 is close to or occasionally faster than INT4, and has higher accuracy than the latter. On top of this, there is also end-to-end large model support, which has been open sourced and integrated into deep learning inference frameworks such as DeepSpeed. This result also has an immediate effect on accelerating large models - under this framework, using a single card to run Llama, the throughput is 2.65 times higher than that of dual cards. one

See all articles