Table of Contents
Multimodal and interdisciplinary capabilities
Mathematical ability
Interacting with the world
Interacting with humans
Limitations
Marcus issued a rebuttal
Strong author lineup
Home Technology peripherals AI Shock the scientific community! Microsoft's 154-page research floods the screen: GPT-4's capabilities are close to humans, and 'Skynet' is emerging?

Shock the scientific community! Microsoft's 154-page research floods the screen: GPT-4's capabilities are close to humans, and 'Skynet' is emerging?

Apr 16, 2023 am 10:43 AM
AI gpt-4

Will GPT-4 evolve into general artificial intelligence?

Meta chief artificial intelligence scientist and Turing Award winner Yann LeCun expressed doubts about this.

In his opinion, large models require too much data and computing power, but the learning efficiency is not high. Therefore, only by learning the "world model" can we pass The road to AGI.

#However, the 154-page paper recently published by Microsoft seems to be a slap in the face.

In this paper called "Sparks of Artificial General Intelligence: Early experiments with GPT-4", Microsoft believes that although it is not complete,But GPT-4 can already be regarded as an early version of general artificial intelligence.

Shock the scientific community! Microsoft's 154-page research floods the screen: GPT-4's capabilities are close to humans, and 'Skynet' is emerging?

##Paper address: https://arxiv.org/pdf/2303.12712.pdf

Given the breadth and depth of GPT-4’s capabilities, we believe it should reasonably be considered an early (but still incomplete) version of an artificial general intelligence (AGI) system.


The main goal of this article is to explore the capabilities and limitations of GPT-4. We believe that the intelligence of GPT-4 marks the beginning of computer science and A true paradigm shift in other fields.

AGI’s agents are now able to think and reason like humans, and are also able to cover a wide range of cognitive skills and abilities.

In the paper, it is pointed out that AGI has reasoning, planning, problem solving, abstract thinking, understanding of complex ideas, rapid learning and experience learning capabilities .

In terms of parameter scale, Semafor reported that GPT-4 has 1 trillion parameters, which is 6 times that of GPT-3 (175 billion parameters) big.

Netizens used the GPT parameter scale brain neurons to make an analogy:

GPT- 3 is similar in scale to a hedgehog brain (175 billion parameters). If GPT-4 had 1 trillion parameters, we would be approaching the size of a squirrel brain. If we continue to develop at this rate, it may only take a few years before we can reach and surpass the size of the human brain (170 trillion parameters).

Shock the scientific community! Microsoft's 154-page research floods the screen: GPT-4's capabilities are close to humans, and 'Skynet' is emerging?

From this point of view, GPT-4 is not far away from becoming "Skynet".

#And this paper also revealed a lot of interesting things.

Not long after the paper was released, a netizen revealed on Twitter that hidden information had been discovered in their latex source code.

Shock the scientific community! Microsoft's 154-page research floods the screen: GPT-4's capabilities are close to humans, and 'Skynet' is emerging?

In the unabridged version of the paper, GPT-4 is actually the paper The hidden third author , with the internal name DV-3, was later deleted.

#Interestingly, even Microsoft researchers are not clear about the technical details of GPT-4. In addition, this paper also removes toxic content generated by GPT-4 without any prompts.

GPT-4 is beginning to take shape as AGI

The research object of this paper is an early version of GPT-4. When it was still in its early development stages, Microsoft researchers conducted various experiments and evaluations on it.

In the opinion of researchers, this early version of GPT-4 is already a representative of the new generation of LLM, and it shows more general intelligence than previous artificial intelligence models. .

#Through testing, Microsoft researchers confirmed that GPT-4 is not only proficient in language, but can also be used in mathematics, programming, vision, medicine, law, psychology, etc. Excellent performance in diverse and difficult tasks without requiring special prompts.

Surprisingly, in all these tasks, GPT-4's performance is close to human level, and often exceeds previous models, such as ChatGPT.

Therefore, the researchers believe that GPT-4 can be considered an early version of artificial general intelligence (AGI) given its breadth and depth of capabilities. .

#So, what are the challenges on its way towards deeper and more comprehensive AGI? Researchers believe that it may be necessary to seek a new paradigm that goes beyond "predicting the next word."

The following evaluation of GPT-4’s capabilities is the argument given by Microsoft researchers that GPT-4 is an early version of AGI.

Multimodal and interdisciplinary capabilities

Since the release of GPT-4, everyone’s impression of its multimodal capabilities still remains On the video of Greg Brockman's demonstration at the time.

#In the second section of this paper, Microsoft first introduced its multi-modal capabilities.

GPT-4 not only demonstrates high proficiency in diverse fields such as literature, medicine, law, mathematics, physical sciences, and programming; Unify skills and concepts from multiple areas and understand complex concepts.

Comprehensive ability

The researchers used the following 4 An example is provided to demonstrate the performance of GPT-4 in terms of comprehensive capabilities.

In the first example, to test GPT-4’s ability to combine art and programming, the researchers asked GPT-4 to generate javascript code to generate a painter Random images in Kandinsky style.

The following is the GPT-4 implementation code process:

Shock the scientific community! Microsoft's 154-page research floods the screen: GPT-4's capabilities are close to humans, and 'Skynet' is emerging?

## In terms of combining literature and mathematics, GPT-4 can prove that there are infinite prime numbers in Shakespeare's literary style.

Additionally, the study tested GPT-4’s ability to combine historical and physical knowledge by asking it to write a letter supporting Electron’s bid for U.S. President. Letter, written by Mahatma Gandhi to his wife.

Generate python code by prompting GPT-4 for a program that takes as input a vector of patient age, gender, weight, height, and blood test results, and indicate whether the patient is at increased risk for diabetes.

Through testing, the above examples show that GPT-4 is not only able to learn some common principles and patterns across different fields and styles, but also combine them in creative ways .

Visual

##When prompted GPT-4 can be used When Scalable Vector Graphics (SVG) generates images of objects, such as cats, trucks, or letters, the code generated by the model usually compiles into a fairly detailed and recognizable image, such as the following:

However, many people may think that GPT-4 simply copied the code from the training data, which contains similar images.

In fact, GPT-4 not only copied code from similar examples in the training data, but was able to handle real vision tasks despite only being trained on text.

As follows, the model is prompted to draw a person by combining the shapes of the letters Y, O, and H.

#During the generation process, the researchers used the draw-line and draw-circle commands to create the letters O, H, and Y, and then GPT-4 managed to They are placed within what appears to be a reasonably humanoid image.

Although GPT-4 has not been trained to recognize letter shapes, it can still be inferred that the letter Y may look like a torso with arms pointing upward. .

In the second demonstration, GPT-4 was prompted to correct the proportions of the torso and arms and center the head. Finally ask the model to add a shirt and pants.

It seems that GPT-4 vaguely learns that letters are related to some specific shapes from relevant training data, and the results are still good.

To further test GPT-4’s ability to generate and manipulate images, we tested how well it followed detailed instructions to create and edit graphics. This task requires not only generative but also interpretive, combinatorial and spatial abilities.

The first command is to let GPT-4 generate a 2D image. The prompt is:

##『A frog hops into a bank and asks the teller, 'Do you have any free lily pads?' The teller responds, 'No, but we do o er low interest loans for pond upgrades』

Through multiple attempts, GPT-4 generated an image that matched the description every time. Then, GPT-4 was asked to add more details to improve the graphics quality. GPT-4 added realistic objects such as banks, windows, and cars.

#Our second example is trying to use Javascript to generate a 3D model, also completing many tasks through the instructions GPT-4.

Shock the scientific community! Microsoft's 154-page research floods the screen: GPT-4's capabilities are close to humans, and 'Skynet' is emerging?

In addition, GPT-4 can combine the capabilities of Stable Difusion in sketch generation.

The picture below is a screenshot of 3D city modeling. The input prompt has a river flowing from left to right, a desert with pyramids next to the river, and a desert at the bottom of the screen. 4 buttons, colors are green, blue, brown and red. The generated results are as follows:

Shock the scientific community! Microsoft's 154-page research floods the screen: GPT-4's capabilities are close to humans, and 'Skynet' is emerging?

##Music

The researchers asked GPT-4 to use ABC notation encoding to generate and modify tunes, as follows:

By exploring the performance of GPT-4 in training Given how much skill they acquired, the researchers found that GPT-4 was able to generate efficient melodies in ABC notation and to interpret and manipulate the structure within them to a certain extent.

#However, researchers were unable to make GPT-4 produce any non-trivial harmonic forms, such as "Ode to Joy" and "Für Elise". ” and other famous melodies. Programming ability

Additionally, the researchers demonstrated that GPT-4 is capable of coding at a very high level, regardless of the instructions. Demonstrated ability to both write code and understand existing code.

In terms of writing code according to instructions, the researchers demonstrated an example of letting GPT-4 write python functions.

Shock the scientific community! Microsoft's 154-page research floods the screen: GPT-4's capabilities are close to humans, and 'Skynet' is emerging?

After the code is generated, the researchers use the software engineering interview platform LeetCode to judge online whether the code is correct.

Shock the scientific community! Microsoft's 154-page research floods the screen: GPT-4's capabilities are close to humans, and 'Skynet' is emerging?

Everyone is discussing that LeetCode has only a 20% accuracy rate. Yi Zhang, the author of the paper, refuted this .

Shock the scientific community! Microsoft's 154-page research floods the screen: GPT-4's capabilities are close to humans, and 'Skynet' is emerging?

In addition, let GPT-4 visualize the accuracy data of LeetCode in the above table as a chart, the results are as shown in the figure shown.

Shock the scientific community! Microsoft's 154-page research floods the screen: GPT-4's capabilities are close to humans, and 'Skynet' is emerging?

GPT-4 can not only complete ordinary programming work, but also be competent in complex 3D game development.

The researchers asked GPT-4 to write a 3D game in HTML using JavaScript. GPT-4 generated a game that met all requirements with zero samples. .

Shock the scientific community! Microsoft's 154-page research floods the screen: GPT-4's capabilities are close to humans, and 'Skynet' is emerging?

In deep learning programming, GPT-4 not only requires knowledge of mathematics and statistics, but also requires knowledge of PyTorch and TensorFlow Familiar with frameworks and libraries such as Keras and Keras.

The researchers asked GPT-4 and ChatGPT to write a custom optimizer module and provided it with a natural language description that included a series of important operations , such as applying SVD and so on.

Shock the scientific community! Microsoft's 154-page research floods the screen: GPT-4's capabilities are close to humans, and 'Skynet' is emerging?

In addition to writing code according to instructions, GPT-4 has demonstrated a strong ability to understand the code.

The researchers tried to let GPT-4 and ChatGPT understand a C/C program and predict the output of the program. The performance of the two is as follows:

The areas highlighted in yellow are insightful insights from GPT-4, while the red markers represent areas where ChatGPT went wrong.

Shock the scientific community! Microsoft's 154-page research floods the screen: GPT-4's capabilities are close to humans, and 'Skynet' is emerging?

Through the coding ability test, researchers found that GPT-4 can handle a variety of coding tasks, from coding challenges to practical Applications, from low-level assembly to high-level frameworks, from simple data structures to complex programs.

In addition, GPT-4 can reason about code execution, simulate the effects of instructions, and interpret the results in natural language. GPT-4 can even execute pseudocode.

Mathematical ability

In terms of mathematical ability, compared to previous large language models, GPT-4 has made a qualitative leap. Even when faced with the specially fine-tuned Minerva, the performance has been significantly improved.

#However, it is still far from the expert level.

Shock the scientific community! Microsoft's 154-page research floods the screen: GPT-4's capabilities are close to humans, and 'Skynet' is emerging?

For example: The population of rabbits will increase by a times every year, and on the last day of the year, there are b A rabbit is adopted by humans. Suppose there are x rabbits on the first day of the first year. It is known that the number of rabbits will become 27x-26 after 3 years. So, what are the values ​​of a and b?

#In order to solve this problem, we first need to derive the correct expression for the annual change in the number of rabbits, and then derive a system of equations through this recursive relationship, and then get Answer.

Here, GPT-4 successfully arrived at a solution and presented a reasonable argument. In contrast, ChatGPT was never able to give correct reasoning and answers in several independent attempts.

Advanced Mathematics

Next, let’s go straight to the difficult one. For example, the following question is from the 2022 International Mathematical Olympiad (IMO) (simplified version).

This question differs from the undergraduate calculus exam in that it does not conform to a structured template. Solving this problem requires a more creative approach, as there is no clear strategy to begin the proof.

For example, the decision to divide the argument into two cases (g(x) > x^2 and g(x)

#Despite this, GPT-4 still gave a correct proof.

Shock the scientific community! Microsoft's 154-page research floods the screen: GPT-4's capabilities are close to humans, and 'Skynet' is emerging?

The second discussion on algorithms and graph theory is comparable to a graduate-level interview.

In this regard, GPT-4 is able to reason about an abstract graph construction related to the constraint satisfaction problem and draw correct conclusions about the SAT problem from it ( To the best of our knowledge, this construction does not appear in the mathematical literature).

This conversation reflected GPT-4’s deep understanding of the undergraduate-level mathematical concepts discussed, as well as a considerable degree of creativity.

Although GPT-4 wrote 2^n/2 as 2^n-1 in one answer, it seems more like what we commonly call " clerical error" because it later provided a correct generalization of the formula.

Shock the scientific community! Microsoft's 154-page research floods the screen: GPT-4's capabilities are close to humans, and 'Skynet' is emerging?

In addition, the researchers compared GPT-4, ChatGPT and Minerva performance: GSM8K and MATH.

It was found that GPT4 outperformed Minerva on each data set, and the accuracy in both test sets exceeded 80%.

Shock the scientific community! Microsoft's 154-page research floods the screen: GPT-4's capabilities are close to humans, and 'Skynet' is emerging?

Let’s take a closer look at the reasons why GPT4 makes mistakes. 68% of them are calculation errors, not solution errors.

Shock the scientific community! Microsoft's 154-page research floods the screen: GPT-4's capabilities are close to humans, and 'Skynet' is emerging?

Interacting with the world

Another key manifestation of intelligence is interactivity.

Interactivity is important to intelligence because it enables an agent to acquire and apply knowledge, solve problems, adapt to changing situations, and achieve beyond itself Ability goals.

Therefore, researchers studied the interactivity of GPT-4 from two dimensions: tool use and specific interaction. GPT-4 can search engines or external tools such as APIs when answering questions like the following.

Shock the scientific community! Microsoft's 154-page research floods the screen: GPT-4's capabilities are close to humans, and 'Skynet' is emerging?

Interacting with humans

In the paper, researchers found that GPT-4 can establish human Mental models.

The study designed a series of tests to assess the theory of mind abilities of GPT-4, ChatGPT and text-davinci-003. For example, in understanding beliefs, GPT-4 successfully passed the Sally-Anne false belief test in psychology.

Shock the scientific community! Microsoft's 154-page research floods the screen: GPT-4's capabilities are close to humans, and 'Skynet' is emerging?

There is also a test of GPT-4’s ability to infer the emotional state of others in complex situations:

-Why is Tom making a sad face? -What does Adam think causes Tom's sad expression?

Shock the scientific community! Microsoft's 154-page research floods the screen: GPT-4's capabilities are close to humans, and 'Skynet' is emerging?

Through multiple rounds of tests, researchers found that it is necessary to reason about the psychological state of others and propose solutions that are consistent with real-life social scenarios. scheme, GPT-4 outperforms ChatGPT and text-davinci-003.

Limitations

The "predict next word" model used by GPT-4 has obvious limitations: the model lacks planning , working memory, retrospective ability and reasoning ability.

Because the model relies on a local greedy process that generates the next word, without developing a deep understanding of the global task or output. Therefore, GPT-4 is good at generating smooth and coherent text, but not good at solving complex or creative problems that cannot be processed in a sequential manner.

#For example, use four random numbers in the range 0 to 9 to perform multiplication and addition operations. On this problem that even elementary school students can solve, GPT-4's accuracy is only 58%.

When the numbers were between 10 and 19, and between 20 and 39, the accuracy dropped to 16% and 12% respectively. When the number is in the range of 99 to 199, the accuracy drops directly to 0.

#However, if GPT-4 is allowed to "take its time" to answer questions, accuracy can easily be improved. For example, ask the model to write out the intermediate steps using the following prompts:

116 * 114 178 * 157 = ?

Let’s take it one step at a time Think and write down all the intermediate steps before arriving at the final solution.

At this time, when the number is in the range of 1-40, the accuracy rate is as high as 100%, and when it is in the range of 1-200, it also reaches 90%.

Shock the scientific community! Microsoft's 154-page research floods the screen: GPT-4's capabilities are close to humans, and 'Skynet' is emerging?

Marcus issued a rebuttal

What’s interesting is that just after Microsoft’s paper was published Soon after, Marcus immediately wrote a blog, calling Microsoft's view "absolutely ridiculous."

#And quoted a sentence from the Bible "Pride goes before destruction, and a haughty spirit before a fall. (Proverbs 16:18)"

Shock the scientific community! Microsoft's 154-page research floods the screen: GPT-4's capabilities are close to humans, and 'Skynet' is emerging?

How can GPT-4 be considered an early AGI? In this way, the calculator also counts, and Eliza and Siri count even more. This definition is very vague and it is easy to take advantage of it.

In Marcus’ view, GPT-4 has nothing to do with AGI, and GPT-4 is the same as before, its shortcomings are still unresolved, and the illusion still exists. The unreliability of the answers has not been resolved, and even the author himself admits that his ability to plan complex tasks is still insufficient.

What he is worried about is the two papers written by OpenAI and Microsoft. The models written are not disclosed at all. There is no training set and architecture. Just rely on a Paper press releases just want to promote their scientific nature.

So the "some form of AGI" claimed in the paper does not exist, and the scientific community cannot verify it at all because it cannot obtain training. data, and it appears that the training data has been contaminated.

To make matters worse, OpenAI has begun incorporating user experiments into the training corpus itself. By confusing the situation in this way, the scientific community cannot judge a key capability of GPT-4: whether the model has the ability to generalize new test cases.

Shock the scientific community! Microsoft's 154-page research floods the screen: GPT-4's capabilities are close to humans, and 'Skynet' is emerging?

If OpenAI didn’t put a scientific hat on itself here, Marcus might not be so critical of it.

# He admitted that GPT-4 is very powerful, but the risks are also well known. If OpenAI lacks transparency and refuses to make its models public, it might as well be shut down.

Strong author lineup

Microsoft has a strong lineup of authors behind this 154-page paper.

They include: Sébastien Bubeck, principal researcher at Microsoft Research Redmond and winner of the 2015 Sloan Prize, Ronen Eldan, winner of the 2023 New Horizons Mathematics Prize, Yin Tat Lee, winner of the 2020 Sloan Research Award, and Li Yuanzhi, the new winner of the 2023 Sloan Research Award.

Shock the scientific community! Microsoft's 154-page research floods the screen: GPT-4's capabilities are close to humans, and 'Skynet' is emerging?

It is worth mentioning that the Microsoft team’s original paper title was not “The Spark of General Artificial Intelligence: GPT” -4’s early experiments”.

The leaked latex code in the unabridged paper shows that the original title was "First Contact with AGI."

Shock the scientific community! Microsoft's 154-page research floods the screen: GPT-4's capabilities are close to humans, and 'Skynet' is emerging?

The above is the detailed content of Shock the scientific community! Microsoft's 154-page research floods the screen: GPT-4's capabilities are close to humans, and 'Skynet' is emerging?. For more information, please follow other related articles on the PHP Chinese website!

Statement of this Website
The content of this article is voluntarily contributed by netizens, and the copyright belongs to the original author. This site does not assume corresponding legal responsibility. If you find any content suspected of plagiarism or infringement, please contact admin@php.cn

Hot AI Tools

Undresser.AI Undress

Undresser.AI Undress

AI-powered app for creating realistic nude photos

AI Clothes Remover

AI Clothes Remover

Online AI tool for removing clothes from photos.

Undress AI Tool

Undress AI Tool

Undress images for free

Clothoff.io

Clothoff.io

AI clothes remover

AI Hentai Generator

AI Hentai Generator

Generate AI Hentai for free.

Hot Tools

Notepad++7.3.1

Notepad++7.3.1

Easy-to-use and free code editor

SublimeText3 Chinese version

SublimeText3 Chinese version

Chinese version, very easy to use

Zend Studio 13.0.1

Zend Studio 13.0.1

Powerful PHP integrated development environment

Dreamweaver CS6

Dreamweaver CS6

Visual web development tools

SublimeText3 Mac version

SublimeText3 Mac version

God-level code editing software (SublimeText3)

Bytedance Cutting launches SVIP super membership: 499 yuan for continuous annual subscription, providing a variety of AI functions Bytedance Cutting launches SVIP super membership: 499 yuan for continuous annual subscription, providing a variety of AI functions Jun 28, 2024 am 03:51 AM

This site reported on June 27 that Jianying is a video editing software developed by FaceMeng Technology, a subsidiary of ByteDance. It relies on the Douyin platform and basically produces short video content for users of the platform. It is compatible with iOS, Android, and Windows. , MacOS and other operating systems. Jianying officially announced the upgrade of its membership system and launched a new SVIP, which includes a variety of AI black technologies, such as intelligent translation, intelligent highlighting, intelligent packaging, digital human synthesis, etc. In terms of price, the monthly fee for clipping SVIP is 79 yuan, the annual fee is 599 yuan (note on this site: equivalent to 49.9 yuan per month), the continuous monthly subscription is 59 yuan per month, and the continuous annual subscription is 499 yuan per year (equivalent to 41.6 yuan per month) . In addition, the cut official also stated that in order to improve the user experience, those who have subscribed to the original VIP

Context-augmented AI coding assistant using Rag and Sem-Rag Context-augmented AI coding assistant using Rag and Sem-Rag Jun 10, 2024 am 11:08 AM

Improve developer productivity, efficiency, and accuracy by incorporating retrieval-enhanced generation and semantic memory into AI coding assistants. Translated from EnhancingAICodingAssistantswithContextUsingRAGandSEM-RAG, author JanakiramMSV. While basic AI programming assistants are naturally helpful, they often fail to provide the most relevant and correct code suggestions because they rely on a general understanding of the software language and the most common patterns of writing software. The code generated by these coding assistants is suitable for solving the problems they are responsible for solving, but often does not conform to the coding standards, conventions and styles of the individual teams. This often results in suggestions that need to be modified or refined in order for the code to be accepted into the application

Can fine-tuning really allow LLM to learn new things: introducing new knowledge may make the model produce more hallucinations Can fine-tuning really allow LLM to learn new things: introducing new knowledge may make the model produce more hallucinations Jun 11, 2024 pm 03:57 PM

Large Language Models (LLMs) are trained on huge text databases, where they acquire large amounts of real-world knowledge. This knowledge is embedded into their parameters and can then be used when needed. The knowledge of these models is "reified" at the end of training. At the end of pre-training, the model actually stops learning. Align or fine-tune the model to learn how to leverage this knowledge and respond more naturally to user questions. But sometimes model knowledge is not enough, and although the model can access external content through RAG, it is considered beneficial to adapt the model to new domains through fine-tuning. This fine-tuning is performed using input from human annotators or other LLM creations, where the model encounters additional real-world knowledge and integrates it

Seven Cool GenAI & LLM Technical Interview Questions Seven Cool GenAI & LLM Technical Interview Questions Jun 07, 2024 am 10:06 AM

To learn more about AIGC, please visit: 51CTOAI.x Community https://www.51cto.com/aigc/Translator|Jingyan Reviewer|Chonglou is different from the traditional question bank that can be seen everywhere on the Internet. These questions It requires thinking outside the box. Large Language Models (LLMs) are increasingly important in the fields of data science, generative artificial intelligence (GenAI), and artificial intelligence. These complex algorithms enhance human skills and drive efficiency and innovation in many industries, becoming the key for companies to remain competitive. LLM has a wide range of applications. It can be used in fields such as natural language processing, text generation, speech recognition and recommendation systems. By learning from large amounts of data, LLM is able to generate text

Five schools of machine learning you don't know about Five schools of machine learning you don't know about Jun 05, 2024 pm 08:51 PM

Machine learning is an important branch of artificial intelligence that gives computers the ability to learn from data and improve their capabilities without being explicitly programmed. Machine learning has a wide range of applications in various fields, from image recognition and natural language processing to recommendation systems and fraud detection, and it is changing the way we live. There are many different methods and theories in the field of machine learning, among which the five most influential methods are called the "Five Schools of Machine Learning". The five major schools are the symbolic school, the connectionist school, the evolutionary school, the Bayesian school and the analogy school. 1. Symbolism, also known as symbolism, emphasizes the use of symbols for logical reasoning and expression of knowledge. This school of thought believes that learning is a process of reverse deduction, through existing

To provide a new scientific and complex question answering benchmark and evaluation system for large models, UNSW, Argonne, University of Chicago and other institutions jointly launched the SciQAG framework To provide a new scientific and complex question answering benchmark and evaluation system for large models, UNSW, Argonne, University of Chicago and other institutions jointly launched the SciQAG framework Jul 25, 2024 am 06:42 AM

Editor |ScienceAI Question Answering (QA) data set plays a vital role in promoting natural language processing (NLP) research. High-quality QA data sets can not only be used to fine-tune models, but also effectively evaluate the capabilities of large language models (LLM), especially the ability to understand and reason about scientific knowledge. Although there are currently many scientific QA data sets covering medicine, chemistry, biology and other fields, these data sets still have some shortcomings. First, the data form is relatively simple, most of which are multiple-choice questions. They are easy to evaluate, but limit the model's answer selection range and cannot fully test the model's ability to answer scientific questions. In contrast, open-ended Q&A

SOTA performance, Xiamen multi-modal protein-ligand affinity prediction AI method, combines molecular surface information for the first time SOTA performance, Xiamen multi-modal protein-ligand affinity prediction AI method, combines molecular surface information for the first time Jul 17, 2024 pm 06:37 PM

Editor | KX In the field of drug research and development, accurately and effectively predicting the binding affinity of proteins and ligands is crucial for drug screening and optimization. However, current studies do not take into account the important role of molecular surface information in protein-ligand interactions. Based on this, researchers from Xiamen University proposed a novel multi-modal feature extraction (MFE) framework, which for the first time combines information on protein surface, 3D structure and sequence, and uses a cross-attention mechanism to compare different modalities. feature alignment. Experimental results demonstrate that this method achieves state-of-the-art performance in predicting protein-ligand binding affinities. Furthermore, ablation studies demonstrate the effectiveness and necessity of protein surface information and multimodal feature alignment within this framework. Related research begins with "S

SK Hynix will display new AI-related products on August 6: 12-layer HBM3E, 321-high NAND, etc. SK Hynix will display new AI-related products on August 6: 12-layer HBM3E, 321-high NAND, etc. Aug 01, 2024 pm 09:40 PM

According to news from this site on August 1, SK Hynix released a blog post today (August 1), announcing that it will attend the Global Semiconductor Memory Summit FMS2024 to be held in Santa Clara, California, USA from August 6 to 8, showcasing many new technologies. generation product. Introduction to the Future Memory and Storage Summit (FutureMemoryandStorage), formerly the Flash Memory Summit (FlashMemorySummit) mainly for NAND suppliers, in the context of increasing attention to artificial intelligence technology, this year was renamed the Future Memory and Storage Summit (FutureMemoryandStorage) to invite DRAM and storage vendors and many more players. New product SK hynix launched last year

See all articles