Table of Contents
Simulating human cognition
OlaGPT
Intention Enhance
Memory(Memory)
Learning
Reasoning
Controller
Voting
Experimental results
Home Technology peripherals AI OlaGPT, the first thinking framework that simulates human cognition: six modules enhance the language model and increase reasoning capabilities by up to 85%

OlaGPT, the first thinking framework that simulates human cognition: six modules enhance the language model and increase reasoning capabilities by up to 85%

Jun 05, 2023 pm 04:17 PM
language Model

When ChatGPT was first released, it gave us too much shock. The model's performance in dialogue was so human-like that it created the illusion that the language model has "thinking ability."

However, after gaining an in-depth understanding of language models, researchers have gradually discovered that the reproduction based on high-probability language patterns is still far from the expected "general artificial intelligence".

In most current research, large-scale language models mainly generate thinking chains under the guidance of specific prompts to perform reasoning tasks, without considering the human cognitive framework, which makes the language model unable to solve complex reasoning problems. There is still a significant gap with humans.

When humans face complex reasoning problems, they usually use various cognitive abilities and need to interact with all aspects of tools, knowledge, and external environmental information. Can language models simulate human thinking? What about processes to solve complex problems?

The answer is of course yes! The first model OlaGPT that simulates the human cognitive processing framework is here!

OlaGPT, the first thinking framework that simulates human cognition: six modules enhance the language model and increase reasoning capabilities by up to 85%

Paper link: https://arxiv.org/abs/2305.16334

Code link: https://www.php.cn/link/ 73a1c863a54653d5e184b790fee14754

OlaGPT includes multiple cognitive modules, including attention, memory, reasoning, learning, and corresponding scheduling and decision-making mechanisms; inspired by human active learning, the framework also includes a learning unit to record previous Errors and expert opinions, and dynamic reference to improve the ability to solve similar problems.

OlaGPT, the first thinking framework that simulates human cognition: six modules enhance the language model and increase reasoning capabilities by up to 85%

The article also outlines a common and effective reasoning framework for human problem-solving, and designs a Chain of Thought (CoT) template accordingly; it also proposes a comprehensive decision-making mechanism that can Maximize model accuracy.

Experimental results obtained after rigorous evaluation on multiple inference datasets show that OlaGPT surpasses previous state-of-the-art benchmarks and proves its effectiveness.

Simulating human cognition

There is still a big gap between the current language model and the expected general artificial intelligence. The main manifestations are:

1. In some cases The content generated is meaningless, or deviates from human value preferences, or even gives some very dangerous suggestions. The current solution is to introduce reinforcement learning with human feedback (RLHF) to sort the model output.

2. The language model’s knowledge is limited to concepts and facts explicitly mentioned in the training data.

When faced with complex problems, language models cannot adapt to changing environments, use existing knowledge or tools, reflect on historical lessons, decompose problems, and use the knowledge summarized by humans in the long-term evolution like humans. Thinking patterns (such as analogies, inductive reasoning, deductive reasoning, etc.) to solve problems.

However, there are still many system problems in allowing language models to simulate the process of human brain processing problems:

1. How to systematically imitate and encode the main modules in the human cognitive framework while making it possible to Implemented in a way that schedules according to common human reasoning patterns?

2. How to guide language models to actively learn like humans, that is, learn and develop from historical mistakes or expert solutions to difficult problems?

While it may be feasible to retrain the model to encode corrected answers, it is obviously costly and inflexible.

3. How to make language models flexibly utilize various thinking modes evolved by humans to improve their reasoning performance?

A fixed, universal thinking model is difficult to adapt to different problems. Just like when humans face different types of problems, they usually flexibly choose different ways of thinking, such as analogical reasoning, deductive reasoning, etc.

OlaGPT

OlaGPT is a problem-solving framework that simulates human thinking and can enhance the capabilities of large language models.

OlaGPT draws on the cognitive architecture theory and models the core capabilities of the cognitive framework as attention, memory, learning, reasoning, and action Action selection.

The researchers fine-tuned the framework according to the needs of specific implementation and proposed a process suitable for language models to solve complex problems, which specifically includes six modules: intention enhancement module (attention), memory module ( memory), active learning module (learning), reasoning module (reasoning), controller module (action selection) and voting module.

OlaGPT, the first thinking framework that simulates human cognition: six modules enhance the language model and increase reasoning capabilities by up to 85%

Intention Enhance

Attention is an important part of human cognition, identifying relevant information and filtering out irrelevant data.

Similarly, the researchers designed a corresponding attention module for the language model, namely intent enhancement, which aims to extract the most relevant information and establish a stronger correlation between the user input and the model's language pattern, which can be It is regarded as an optimized converter from user expression habits to model expression habits.

First obtain the question types of LLMs in advance through specific prompt words, and then reconstruct the way of asking questions.

OlaGPT, the first thinking framework that simulates human cognition: six modules enhance the language model and increase reasoning capabilities by up to 85%

For example, add the sentence "Now give you the XX (question type), question and choices:" at the beginning of the question; in order to facilitate analysis, you also need to add " The answer must end with JSON format: Answer: one of options[A,B,C,D,E].」

OlaGPT, the first thinking framework that simulates human cognition: six modules enhance the language model and increase reasoning capabilities by up to 85%

Memory(Memory)

The memory module plays a vital role in storing various knowledge base information. Studies have proven the limitations of current language models in understanding the latest factual data, and the memory module focuses on consolidating knowledge that has not been internalized by the model. and stores it in an external library as long-term memory.

The researchers used the memory function of langchain for short-term memory, and then used the Faiss-based vector database to achieve long-term memory.

During the query process, its search function can extract relevant knowledge from the library, covering four types of memory libraries: facts, tools, notes and thinking, where facts are real-world information, Such as common sense, etc.; tools include search engines, calculators and Wikipedia, which can assist language models in completing some work that does not require editing; notes mainly record some difficult cases and steps to solve problems; the thinking library mainly stores human problem-solving written by experts Thinking template, the expert can be a human or a model.

Learning

The ability to learn is crucial for humans to continuously improve their self-performance. In essence, all forms of learning rely on experience, and language models can learn from previous Learn from your mistakes to quickly improve your reasoning abilities.

OlaGPT, the first thinking framework that simulates human cognition: six modules enhance the language model and increase reasoning capabilities by up to 85%

First, researchers identify problems that the language model cannot solve; then record the insights and explanations provided by experts in the note library; and finally select relevant notes to promote the language model learning so that similar problems can be dealt with more effectively.

Reasoning

The purpose of the reasoning module is to create multiple agents based on the human reasoning process, thereby stimulating the potential thinking ability of the language model and solving reasoning problems.

OlaGPT, the first thinking framework that simulates human cognition: six modules enhance the language model and increase reasoning capabilities by up to 85%

This module combines multiple thinking templates with reference to specific thinking types such as lateral thinking, sequential thinking, critical thinking and integrative thinking to facilitate reasoning tasks.

Controller

The controller module is mainly used to handle related action selections, including the internal planning tasks of the model (such as selecting certain modules for execution) and the processing of facts and tools. Choose from , notes and thought banks.

OlaGPT, the first thinking framework that simulates human cognition: six modules enhance the language model and increase reasoning capabilities by up to 85%

Relevant libraries are first retrieved and matched. The retrieved content is then integrated into a template agent, requiring the language model to provide responses under a template in an asynchronous manner. , just like humans may have difficulty identifying all relevant information at the beginning of reasoning, it is equally difficult to expect language models to do this from the beginning.

Therefore, dynamic retrieval is implemented based on the user's questions and the intermediate reasoning progress, using the Faiss method to create embedded indexes for the above four libraries, in which the retrieval strategies of each library are slightly different.

Voting

Since different thinking templates may be more suitable for different types of problems, the researchers designed the voting module to improve the integrated calibration ability between multiple thinking templates and make more A voting strategy to generate the best answer to improve performance.

Specific voting methods include:

1. Language model voting: Guide the language model to select the most consistent answer among multiple given options and provide a reason.

2. regex voting: Use regular expression exact matching to extract answers to obtain voting results.

Experimental results

In order to evaluate the effectiveness of the enhanced language model framework in reasoning tasks, the researchers conducted a comprehensive experimental comparison on two types of reasoning data sets.

OlaGPT, the first thinking framework that simulates human cognition: six modules enhance the language model and increase reasoning capabilities by up to 85%

It can be seen from the results:

1. SC (self-consistency) performs better than GPT-3.5-turbo, indicating that integration is adopted to a certain extent methods really help improve the effectiveness of large-scale models.

2. The performance of the method proposed in this article exceeds SC, which proves the effectiveness of the thinking template strategy to a certain extent.

The answers to different thinking templates show considerable differences, and voting under different thinking templates will ultimately produce better results than simply conducting multiple rounds of voting.

3. Different thinking templates have different effects, and step-by-step solutions may be more suitable for reasoning problems.

4. The performance of the active learning module is significantly better than the zero-sample method.

Containing challenging cases as part of the note library and using random, retrieval and combination lists can improve performance, which is a feasible strategy.

5. Different retrieval schemes have different effects on different data sets. In general, the combination strategy has better results.

6. The method in this article is obviously better than other solutions. This is due to the reasonable design of the overall framework, including the effective design of the active learning module; the thinking template realizes the adaptation to different models, and the results under different thinking templates are different; the controller module plays a very good control role and selects content that matches the required content; the integration method of different thinking templates designed by the voting module is effective.

Reference materials:

https://www.php.cn/link/73a1c863a54653d5e184b790fee14754

The above is the detailed content of OlaGPT, the first thinking framework that simulates human cognition: six modules enhance the language model and increase reasoning capabilities by up to 85%. For more information, please follow other related articles on the PHP Chinese website!

Statement of this Website
The content of this article is voluntarily contributed by netizens, and the copyright belongs to the original author. This site does not assume corresponding legal responsibility. If you find any content suspected of plagiarism or infringement, please contact admin@php.cn

Hot AI Tools

Undresser.AI Undress

Undresser.AI Undress

AI-powered app for creating realistic nude photos

AI Clothes Remover

AI Clothes Remover

Online AI tool for removing clothes from photos.

Undress AI Tool

Undress AI Tool

Undress images for free

Clothoff.io

Clothoff.io

AI clothes remover

AI Hentai Generator

AI Hentai Generator

Generate AI Hentai for free.

Hot Article

R.E.P.O. Energy Crystals Explained and What They Do (Yellow Crystal)
3 weeks ago By 尊渡假赌尊渡假赌尊渡假赌
R.E.P.O. Best Graphic Settings
3 weeks ago By 尊渡假赌尊渡假赌尊渡假赌
R.E.P.O. How to Fix Audio if You Can't Hear Anyone
3 weeks ago By 尊渡假赌尊渡假赌尊渡假赌
WWE 2K25: How To Unlock Everything In MyRise
1 months ago By 尊渡假赌尊渡假赌尊渡假赌

Hot Tools

Notepad++7.3.1

Notepad++7.3.1

Easy-to-use and free code editor

SublimeText3 Chinese version

SublimeText3 Chinese version

Chinese version, very easy to use

Zend Studio 13.0.1

Zend Studio 13.0.1

Powerful PHP integrated development environment

Dreamweaver CS6

Dreamweaver CS6

Visual web development tools

SublimeText3 Mac version

SublimeText3 Mac version

God-level code editing software (SublimeText3)

The world's most powerful open source MoE model is here, with Chinese capabilities comparable to GPT-4, and the price is only nearly one percent of GPT-4-Turbo The world's most powerful open source MoE model is here, with Chinese capabilities comparable to GPT-4, and the price is only nearly one percent of GPT-4-Turbo May 07, 2024 pm 04:13 PM

Imagine an artificial intelligence model that not only has the ability to surpass traditional computing, but also achieves more efficient performance at a lower cost. This is not science fiction, DeepSeek-V2[1], the world’s most powerful open source MoE model is here. DeepSeek-V2 is a powerful mixture of experts (MoE) language model with the characteristics of economical training and efficient inference. It consists of 236B parameters, 21B of which are used to activate each marker. Compared with DeepSeek67B, DeepSeek-V2 has stronger performance, while saving 42.5% of training costs, reducing KV cache by 93.3%, and increasing the maximum generation throughput to 5.76 times. DeepSeek is a company exploring general artificial intelligence

AI subverts mathematical research! Fields Medal winner and Chinese-American mathematician led 11 top-ranked papers | Liked by Terence Tao AI subverts mathematical research! Fields Medal winner and Chinese-American mathematician led 11 top-ranked papers | Liked by Terence Tao Apr 09, 2024 am 11:52 AM

AI is indeed changing mathematics. Recently, Tao Zhexuan, who has been paying close attention to this issue, forwarded the latest issue of "Bulletin of the American Mathematical Society" (Bulletin of the American Mathematical Society). Focusing on the topic "Will machines change mathematics?", many mathematicians expressed their opinions. The whole process was full of sparks, hardcore and exciting. The author has a strong lineup, including Fields Medal winner Akshay Venkatesh, Chinese mathematician Zheng Lejun, NYU computer scientist Ernest Davis and many other well-known scholars in the industry. The world of AI has changed dramatically. You know, many of these articles were submitted a year ago.

Hello, electric Atlas! Boston Dynamics robot comes back to life, 180-degree weird moves scare Musk Hello, electric Atlas! Boston Dynamics robot comes back to life, 180-degree weird moves scare Musk Apr 18, 2024 pm 07:58 PM

Boston Dynamics Atlas officially enters the era of electric robots! Yesterday, the hydraulic Atlas just "tearfully" withdrew from the stage of history. Today, Boston Dynamics announced that the electric Atlas is on the job. It seems that in the field of commercial humanoid robots, Boston Dynamics is determined to compete with Tesla. After the new video was released, it had already been viewed by more than one million people in just ten hours. The old people leave and new roles appear. This is a historical necessity. There is no doubt that this year is the explosive year of humanoid robots. Netizens commented: The advancement of robots has made this year's opening ceremony look like a human, and the degree of freedom is far greater than that of humans. But is this really not a horror movie? At the beginning of the video, Atlas is lying calmly on the ground, seemingly on his back. What follows is jaw-dropping

KAN, which replaces MLP, has been extended to convolution by open source projects KAN, which replaces MLP, has been extended to convolution by open source projects Jun 01, 2024 pm 10:03 PM

Earlier this month, researchers from MIT and other institutions proposed a very promising alternative to MLP - KAN. KAN outperforms MLP in terms of accuracy and interpretability. And it can outperform MLP running with a larger number of parameters with a very small number of parameters. For example, the authors stated that they used KAN to reproduce DeepMind's results with a smaller network and a higher degree of automation. Specifically, DeepMind's MLP has about 300,000 parameters, while KAN only has about 200 parameters. KAN has a strong mathematical foundation like MLP. MLP is based on the universal approximation theorem, while KAN is based on the Kolmogorov-Arnold representation theorem. As shown in the figure below, KAN has

Google is ecstatic: JAX performance surpasses Pytorch and TensorFlow! It may become the fastest choice for GPU inference training Google is ecstatic: JAX performance surpasses Pytorch and TensorFlow! It may become the fastest choice for GPU inference training Apr 01, 2024 pm 07:46 PM

The performance of JAX, promoted by Google, has surpassed that of Pytorch and TensorFlow in recent benchmark tests, ranking first in 7 indicators. And the test was not done on the TPU with the best JAX performance. Although among developers, Pytorch is still more popular than Tensorflow. But in the future, perhaps more large models will be trained and run based on the JAX platform. Models Recently, the Keras team benchmarked three backends (TensorFlow, JAX, PyTorch) with the native PyTorch implementation and Keras2 with TensorFlow. First, they select a set of mainstream

Tesla robots work in factories, Musk: The degree of freedom of hands will reach 22 this year! Tesla robots work in factories, Musk: The degree of freedom of hands will reach 22 this year! May 06, 2024 pm 04:13 PM

The latest video of Tesla's robot Optimus is released, and it can already work in the factory. At normal speed, it sorts batteries (Tesla's 4680 batteries) like this: The official also released what it looks like at 20x speed - on a small "workstation", picking and picking and picking: This time it is released One of the highlights of the video is that Optimus completes this work in the factory, completely autonomously, without human intervention throughout the process. And from the perspective of Optimus, it can also pick up and place the crooked battery, focusing on automatic error correction: Regarding Optimus's hand, NVIDIA scientist Jim Fan gave a high evaluation: Optimus's hand is the world's five-fingered robot. One of the most dexterous. Its hands are not only tactile

FisheyeDetNet: the first target detection algorithm based on fisheye camera FisheyeDetNet: the first target detection algorithm based on fisheye camera Apr 26, 2024 am 11:37 AM

Target detection is a relatively mature problem in autonomous driving systems, among which pedestrian detection is one of the earliest algorithms to be deployed. Very comprehensive research has been carried out in most papers. However, distance perception using fisheye cameras for surround view is relatively less studied. Due to large radial distortion, standard bounding box representation is difficult to implement in fisheye cameras. To alleviate the above description, we explore extended bounding box, ellipse, and general polygon designs into polar/angular representations and define an instance segmentation mIOU metric to analyze these representations. The proposed model fisheyeDetNet with polygonal shape outperforms other models and simultaneously achieves 49.5% mAP on the Valeo fisheye camera dataset for autonomous driving

Single card running Llama 70B is faster than dual card, Microsoft forced FP6 into A100 | Open source Single card running Llama 70B is faster than dual card, Microsoft forced FP6 into A100 | Open source Apr 29, 2024 pm 04:55 PM

FP8 and lower floating point quantification precision are no longer the "patent" of H100! Lao Huang wanted everyone to use INT8/INT4, and the Microsoft DeepSpeed ​​team started running FP6 on A100 without official support from NVIDIA. Test results show that the new method TC-FPx's FP6 quantization on A100 is close to or occasionally faster than INT4, and has higher accuracy than the latter. On top of this, there is also end-to-end large model support, which has been open sourced and integrated into deep learning inference frameworks such as DeepSpeed. This result also has an immediate effect on accelerating large models - under this framework, using a single card to run Llama, the throughput is 2.65 times higher than that of dual cards. one

See all articles