Home Technology peripherals AI The two giants are optimistic at the same time! Artificial intelligence adds a new term, is it the next wave of AI?

The two giants are optimistic at the same time! Artificial intelligence adds a new term, is it the next wave of AI?

May 27, 2023 am 09:52 AM
AI ai (abbreviation for artificial intelligence) wave

On May 18, robot concept stocks set off a rising limit trend. As of the time of publication, Fengli Intelligent and Youde Precision had reached their daily limit of 20cm. Robot, Haozhi Electromechanical, Jiangsu Leili, and Green Harmonic had risen by more than 10%. Southern Seiko and Yuanda had risen by more than 10%. Intelligence, Huagong Technology, etc. have hit the daily limit

The two giants are optimistic at the same time! Artificial intelligence adds a new term, is it the next wave of AI?

News, at the ITF World 2023 Semiconductor Conference, Huang Renxun said that the next wave of artificial intelligence will be embodied intelligence (embodied AI), that is, intelligent systems that can understand, reason, and interact with the physical world.

In addition, at the Tesla shareholders meeting on May 16, Musk stated that humanoid robots will be Tesla’s main long-term source of value in the future. He spent a lot of space explaining robots and speculated that the demand for the humanoid robot Optimus will reach 10 billion units, far exceeding the demand for automobiles, and may be in the tens of billions of dollars.

Being favored by the two giants Tesla and NVIDIA at the same time, the humanoid robot sector is indeed worthy of a high look.

In addition, Beijing has released several measures to promote the development of general artificial intelligence. There is also an expression of embodied intelligence: promoting the research and application of embodied intelligence systems, and breaking through the perception, cognition, and decision-making technologies of robots under complex conditions such as open environments, generalized scenarios, and continuous tasks.

The two giants are optimistic at the same time! Artificial intelligence adds a new term, is it the next wave of AI?

Artificial intelligence adds a new term

"Embodied intelligence" originally dates back to 1950, when Turing first proposed this concept in the paper "Computing Machinery and Intelligence". It refers to a kind of machine intelligence with autonomous decision-making and action capabilities. It can perceive and understand the environment like humans, and complete tasks through autonomous learning and adaptive behavior.

To be more specific, robots/simulators (referring to virtual environments) that can interact with the environment and perceive the environment like humans, and have autonomous planning, decision-making, action, and execution capabilities (referring to virtual environments) are the ultimate form of AI. We temporarily call it embodied intelligence. robot. Its implementation covers a variety of artificial intelligence technologies, such as computer vision, natural language processing, and robotics.

As Li Feifei, a professor of computer science at Stanford University, said, "The meaning of embodiment is not the body itself, but the overall needs and functions of interacting with the environment and doing things in the environment."

The two giants are optimistic at the same time! Artificial intelligence adds a new term, is it the next wave of AI?

Simply put, embodied intelligent robots can understand human language and complete corresponding tasks. Although the ideal is grand, the reality is that it can only "understand human language", and people still need to rely heavily on handwritten code to control robots.

Dieter Fox, senior director of robotics research at NVIDIA and a professor at the University of Washington, pointed out that a key goal of robotics research is to build robots that are helpful to humans in the real world. But to do this, they must first be exposed to and learn how to interact with humans.

The next wave of AI

Giants are investing in the humanoid robot track. After Tesla launched the Optimus prototype last year, and recently including OpenAI Venture Fund, which led the Norwegian robot manufacturer 1X Technologies’ A2 round of financing, chatGPT will also help improve the perception capabilities of humanoid robots and accelerate the industry. change.

The emergence of large models such as GPT has provided new ideas - many researchers have tried to combine multi-modal large language models with robots, through joint training of images, text, and embodied data, and introduced Multi-modal input enhances the model's understanding of real-life objects and helps robots handle embodied reasoning tasks.

The AI ​​teams of Google and Microsoft are at the forefront, trying to inject soul into robots with large models.

On March 8, the team from Google and the Technical University of Berlin launched the largest visual language model in history - PaLM-E, with a parameter volume of up to 562 billion (GPT-3 has a parameter volume of 175 billion). As a multi-modal embodied visual language model (VLM), PaLM-E can not only understand images, but also understand and generate language, and can execute various complex robot instructions without retraining. Google researchers plan to study more real-world applications of PaLM-E, such as home automation or industrial robots. They hope that PaLM-E can promote more research on multi-modal reasoning and embodied AI.

In a recent study, the Microsoft team is exploring how to extend ChatGPT developed by OpenA into the field of robotics, allowing us to use language to intuitively control multiple platforms such as robotic arms, drones, and home assistance robots. The researchers cited multiple examples of ChatGPT solving robotic problems, and also introduced the use of ChatGPT to solve complex robot deployments in areas such as operations, aerial and navigation.

The two giants are optimistic at the same time! Artificial intelligence adds a new term, is it the next wave of AI?

It is obvious that Google and Microsoft have highly similar expectations for embodied AI: humans operate robots and do not need to learn complex programming languages ​​or the details of the robot system. "Speech" (dictating/gesturing requirements to the robot) is " "The robot completes the task" (the robot completes the task), reaching the state of "handy, like an arm using fingers".

Therefore, large language models such as ChatGPT play a crucial role in realizing convenient human-computer interaction in embodied intelligence.

If the large model represented by chatGPT has opened a new era of general AI, then multi-modal, embodied, active and interactive artificial intelligence must be the only way forward in this era.

The above is the detailed content of The two giants are optimistic at the same time! Artificial intelligence adds a new term, is it the next wave of AI?. For more information, please follow other related articles on the PHP Chinese website!

Statement of this Website
The content of this article is voluntarily contributed by netizens, and the copyright belongs to the original author. This site does not assume corresponding legal responsibility. If you find any content suspected of plagiarism or infringement, please contact admin@php.cn

Hot AI Tools

Undresser.AI Undress

Undresser.AI Undress

AI-powered app for creating realistic nude photos

AI Clothes Remover

AI Clothes Remover

Online AI tool for removing clothes from photos.

Undress AI Tool

Undress AI Tool

Undress images for free

Clothoff.io

Clothoff.io

AI clothes remover

Video Face Swap

Video Face Swap

Swap faces in any video effortlessly with our completely free AI face swap tool!

Hot Tools

Notepad++7.3.1

Notepad++7.3.1

Easy-to-use and free code editor

SublimeText3 Chinese version

SublimeText3 Chinese version

Chinese version, very easy to use

Zend Studio 13.0.1

Zend Studio 13.0.1

Powerful PHP integrated development environment

Dreamweaver CS6

Dreamweaver CS6

Visual web development tools

SublimeText3 Mac version

SublimeText3 Mac version

God-level code editing software (SublimeText3)

Bytedance Cutting launches SVIP super membership: 499 yuan for continuous annual subscription, providing a variety of AI functions Bytedance Cutting launches SVIP super membership: 499 yuan for continuous annual subscription, providing a variety of AI functions Jun 28, 2024 am 03:51 AM

This site reported on June 27 that Jianying is a video editing software developed by FaceMeng Technology, a subsidiary of ByteDance. It relies on the Douyin platform and basically produces short video content for users of the platform. It is compatible with iOS, Android, and Windows. , MacOS and other operating systems. Jianying officially announced the upgrade of its membership system and launched a new SVIP, which includes a variety of AI black technologies, such as intelligent translation, intelligent highlighting, intelligent packaging, digital human synthesis, etc. In terms of price, the monthly fee for clipping SVIP is 79 yuan, the annual fee is 599 yuan (note on this site: equivalent to 49.9 yuan per month), the continuous monthly subscription is 59 yuan per month, and the continuous annual subscription is 499 yuan per year (equivalent to 41.6 yuan per month) . In addition, the cut official also stated that in order to improve the user experience, those who have subscribed to the original VIP

Context-augmented AI coding assistant using Rag and Sem-Rag Context-augmented AI coding assistant using Rag and Sem-Rag Jun 10, 2024 am 11:08 AM

Improve developer productivity, efficiency, and accuracy by incorporating retrieval-enhanced generation and semantic memory into AI coding assistants. Translated from EnhancingAICodingAssistantswithContextUsingRAGandSEM-RAG, author JanakiramMSV. While basic AI programming assistants are naturally helpful, they often fail to provide the most relevant and correct code suggestions because they rely on a general understanding of the software language and the most common patterns of writing software. The code generated by these coding assistants is suitable for solving the problems they are responsible for solving, but often does not conform to the coding standards, conventions and styles of the individual teams. This often results in suggestions that need to be modified or refined in order for the code to be accepted into the application

Can fine-tuning really allow LLM to learn new things: introducing new knowledge may make the model produce more hallucinations Can fine-tuning really allow LLM to learn new things: introducing new knowledge may make the model produce more hallucinations Jun 11, 2024 pm 03:57 PM

Large Language Models (LLMs) are trained on huge text databases, where they acquire large amounts of real-world knowledge. This knowledge is embedded into their parameters and can then be used when needed. The knowledge of these models is "reified" at the end of training. At the end of pre-training, the model actually stops learning. Align or fine-tune the model to learn how to leverage this knowledge and respond more naturally to user questions. But sometimes model knowledge is not enough, and although the model can access external content through RAG, it is considered beneficial to adapt the model to new domains through fine-tuning. This fine-tuning is performed using input from human annotators or other LLM creations, where the model encounters additional real-world knowledge and integrates it

Seven Cool GenAI & LLM Technical Interview Questions Seven Cool GenAI & LLM Technical Interview Questions Jun 07, 2024 am 10:06 AM

To learn more about AIGC, please visit: 51CTOAI.x Community https://www.51cto.com/aigc/Translator|Jingyan Reviewer|Chonglou is different from the traditional question bank that can be seen everywhere on the Internet. These questions It requires thinking outside the box. Large Language Models (LLMs) are increasingly important in the fields of data science, generative artificial intelligence (GenAI), and artificial intelligence. These complex algorithms enhance human skills and drive efficiency and innovation in many industries, becoming the key for companies to remain competitive. LLM has a wide range of applications. It can be used in fields such as natural language processing, text generation, speech recognition and recommendation systems. By learning from large amounts of data, LLM is able to generate text

Five schools of machine learning you don't know about Five schools of machine learning you don't know about Jun 05, 2024 pm 08:51 PM

Machine learning is an important branch of artificial intelligence that gives computers the ability to learn from data and improve their capabilities without being explicitly programmed. Machine learning has a wide range of applications in various fields, from image recognition and natural language processing to recommendation systems and fraud detection, and it is changing the way we live. There are many different methods and theories in the field of machine learning, among which the five most influential methods are called the "Five Schools of Machine Learning". The five major schools are the symbolic school, the connectionist school, the evolutionary school, the Bayesian school and the analogy school. 1. Symbolism, also known as symbolism, emphasizes the use of symbols for logical reasoning and expression of knowledge. This school of thought believes that learning is a process of reverse deduction, through existing

To provide a new scientific and complex question answering benchmark and evaluation system for large models, UNSW, Argonne, University of Chicago and other institutions jointly launched the SciQAG framework To provide a new scientific and complex question answering benchmark and evaluation system for large models, UNSW, Argonne, University of Chicago and other institutions jointly launched the SciQAG framework Jul 25, 2024 am 06:42 AM

Editor |ScienceAI Question Answering (QA) data set plays a vital role in promoting natural language processing (NLP) research. High-quality QA data sets can not only be used to fine-tune models, but also effectively evaluate the capabilities of large language models (LLM), especially the ability to understand and reason about scientific knowledge. Although there are currently many scientific QA data sets covering medicine, chemistry, biology and other fields, these data sets still have some shortcomings. First, the data form is relatively simple, most of which are multiple-choice questions. They are easy to evaluate, but limit the model's answer selection range and cannot fully test the model's ability to answer scientific questions. In contrast, open-ended Q&A

SOTA performance, Xiamen multi-modal protein-ligand affinity prediction AI method, combines molecular surface information for the first time SOTA performance, Xiamen multi-modal protein-ligand affinity prediction AI method, combines molecular surface information for the first time Jul 17, 2024 pm 06:37 PM

Editor | KX In the field of drug research and development, accurately and effectively predicting the binding affinity of proteins and ligands is crucial for drug screening and optimization. However, current studies do not take into account the important role of molecular surface information in protein-ligand interactions. Based on this, researchers from Xiamen University proposed a novel multi-modal feature extraction (MFE) framework, which for the first time combines information on protein surface, 3D structure and sequence, and uses a cross-attention mechanism to compare different modalities. feature alignment. Experimental results demonstrate that this method achieves state-of-the-art performance in predicting protein-ligand binding affinities. Furthermore, ablation studies demonstrate the effectiveness and necessity of protein surface information and multimodal feature alignment within this framework. Related research begins with "S

SK Hynix will display new AI-related products on August 6: 12-layer HBM3E, 321-high NAND, etc. SK Hynix will display new AI-related products on August 6: 12-layer HBM3E, 321-high NAND, etc. Aug 01, 2024 pm 09:40 PM

According to news from this site on August 1, SK Hynix released a blog post today (August 1), announcing that it will attend the Global Semiconductor Memory Summit FMS2024 to be held in Santa Clara, California, USA from August 6 to 8, showcasing many new technologies. generation product. Introduction to the Future Memory and Storage Summit (FutureMemoryandStorage), formerly the Flash Memory Summit (FlashMemorySummit) mainly for NAND suppliers, in the context of increasing attention to artificial intelligence technology, this year was renamed the Future Memory and Storage Summit (FutureMemoryandStorage) to invite DRAM and storage vendors and many more players. New product SK hynix launched last year

See all articles