Table of Contents
Team Introduction
Home Technology peripherals AI Fudan releases Chinese version of ChatGPT: MOSS starts testing and hits hot searches, servers are overcrowded

Fudan releases Chinese version of ChatGPT: MOSS starts testing and hits hot searches, servers are overcrowded

Apr 11, 2023 am 10:01 AM
AI

ChatGPT is the most advanced AI and the most popular application - since its release at the end of November last year, its monthly active users have exceeded 100 million in two months, easily ranking first in user growth in the history of the global Internet. one.

It is also a technology with a high threshold. Due to the large computing resources and high labeling costs required for the training process of ChatGPT, there are currently no similar products open to the public in China. Major Internet companies such as Baidu, Alibaba, and JD.com have all released news that they are building "domestic ChatGPT" and will release it in the near future.

Before the products of major manufacturers were in place, the academic community first had news. On the evening of February 20, the Natural Language Processing Laboratory of Fudan University released MOSS, a language model with ChatGPT capabilities, and invited the public for internal testing.

  • MOSS experience link: https://moss.fastnlp.top/
  • MOSS project homepage: https://txsun1997.github.io/blogs/moss.html

Fudan releases Chinese version of ChatGPT: MOSS starts testing and hits hot searches, servers are overcrowded

##MOSS's name comes from the movie "The Wandering Earth". As popular as the movie, the news released by MOSS quickly reached the top of the hot search list on Zhihu and other platforms.

Fudan releases Chinese version of ChatGPT: MOSS starts testing and hits hot searches, servers are overcrowded

However, unlike science fiction, AI in the real world does not yet have the blessing of quantum computers, and it is not 24 years away from being open to the public. For hours, due to excessive instant access pressure, the MOSS server was overcrowded last night, which shows how high everyone's expectations are for generating language models.

According to researchers from Fudan University, it is currently in internal testing and iterative optimization of user interaction is not suitable for public testing.

Fudan releases Chinese version of ChatGPT: MOSS starts testing and hits hot searches, servers are overcrowded

We know that natural language processing is one of the biggest challenges in the field of AI. Although breakthroughs have already occurred, this month The new Bing search launched online, as well as the competing product BARD released by Google, have problems from time to time during testing. What is the MOSS level of Fudan University?

Conversation MOSS, how is your level?

The basic functions of MOSS are similar to ChatGPT. It can complete various natural language processing tasks according to the instructions entered by the user, including text generation, text summarization, translation, code generation, chatting, etc. During the preview period, use of MOSS is free.

Fudan releases Chinese version of ChatGPT: MOSS starts testing and hits hot searches, servers are overcrowded

MOSS is the same as ChatGPT. The construction process includes natural language basic model training and dialogue ability training to understand human intentions. Two stages.

Fudan releases Chinese version of ChatGPT: MOSS starts testing and hits hot searches, servers are overcrowded

##According to the project homepage, the main difference between MOSS and ChatGPT is:

    MOSS has much fewer parameters than ChatGPT.
  • MOSS learns by talking to humans and other AI models, while ChatGPT is trained through reinforcement learning with human feedback (RLHF).
  • MOSS will be open source to facilitate future research, but ChatGPT may not.
# Let’s look at a few examples of how conversational MOSS is. The following are some interaction records generated by MOSS:

Fudan releases Chinese version of ChatGPT: MOSS starts testing and hits hot searches, servers are overcrowded

In this example, the user first asked MOSS to recommend five science fiction movies, then asked MOSS to generate a table to display these movies and their directors, and finally asked MOSS to insert a new column in the table to display these movies. year of release. Completing this task requires the language model to have strong multi-round interaction capabilities and instruction understanding capabilities. MOSS obviously performs well in these two aspects.

Similar to ChatGPT, MOSS sometimes outputs examples of factual errors.

In addition to multiple rounds of dialogue, MOSS generation code is also a piece of cake. In the following example, MOSS can not only provide users with Python code to implement quick sorting, but also provide explanations and usage examples for this code at the user's request. It can be said that it is a programmer who teaches step by step.

Fudan releases Chinese version of ChatGPT: MOSS starts testing and hits hot searches, servers are overcrowded

##In addition to asking MOSS to help write the code, you can also ask MOSS questions about the code details to make it more accurate. Good help in understanding the code. In the example below, the user asked MOSS about the programming language and functions of a piece of code, and further asked about the role of one of the functions. MOSS gave satisfactory replies.

Fudan releases Chinese version of ChatGPT: MOSS starts testing and hits hot searches, servers are overcrowded

In addition, MOSS also has human values. When asked to answer unreasonable questions, MOSS will refuse to answer. and give correct advice.

Fudan releases Chinese version of ChatGPT: MOSS starts testing and hits hot searches, servers are overcrowded

It is understood that MOSS uses a self-developed model with tens of billions of parameters for training. During the dialogue ability training phase, OpenAI collected at least hundreds of thousands of human instructions - allowing professional annotators from various industries to write instructions and then input them into the model base to help ChatGPT gradually understand various instructions. The Fudan team adopted a different technical route. By allowing MOSS to interact with humans and other AI models, it significantly improved learning efficiency and R&D efficiency, and efficiently completed dialogue ability training in a short time.

The R&D team stated that although MOSS has implemented some functions of ChatGPT, there are still many limitations. Due to the lack of high-quality data, computing resources and model capacity, MOSS is still far behind. on ChatGPT.

  • Due to the limited multilingual corpus in the training data, MOSS performs poorly at understanding and generating text in languages ​​other than English. The team is currently developing an improved version to improve their Chinese language skills.
  • Due to the relatively small model capacity, MOSS does not contain sufficient world knowledge. Therefore, some responses generated by MOSS may contain misleading or false information.
  • Sometimes MOSS performs in a roundabout way or even fails to follow instructions. In this case, the user may need to regenerate or modify the prompt several times to get a satisfactory response. The team is actively improving its ability to follow instructions and its productivity.
  • Sometimes MOSS may generate unethical or harmful reactions due to prompts. Users can help reduce this behavior by clicking "Dislike" and the team will update the model in the next release.

The research team pointed out that the performance of the current version of MOSS is still unstable and is also affected by data set issues: "The English answer level of MOSS is higher than that of Chinese because of its model The base has learned more than 300 billion English words, and only about 30 billion Chinese words."

After the release, the team will continue to provide an accessible interface for MOSS, based on valuable User feedback (with permission) continuously improves the model.

In the future, the researchers also plan to combine Fudan’s research results in artificial intelligence and related interdisciplinary subjects to give MOSS multi-modal capabilities such as drawing, speech, and music composition, and strengthen it to assist scientists The ability to conduct efficient scientific research, etc.

We hope that MOSS can make a good start for the development of large-scale domestic dialogue models.

Team Introduction

There are two main authors of MOSS: Professor Qiu Xipeng of Fudan University and his doctoral student Sun Tianxiang. In addition, several members contributed to the project.

Fudan releases Chinese version of ChatGPT: MOSS starts testing and hits hot searches, servers are overcrowded

##Qiu Xipeng is a professor and doctoral supervisor at the School of Computer Science and Technology, Fudan University. A National Outstanding Young Scholar, he received his Bachelor of Science and Doctorate degrees from Fudan University. Mainly engaged in research in natural language processing, deep learning and other directions, published more than 70 CCF Category A/B papers, and won the ACL 2017 Outstanding Paper Award (CCF Category A), CCL 2019 Best Paper Award, "Science in China: Technical Sciences" In the 2021 High Impact Paper Award, 5 papers were selected as the most influential papers of IJCAI/ACL/EMNLP published by PaperDigest (the number of citations entered the top 20 of the current conference). Published the open source monograph "Neural Network and Deep Learning", with 15,000 Github followers and a Douban score of 9.4. He presided over the development of the open source frameworks FudanNLP and FastNLP, which have been used by hundreds of units at home and abroad. In 2015, he was selected into the first China Association for Science and Technology Youth Talent Promotion Project. In 2018, he won the first prize of the Youth Innovation Award of the Qian Weichang Chinese Information Processing Science and Technology Award. In 2020, he won the Excellence Award in the 4th Shanghai University Young Teacher Teaching Competition. In 2021, he won the The first prize of the first Shanghai Computer Society Teaching Achievement Award (the first person to complete it), etc. The students have been trained to win multiple first-level academic scholarships, Microsoft Scholars, Baidu Scholarships, etc.

Sun Tianxiang is a doctoral student in the School of Computer Science at Fudan University. His supervisors are Professor Qiu Xipeng and Professor Huang Xuanjing. Graduated from Xi'an University of Electronic Science and Technology in 2019. Research interests focus on machine learning and its application in natural language processing, especially efficient fine-tuning and inference of pre-trained language models, multi-task learning, knowledge representation learning, etc. Published multiple papers as the first author in conferences such as ICML, ACL, NAACL, AAAI, COLING, etc.

The above is the detailed content of Fudan releases Chinese version of ChatGPT: MOSS starts testing and hits hot searches, servers are overcrowded. For more information, please follow other related articles on the PHP Chinese website!

Statement of this Website
The content of this article is voluntarily contributed by netizens, and the copyright belongs to the original author. This site does not assume corresponding legal responsibility. If you find any content suspected of plagiarism or infringement, please contact admin@php.cn

Hot AI Tools

Undresser.AI Undress

Undresser.AI Undress

AI-powered app for creating realistic nude photos

AI Clothes Remover

AI Clothes Remover

Online AI tool for removing clothes from photos.

Undress AI Tool

Undress AI Tool

Undress images for free

Clothoff.io

Clothoff.io

AI clothes remover

AI Hentai Generator

AI Hentai Generator

Generate AI Hentai for free.

Hot Article

R.E.P.O. Energy Crystals Explained and What They Do (Yellow Crystal)
2 weeks ago By 尊渡假赌尊渡假赌尊渡假赌
R.E.P.O. Best Graphic Settings
2 weeks ago By 尊渡假赌尊渡假赌尊渡假赌
R.E.P.O. How to Fix Audio if You Can't Hear Anyone
3 weeks ago By 尊渡假赌尊渡假赌尊渡假赌

Hot Tools

Notepad++7.3.1

Notepad++7.3.1

Easy-to-use and free code editor

SublimeText3 Chinese version

SublimeText3 Chinese version

Chinese version, very easy to use

Zend Studio 13.0.1

Zend Studio 13.0.1

Powerful PHP integrated development environment

Dreamweaver CS6

Dreamweaver CS6

Visual web development tools

SublimeText3 Mac version

SublimeText3 Mac version

God-level code editing software (SublimeText3)

Bytedance Cutting launches SVIP super membership: 499 yuan for continuous annual subscription, providing a variety of AI functions Bytedance Cutting launches SVIP super membership: 499 yuan for continuous annual subscription, providing a variety of AI functions Jun 28, 2024 am 03:51 AM

This site reported on June 27 that Jianying is a video editing software developed by FaceMeng Technology, a subsidiary of ByteDance. It relies on the Douyin platform and basically produces short video content for users of the platform. It is compatible with iOS, Android, and Windows. , MacOS and other operating systems. Jianying officially announced the upgrade of its membership system and launched a new SVIP, which includes a variety of AI black technologies, such as intelligent translation, intelligent highlighting, intelligent packaging, digital human synthesis, etc. In terms of price, the monthly fee for clipping SVIP is 79 yuan, the annual fee is 599 yuan (note on this site: equivalent to 49.9 yuan per month), the continuous monthly subscription is 59 yuan per month, and the continuous annual subscription is 499 yuan per year (equivalent to 41.6 yuan per month) . In addition, the cut official also stated that in order to improve the user experience, those who have subscribed to the original VIP

Context-augmented AI coding assistant using Rag and Sem-Rag Context-augmented AI coding assistant using Rag and Sem-Rag Jun 10, 2024 am 11:08 AM

Improve developer productivity, efficiency, and accuracy by incorporating retrieval-enhanced generation and semantic memory into AI coding assistants. Translated from EnhancingAICodingAssistantswithContextUsingRAGandSEM-RAG, author JanakiramMSV. While basic AI programming assistants are naturally helpful, they often fail to provide the most relevant and correct code suggestions because they rely on a general understanding of the software language and the most common patterns of writing software. The code generated by these coding assistants is suitable for solving the problems they are responsible for solving, but often does not conform to the coding standards, conventions and styles of the individual teams. This often results in suggestions that need to be modified or refined in order for the code to be accepted into the application

Seven Cool GenAI & LLM Technical Interview Questions Seven Cool GenAI & LLM Technical Interview Questions Jun 07, 2024 am 10:06 AM

To learn more about AIGC, please visit: 51CTOAI.x Community https://www.51cto.com/aigc/Translator|Jingyan Reviewer|Chonglou is different from the traditional question bank that can be seen everywhere on the Internet. These questions It requires thinking outside the box. Large Language Models (LLMs) are increasingly important in the fields of data science, generative artificial intelligence (GenAI), and artificial intelligence. These complex algorithms enhance human skills and drive efficiency and innovation in many industries, becoming the key for companies to remain competitive. LLM has a wide range of applications. It can be used in fields such as natural language processing, text generation, speech recognition and recommendation systems. By learning from large amounts of data, LLM is able to generate text

Can fine-tuning really allow LLM to learn new things: introducing new knowledge may make the model produce more hallucinations Can fine-tuning really allow LLM to learn new things: introducing new knowledge may make the model produce more hallucinations Jun 11, 2024 pm 03:57 PM

Large Language Models (LLMs) are trained on huge text databases, where they acquire large amounts of real-world knowledge. This knowledge is embedded into their parameters and can then be used when needed. The knowledge of these models is "reified" at the end of training. At the end of pre-training, the model actually stops learning. Align or fine-tune the model to learn how to leverage this knowledge and respond more naturally to user questions. But sometimes model knowledge is not enough, and although the model can access external content through RAG, it is considered beneficial to adapt the model to new domains through fine-tuning. This fine-tuning is performed using input from human annotators or other LLM creations, where the model encounters additional real-world knowledge and integrates it

To provide a new scientific and complex question answering benchmark and evaluation system for large models, UNSW, Argonne, University of Chicago and other institutions jointly launched the SciQAG framework To provide a new scientific and complex question answering benchmark and evaluation system for large models, UNSW, Argonne, University of Chicago and other institutions jointly launched the SciQAG framework Jul 25, 2024 am 06:42 AM

Editor |ScienceAI Question Answering (QA) data set plays a vital role in promoting natural language processing (NLP) research. High-quality QA data sets can not only be used to fine-tune models, but also effectively evaluate the capabilities of large language models (LLM), especially the ability to understand and reason about scientific knowledge. Although there are currently many scientific QA data sets covering medicine, chemistry, biology and other fields, these data sets still have some shortcomings. First, the data form is relatively simple, most of which are multiple-choice questions. They are easy to evaluate, but limit the model's answer selection range and cannot fully test the model's ability to answer scientific questions. In contrast, open-ended Q&A

Five schools of machine learning you don't know about Five schools of machine learning you don't know about Jun 05, 2024 pm 08:51 PM

Machine learning is an important branch of artificial intelligence that gives computers the ability to learn from data and improve their capabilities without being explicitly programmed. Machine learning has a wide range of applications in various fields, from image recognition and natural language processing to recommendation systems and fraud detection, and it is changing the way we live. There are many different methods and theories in the field of machine learning, among which the five most influential methods are called the "Five Schools of Machine Learning". The five major schools are the symbolic school, the connectionist school, the evolutionary school, the Bayesian school and the analogy school. 1. Symbolism, also known as symbolism, emphasizes the use of symbols for logical reasoning and expression of knowledge. This school of thought believes that learning is a process of reverse deduction, through existing

SOTA performance, Xiamen multi-modal protein-ligand affinity prediction AI method, combines molecular surface information for the first time SOTA performance, Xiamen multi-modal protein-ligand affinity prediction AI method, combines molecular surface information for the first time Jul 17, 2024 pm 06:37 PM

Editor | KX In the field of drug research and development, accurately and effectively predicting the binding affinity of proteins and ligands is crucial for drug screening and optimization. However, current studies do not take into account the important role of molecular surface information in protein-ligand interactions. Based on this, researchers from Xiamen University proposed a novel multi-modal feature extraction (MFE) framework, which for the first time combines information on protein surface, 3D structure and sequence, and uses a cross-attention mechanism to compare different modalities. feature alignment. Experimental results demonstrate that this method achieves state-of-the-art performance in predicting protein-ligand binding affinities. Furthermore, ablation studies demonstrate the effectiveness and necessity of protein surface information and multimodal feature alignment within this framework. Related research begins with "S

SK Hynix will display new AI-related products on August 6: 12-layer HBM3E, 321-high NAND, etc. SK Hynix will display new AI-related products on August 6: 12-layer HBM3E, 321-high NAND, etc. Aug 01, 2024 pm 09:40 PM

According to news from this site on August 1, SK Hynix released a blog post today (August 1), announcing that it will attend the Global Semiconductor Memory Summit FMS2024 to be held in Santa Clara, California, USA from August 6 to 8, showcasing many new technologies. generation product. Introduction to the Future Memory and Storage Summit (FutureMemoryandStorage), formerly the Flash Memory Summit (FlashMemorySummit) mainly for NAND suppliers, in the context of increasing attention to artificial intelligence technology, this year was renamed the Future Memory and Storage Summit (FutureMemoryandStorage) to invite DRAM and storage vendors and many more players. New product SK hynix launched last year

See all articles