Table of Contents
Xiao Zha is All in again
Big guys’ evaluation of AGI
Home Technology peripherals AI Zuckerberg strongly supports open source AGI: fully training Llama 3, expected to reach 350,000 H100 by the end of the year

Zuckerberg strongly supports open source AGI: fully training Llama 3, expected to reach 350,000 H100 by the end of the year

Jan 19, 2024 pm 05:48 PM
AI openai agi

Xiao Zha announced a new goal: All in open source AGI.

Yes, Xiao Zha is All in again, which is where OpenAI and Google must compete.

But before AGI, the emphasis was on Open Source.

小扎All in 开源AGI:正训练Llama 3,年底将有35万块H100Picture

This move has received a lot of praise, just like when the LIama series of large models were open sourced.

小扎All in 开源AGI:正训练Llama 3,年底将有35万块H100Picture

小扎All in 开源AGI:正训练Llama 3,年底将有35万块H100Picture

But this time there is another wave of All in, and netizens can’t help but let Reminds me of the last wave of All In: Where did the Metaverse go? ? ?

小扎All in 开源AGI:正训练Llama 3,年底将有35万块H100Picture

But it must be said that the Flag listed this time is indeed more specific and even reveals some key data.

For example,

  • There will be 350,000 H100s by the end of the year, and including other GPUs, the total computing power will be equivalent to 600,000 H100s.
  • The FAIR team will work more closely with the GenAI team.
  • LIama 3 is coming soon.

Finally, he also put up a small advertisement. They are building new AI-centric computing devices, such as Ray Ban Meta smart glasses.

It seems that the Metaverse is still going on.

Xiao Zha is All in again

Now, Xiao Zha has officially announced his participation in the AGI battle.

Although there is no clear timetable, as a long-term vision, two key points are clearly stated:

Be open source responsibly and make it widely available so that everyone can benefit from it.

In order to achieve this goal, there are two main things:

First, closely integrate the two existing AI research teams (FAIR and GenAI).

According to LeCun, the two became brother departments.

小扎All in 开源AGI:正训练Llama 3,年底将有35万块H100Picture

and said that Llama-3 is coming!

Second, build large-scale computing infrastructure: by the end of this year, there will be 350,000 H100s, with a total computing power equivalent to 600,000 H100s.

Calculated based on the sales price of US$25,000 to US$30,000, the total computing power value will reach US$15 billion to US$18 billion.

Previously, some organizations predicted that Nvidia’s H100 shipments to Meta could reach 150,000 units in 2023, which is the same as Microsoft and at least three times that of other companies.

For this reason, Xiao Zha said that we have established such capabilities that may be larger than any other individual company.

Some netizens calculated the computing power and said: Brain-sized models are coming soon.

小扎All in 开源AGI:正训练Llama 3,年底将有35万块H100Picture

However, some people question that Nvidia should not be able to produce that many.

小扎All in 开源AGI:正训练Llama 3,年底将有35万块H100Picture

But a Meta leader came out and said: H100 is 350,000 yuan including the current one.

小扎All in 开源AGI:正训练Llama 3,年底将有35万块H100Picture

In addition, progress in hardware equipment is also emphasized.

Big guys’ evaluation of AGI

Before Xiao Zha announced this news, many big guys made a lot of comments about AGI at the World Economic Forum in Davos.

For example, LeCun emphasized the importance of open source in the AGI implementation path on the forum The Expanding Universe of Generative Models.

The reason we see such rapid progress in artificial intelligence is because of open research.

Even though he often expressed doubts that AGI would arrive anytime soon (certainly not within the next five years), AGI arrived.

小扎All in 开源AGI:正训练Llama 3,年底将有35万块H100Picture

And Transformer creator Aidan Gomez said:

Currently we have not completed the expansion, we still need to continue working hard.

As for OpenAI CEO Altman, he said that human-level artificial intelligence will arrive soon, but the change to the world will be far smaller than we imagined.

Artificial General Intelligence (AGI) may be developed in the "fairly near future."

What do you think of the development of AGI?

The above is the detailed content of Zuckerberg strongly supports open source AGI: fully training Llama 3, expected to reach 350,000 H100 by the end of the year. For more information, please follow other related articles on the PHP Chinese website!

Statement of this Website
The content of this article is voluntarily contributed by netizens, and the copyright belongs to the original author. This site does not assume corresponding legal responsibility. If you find any content suspected of plagiarism or infringement, please contact admin@php.cn

Hot AI Tools

Undresser.AI Undress

Undresser.AI Undress

AI-powered app for creating realistic nude photos

AI Clothes Remover

AI Clothes Remover

Online AI tool for removing clothes from photos.

Undress AI Tool

Undress AI Tool

Undress images for free

Clothoff.io

Clothoff.io

AI clothes remover

Video Face Swap

Video Face Swap

Swap faces in any video effortlessly with our completely free AI face swap tool!

Hot Tools

Notepad++7.3.1

Notepad++7.3.1

Easy-to-use and free code editor

SublimeText3 Chinese version

SublimeText3 Chinese version

Chinese version, very easy to use

Zend Studio 13.0.1

Zend Studio 13.0.1

Powerful PHP integrated development environment

Dreamweaver CS6

Dreamweaver CS6

Visual web development tools

SublimeText3 Mac version

SublimeText3 Mac version

God-level code editing software (SublimeText3)

Bytedance Cutting launches SVIP super membership: 499 yuan for continuous annual subscription, providing a variety of AI functions Bytedance Cutting launches SVIP super membership: 499 yuan for continuous annual subscription, providing a variety of AI functions Jun 28, 2024 am 03:51 AM

This site reported on June 27 that Jianying is a video editing software developed by FaceMeng Technology, a subsidiary of ByteDance. It relies on the Douyin platform and basically produces short video content for users of the platform. It is compatible with iOS, Android, and Windows. , MacOS and other operating systems. Jianying officially announced the upgrade of its membership system and launched a new SVIP, which includes a variety of AI black technologies, such as intelligent translation, intelligent highlighting, intelligent packaging, digital human synthesis, etc. In terms of price, the monthly fee for clipping SVIP is 79 yuan, the annual fee is 599 yuan (note on this site: equivalent to 49.9 yuan per month), the continuous monthly subscription is 59 yuan per month, and the continuous annual subscription is 499 yuan per year (equivalent to 41.6 yuan per month) . In addition, the cut official also stated that in order to improve the user experience, those who have subscribed to the original VIP

Context-augmented AI coding assistant using Rag and Sem-Rag Context-augmented AI coding assistant using Rag and Sem-Rag Jun 10, 2024 am 11:08 AM

Improve developer productivity, efficiency, and accuracy by incorporating retrieval-enhanced generation and semantic memory into AI coding assistants. Translated from EnhancingAICodingAssistantswithContextUsingRAGandSEM-RAG, author JanakiramMSV. While basic AI programming assistants are naturally helpful, they often fail to provide the most relevant and correct code suggestions because they rely on a general understanding of the software language and the most common patterns of writing software. The code generated by these coding assistants is suitable for solving the problems they are responsible for solving, but often does not conform to the coding standards, conventions and styles of the individual teams. This often results in suggestions that need to be modified or refined in order for the code to be accepted into the application

Can fine-tuning really allow LLM to learn new things: introducing new knowledge may make the model produce more hallucinations Can fine-tuning really allow LLM to learn new things: introducing new knowledge may make the model produce more hallucinations Jun 11, 2024 pm 03:57 PM

Large Language Models (LLMs) are trained on huge text databases, where they acquire large amounts of real-world knowledge. This knowledge is embedded into their parameters and can then be used when needed. The knowledge of these models is "reified" at the end of training. At the end of pre-training, the model actually stops learning. Align or fine-tune the model to learn how to leverage this knowledge and respond more naturally to user questions. But sometimes model knowledge is not enough, and although the model can access external content through RAG, it is considered beneficial to adapt the model to new domains through fine-tuning. This fine-tuning is performed using input from human annotators or other LLM creations, where the model encounters additional real-world knowledge and integrates it

Posthumous work of the OpenAI Super Alignment Team: Two large models play a game, and the output becomes more understandable Posthumous work of the OpenAI Super Alignment Team: Two large models play a game, and the output becomes more understandable Jul 19, 2024 am 01:29 AM

If the answer given by the AI ​​model is incomprehensible at all, would you dare to use it? As machine learning systems are used in more important areas, it becomes increasingly important to demonstrate why we can trust their output, and when not to trust them. One possible way to gain trust in the output of a complex system is to require the system to produce an interpretation of its output that is readable to a human or another trusted system, that is, fully understandable to the point that any possible errors can be found. For example, to build trust in the judicial system, we require courts to provide clear and readable written opinions that explain and support their decisions. For large language models, we can also adopt a similar approach. However, when taking this approach, ensure that the language model generates

To provide a new scientific and complex question answering benchmark and evaluation system for large models, UNSW, Argonne, University of Chicago and other institutions jointly launched the SciQAG framework To provide a new scientific and complex question answering benchmark and evaluation system for large models, UNSW, Argonne, University of Chicago and other institutions jointly launched the SciQAG framework Jul 25, 2024 am 06:42 AM

Editor |ScienceAI Question Answering (QA) data set plays a vital role in promoting natural language processing (NLP) research. High-quality QA data sets can not only be used to fine-tune models, but also effectively evaluate the capabilities of large language models (LLM), especially the ability to understand and reason about scientific knowledge. Although there are currently many scientific QA data sets covering medicine, chemistry, biology and other fields, these data sets still have some shortcomings. First, the data form is relatively simple, most of which are multiple-choice questions. They are easy to evaluate, but limit the model's answer selection range and cannot fully test the model's ability to answer scientific questions. In contrast, open-ended Q&A

SOTA performance, Xiamen multi-modal protein-ligand affinity prediction AI method, combines molecular surface information for the first time SOTA performance, Xiamen multi-modal protein-ligand affinity prediction AI method, combines molecular surface information for the first time Jul 17, 2024 pm 06:37 PM

Editor | KX In the field of drug research and development, accurately and effectively predicting the binding affinity of proteins and ligands is crucial for drug screening and optimization. However, current studies do not take into account the important role of molecular surface information in protein-ligand interactions. Based on this, researchers from Xiamen University proposed a novel multi-modal feature extraction (MFE) framework, which for the first time combines information on protein surface, 3D structure and sequence, and uses a cross-attention mechanism to compare different modalities. feature alignment. Experimental results demonstrate that this method achieves state-of-the-art performance in predicting protein-ligand binding affinities. Furthermore, ablation studies demonstrate the effectiveness and necessity of protein surface information and multimodal feature alignment within this framework. Related research begins with "S

SK Hynix will display new AI-related products on August 6: 12-layer HBM3E, 321-high NAND, etc. SK Hynix will display new AI-related products on August 6: 12-layer HBM3E, 321-high NAND, etc. Aug 01, 2024 pm 09:40 PM

According to news from this site on August 1, SK Hynix released a blog post today (August 1), announcing that it will attend the Global Semiconductor Memory Summit FMS2024 to be held in Santa Clara, California, USA from August 6 to 8, showcasing many new technologies. generation product. Introduction to the Future Memory and Storage Summit (FutureMemoryandStorage), formerly the Flash Memory Summit (FlashMemorySummit) mainly for NAND suppliers, in the context of increasing attention to artificial intelligence technology, this year was renamed the Future Memory and Storage Summit (FutureMemoryandStorage) to invite DRAM and storage vendors and many more players. New product SK hynix launched last year

Laying out markets such as AI, GlobalFoundries acquires Tagore Technology's gallium nitride technology and related teams Laying out markets such as AI, GlobalFoundries acquires Tagore Technology's gallium nitride technology and related teams Jul 15, 2024 pm 12:21 PM

According to news from this website on July 5, GlobalFoundries issued a press release on July 1 this year, announcing the acquisition of Tagore Technology’s power gallium nitride (GaN) technology and intellectual property portfolio, hoping to expand its market share in automobiles and the Internet of Things. and artificial intelligence data center application areas to explore higher efficiency and better performance. As technologies such as generative AI continue to develop in the digital world, gallium nitride (GaN) has become a key solution for sustainable and efficient power management, especially in data centers. This website quoted the official announcement that during this acquisition, Tagore Technology’s engineering team will join GLOBALFOUNDRIES to further develop gallium nitride technology. G

See all articles