Some time ago, Microsoft made an own mistake: it grandly open sourced WizardLM-2, and then withdrew it cleanly soon after.
Currently queryable WizardLM-2 release information, this is an open source large model "truly comparable to GPT-4", with performance in complex chat, multi-language, reasoning and agency has been improved.
The series includes three models: WizardLM-2 8x22B, WizardLM-2 70B and WizardLM-2 7B. Among them:
Additionally, based on human preference evaluation, WizardLM-28x22B’s capabilities “were only slightly behind the GPT-4-1106 preview, but Significantly stronger than CommandRPlus and GPT4-0314."
##It will become another popular version like LLaMa 3. Open source milestone?While everyone was busy downloading the model, the team suddenly withdrew everything: blog, GitHub, HuggingFace all got 404.
Picture source: https://wizardlm.github.io/WizardLM2/
Team’s The explanation is:
Hello to all Huggingface friends! Sorry, we removed the model. It's been a while since we released a model from a few months ago, so we're not familiar with the new release process now: we accidentally left out a necessary item in the model release process - toxicity testing. This is a step that all new models currently need to complete.
We are currently completing this test quickly and will re-release our model as soon as possible. Don't worry, thank you for your concern and understanding.
However, the AI community’s attention and discussion on WizardLM-2 has not stopped. There are several doubts:
First, the deleted open source projects are not just WizardLM-2, all the Wizard series work of the team is gone, including the previous WizardMath and WizardCoder.
Secondly, some people question why the blog is also deleted when the model weights are deleted? If it is just missing the testing part, there is no need to withdraw it completely.
#The team’s explanation is: “According to relevant regulations.” What are the specific regulations? No one knows yet.
Third, there are also speculations that the team behind WizardLM has been fired, and that the withdrawal of the Wizard series project was also forced.
However, this speculation was denied by the team:
Picture source: https://x.com/_Mira___Mira_/status/1783716276944486751
##Picture source: https://x.com/ DavidFSWD/status/1783682898786152470
And when we search for the author’s name now, it has not completely disappeared from Microsoft’s official website:
Image source: https://www.microsoft.com/en-us/research/people/qins/
Fourth, some people speculate that Microsoft has withdrawn this The open source model is, firstly, because its performance is too close to GPT-4, and secondly, because it “collides” with OpenAI’s technical route.
What is the specific route? We can take a look at the technical details of the original blog page.
The team stated that through LLM training, human-generated data in nature is increasingly exhausted, and AI-carefully created data and AI Step-by-Step supervised models will be the gateway to more powerful The only way to go with AI.
Over the past year, the Microsoft team has built a synthetic training system fully powered by artificial intelligence, as shown in the figure below.
is roughly divided into several sections:
Data preprocessing:
Evol Lab:
AI Align AI (AAA):
Learning:
The last thing to say is that any speculation is in vain, let us look forward to the comeback of WizardLM-2.
The above is the detailed content of WizardLM-2, which is 'very close to GPT-4', was urgently withdrawn by Microsoft. What's the inside story?. For more information, please follow other related articles on the PHP Chinese website!