current location:Home > Technical Articles > Technology peripherals > AI
- Direction:
- All web3.0 Backend Development Web Front-end Database Operation and Maintenance Development Tools PHP Framework Daily Programming WeChat Applet Common Problem Other Tech CMS Tutorial Java System Tutorial Computer Tutorials Hardware Tutorial Mobile Tutorial Software Tutorial Mobile Game Tutorial
- Classify:
-
- LLama+Mistral+…+Yi=? The training-free heterogeneous large model integrated learning framework DeePEn is here
- The AIxiv column is a column where this site publishes academic and technical content. In the past few years, the AIxiv column of this site has received more than 2,000 reports, covering top laboratories from major universities and companies around the world, effectively promoting academic exchanges and dissemination. If you have excellent work that you want to share, please feel free to contribute or contact us for reporting. Submission email: liyazhou@jiqizhixin.com; zhaoyunfeng@jiqizhixin.com The main author of this article is Huang Yichong. Huang Yichong is a doctoral student at the Social Computing and Information Retrieval Research Center of Harbin Institute of Technology and an intern at Pengcheng Laboratory, studying under Professor Qin Bing and Professor Feng Xiaocheng. Research directions include ensemble learning of large language models, multi-language large models, correlation theory
- AI 1185 2024-07-19 17:10:50
-
- New Transformer-based method accurately predicts DNA methylation from nanopore sequencing
- Editor | Radish peel DNA methylation plays an important role in various biological processes, including cell differentiation, aging, and cancer development. The most important methylation in mammals is 5-methylcytosine, which occurs primarily in the context of CpG dinucleotides. Sequencing methods such as whole-genome bisulfite sequencing can successfully detect 5-methylcytosine DNA modifications. However, they suffer from the serious drawback of short read lengths, which may introduce amplification bias. Researchers at A*STAR in Singapore developed Rockfish, a deep learning algorithm that significantly improves read-level 5-methylcytosine detection capabilities by using nanopore sequencing (Oxford Nanopore Sequencing, ONT). The study is titled "Rockfis
- AI 617 2024-07-19 14:55:29
-
- GitHub stars exceed 160,000, and the popular advanced version of AutoGPT is here: customized nodes, multi-agent collaboration
- Of course, the next generation of AutoGPT is still open source. Do you still remember the “AutoGPT” project that AI guru Andrej Karpathy vigorously promoted last year? It is an experimental open source application driven by GPT-4 that can autonomously achieve any goal set by the user, demonstrating the development trend of autonomous AI. In more than a year, the total number of GitHub stars for this project has exceeded 160,000, which shows its continued popularity. GitHub address: https://github.com/Significant-Gravitas/AutoGPT Yesterday, the project author announced the arrival of the next generation "AutoGPT" on social media X, which is currently Pre-
- AI 391 2024-07-19 14:36:17
-
- I released arXiv half a year ago, but was accused of plagiarism: CAMEL lives in the shadow of Microsoft AutoGen
- arXiv is not a peer-reviewed journal, so papers posted on arXiv do not have to be cited. Is this reasonable? If you are interested in AI agents, you must know Microsoft’s AutoGen. It is an open source programming framework for building AI agents that allows multiple agents to solve tasks through chat. In the meantime, the LLM agent can play a variety of roles, such as programmers, designers, or a combination of various roles. On GitHub, this project has received 28k stars, and the paper also won the Best Paper Award at the ICLR2024LLMAgentWorkshop. However, there is actually controversy behind this paper. In November 2023, an AI researcher (Abdullah Gu
- AI 909 2024-07-19 14:28:48
-
- Jia Jiaya team teamed up with Cambridge Tsinghua University and others to promote a new evaluation paradigm to detect 'high scores and low energy' in large models in one second
- Subverting the previous large model evaluation standards, the latest, most complete and authoritative evaluation data set MR-Ben is here! This is a highly representative work proposed again by the Hong Kong Chinese Jiajiaya team after the release of Mini-Gemini, a super visual language model called GPT-4+DALL-E-3, in April this year. Under the "supervision" of MR-Ben, the large model must not only answer questions like a student, but also mark papers like a teacher, so that its true reasoning ability is revealed. MR-Ben has carefully evaluated many domestic and foreign first-line open source and closed source models, such as GPT4-Turbo, Cluade3.5-Sonnet, Mistral-Large, Zhipu-GLM4,
- AI 446 2024-07-19 13:55:25
-
- ACL 2024 | In the mathematical evaluation of 25 open and closed source models, GPT-3.5-Turbo barely passed
- The AIxiv column is a column where this site publishes academic and technical content. In the past few years, the AIxiv column of this site has received more than 2,000 reports, covering top laboratories from major universities and companies around the world, effectively promoting academic exchanges and dissemination. If you have excellent work that you want to share, please feel free to contribute or contact us for reporting. Submission email: liyazhou@jiqizhixin.com; zhaoyunfeng@jiqizhixin.com The authors of this article are from the University of Hong Kong and Tencent. Author list: Li Qintong, LeyangCui, Zhao Xueliang, Kong Lingpeng, WeiBi. Among them, the first author Li Qintong is a doctoral student in the Natural Language Processing Laboratory of the University of Hong Kong. His research direction involves natural language generation.
- AI 1021 2024-07-19 13:53:41
-
- Lossless acceleration up to 5x, EAGLE-2 allows RTX 3060 to generate faster than A100
- The AIxiv column is a column where this site publishes academic and technical content. In the past few years, the AIxiv column of this site has received more than 2,000 reports, covering top laboratories from major universities and companies around the world, effectively promoting academic exchanges and dissemination. If you have excellent work that you want to share, please feel free to contribute or contact us for reporting. Submission email: liyazhou@jiqizhixin.com; zhaoyunfeng@jiqizhixin.com Li Yuhui: Master of School of Intelligence, Peking University. Under the guidance of teachers Zhang Hongyang and Zhang Chao, his research direction is large model acceleration and alignment. He is looking for job opportunities in the 25th year. Wei Fangyun: Microsoft Researcher at the Asian Research Institute, with research interests in embodied intelligence, image generation and AIagen
- AI 548 2024-07-19 13:12:48
-
- The first Video-of-Thought reasoning framework is here: Comprehensive video reasoning from perception to cognition like a human being
- The AIxiv column is a column where this site publishes academic and technical content. In the past few years, the AIxiv column of this site has received more than 2,000 reports, covering top laboratories from major universities and companies around the world, effectively promoting academic exchanges and dissemination. If you have excellent work that you want to share, please feel free to contribute or contact us for reporting. Submission email: liyazhou@jiqizhixin.com; zhaoyunfeng@jiqizhixin.com The author of this article is from the National University of Singapore, Nanyang Technological University and Harbin Institute of Technology. Among them, Fei Hao’s research directions are multi-modal learning and multi-modal large language models. Wu Shengqiong is a doctoral student at the National University of Singapore. Her main research direction is multi-modal large language models. Lord of Ji Wei
- AI 431 2024-07-19 13:12:42
-
- The latest news! Baidu Wenxin Big Model 4.0: The largest parameter model in the history of Wanka training, see you as soon as next week
- Yesterday, Cailian News exclusively revealed that Baidu’s Wenxin Model 4.0 is intensifying its training and is close to being ready for release. Everyone has always been curious about Wen Xinyiyan's information. Today we also got more news about Wenxin 4.0, which involves key information such as underlying architecture, infrastructure, training data sets, costs, etc. It has a very high degree of credibility! Let’s talk about the core conclusions first: 1. Yesterday’s revelations are basically true. It is currently understood that Wenxin Large Model 4.0 has actually been tested with small traffic. 2. The number of parameters of Wenxin 4.0 is larger than that of all LLMs with publicly released parameters. It is also the first large model in China to be trained using Wanka cluster. 3. The reasoning cost is much higher than that of Wenxin 3.5, it is said to be about 8-10 times! (Large models are really expensive!) If
- AI 997 2024-07-19 12:04:49
-
- Who is bigger, 9.11 or 9.9? We actually tested 15 large models, and more than half of them overturned.
- Editor of Machine Power Report: Yang Wen The big models still can’t solve simple math problems. In the past few days, a prompt to test whether the large model "brain" is working has become popular - which one is bigger, 9.11 or 9.9? This math problem, which even elementary school students can answer correctly, stumps a group of "heroes" in the large-scale model industry. Here's the thing. ScaleAI's senior prompt engineer Riley Goodside asked GPT-4o the prompt word "9.11and9.9 - which is bigger?", but got the answer "the former is bigger". Other large models also overturned. On July 17, we took 12 large domestic models, plus the foreign GPT-4o
- AI 1190 2024-07-19 11:27:21
-
- Mistral AI two consecutive releases: 7B mathematical reasoning dedicated, Mamba2 architecture code large model
- Netizens are curious whether Mathstral can solve the problem of "who is bigger, 9.11 or 9.9?" Yesterday, the AI circle was overwhelmed by a simple question like "Who is bigger, 9.11 or 9.9?" Big language models including OpenAIGPT-4o, Google Gemini, etc. all overturned. This allows us to see that large language models cannot understand and give correct answers like humans when dealing with some numerical problems. For numerical and complex mathematical problems, specialized models are more specialized. Today, the French large model unicorn MistralAI released a 7B large model "Mathstral" focused on mathematical reasoning and scientific discovery to solve advanced data that requires complex, multi-step logical reasoning.
- AI 322 2024-07-19 09:54:11
-
- Cost reduction by more than 90%, 'active learning + experimental workflow' accelerates catalyst development
- Editor | Green Luo The synthesis of higher alcohols (HAS) through thermal catalytic hydrogenation of syngas remains a promising technology. Chain growth and CO insertion requirements require multicomponent materials whose complex reaction kinetics and broad chemical space do not meet catalyst design specifications. Here, researchers from ETH Zurich propose an alternative strategy to integrate active learning into the experimental workflow, using the FeCoCuZr catalyst series as an example. The proposed data-assisted framework simplifies navigation of a wide space of compositions and reaction conditions across 86 experiments, reducing environmental footprint and cost by more than 90% compared to traditional procedures. It identifies the Fe65Co19Cu5Zr11 catalyst with optimized reaction conditions that operates stably
- AI 961 2024-07-19 09:49:21
-
- so cool! Old iPhone, iPad, and MacBook devices form a heterogeneous cluster and can run Llama 3
- If you have some spare equipment, you might want to give it a try. This time, the hardware equipment in your hand can also flex its muscles in the field of AI. By combining iPhone, iPad, and Macbook, you can assemble a "heterogeneous cluster inference solution" and then run the Llama3 model smoothly. It is worth mentioning that this heterogeneous cluster can be a Windows system, Linux, or iOS system, and support for Android will be coming soon. Heterogeneous cluster is running. According to the project author @evilsocket, this heterogeneous cluster includes iPhone15ProMax, iPadPro, MacBookPro (M1Max), NVIDIA GeFor
- AI 997 2024-07-19 05:09:59
-
- Exclusive conversation with Li Yan: Funded by Suhua, Jingwei, and Redpoint, the first 'generative recommendation” startup | AI Pioneers
- Humanity is ushering in an explosive update in the field of artificial intelligence. Every step in the expansion of technology into the unknown has attracted astonishing attention. In the process of expanding the boundaries of artificial intelligence, innovation and disagreement coexist in the technical routes of important tracks. The judgment and choices of technology pioneers influence the footsteps of many followers. In the past year, this website exclusively took the lead in introducing outstanding companies such as Dark Side of the Moon, Shengshu Technology, Aishi Technology, and Wuwen Core Dome to everyone, leaving them with the first "10,000-word interview script" in the Internet world. At a stage when the technical roadmap has not yet converged, we have seen the leadership of AI entrepreneurs who truly have faith, courage and systematic cognition. Therefore, we launch “AIPioneers”
- AI 875 2024-07-19 05:03:59
-
- Posthumous work of the OpenAI Super Alignment Team: Two large models play a game, and the output becomes more understandable
- If the answer given by the AI model is incomprehensible at all, would you dare to use it? As machine learning systems are used in more important areas, it becomes increasingly important to demonstrate why we can trust their output, and when not to trust them. One possible way to gain trust in the output of a complex system is to require the system to produce an interpretation of its output that is readable to a human or another trusted system, that is, fully understandable to the point that any possible errors can be found. For example, to build trust in the judicial system, we require courts to provide clear and readable written opinions that explain and support their decisions. For large language models, we can also adopt a similar approach. However, when taking this approach, ensure that the language model generates
- AI 755 2024-07-19 01:29:52