current location:Home > Technical Articles > Technology peripherals > AI
- Direction:
- All web3.0 Backend Development Web Front-end Database Operation and Maintenance Development Tools PHP Framework Daily Programming WeChat Applet Common Problem Other Tech CMS Tutorial Java System Tutorial Computer Tutorials Hardware Tutorial Mobile Tutorial Software Tutorial Mobile Game Tutorial
- Classify:
-
- Why Gemini 1.5\'s One Million Token Context Is a Game Changer
- GoogleGemini1.5nowcomeswithamassiveonemilliontokencontextwindow,dwarfingitsdirectcompetitioninChatGPT,Claude,andotherAIchatbots.ItsoundslikeamassiveupgradeandcouldsetGeminiapart.It'salittledifficulttograspits
- AI 1028 2024-06-14 15:13:23
-
- GPT-4o Brings GPT-4 to Everyone, and This Is How It Works
- So,whatisGPT-4o?WhatIsGPT-4o?GPT-4oistheChatGPTdeveloperOpenAI'snewestAImodel,revealedatitsearlyMay2024"SpringUpdate"event.Itwillcoexistwithitsprevioustop-performingmodel,GPT-4Turbo,atleastfornow,and
- AI 1222 2024-06-14 15:07:31
-
- How to Use ChatGPT to Write Your Resume
- Standingoutwithyourresumecanbeahugechallenge,butChatGPTisagreattooltogenerateideasandautomatesomeofyourediting.Learnhowtowriteahigh-qualityresumewithChatGPTandhowtofixcommonerrors.HowtoWriteYourResumeWithCha
- AI 746 2024-06-14 15:03:01
-
- Can\'t Wait for the ChatGPT Desktop App? Try This Open Source Alternative Instead
- Aspartofitsmassive2024SpringUpdate,OpenAIrevealedthatalong-awaitedChatGPTdesktopversionwasincoming.StartingwithmacOS,OpenAIisofficiallybringingChatGPTtodesktops,withWindowssupportcominglater.Butifyoucan'twait&md
- AI 1217 2024-06-14 14:31:36
-
- ACL 2024|PsySafe: Research on Agent System Security from an Interdisciplinary Perspective
- The AIxiv column is a column where this site publishes academic and technical content. In the past few years, the AIxiv column of this site has received more than 2,000 reports, covering top laboratories from major universities and companies around the world, effectively promoting academic exchanges and dissemination. If you have excellent work that you want to share, please feel free to contribute or contact us for reporting. Submission email: liyazhou@jiqizhixin.com; zhaoyunfeng@jiqizhixin.com This article was completed by Shanghai Artificial Intelligence Laboratory, Dalian University of Technology and University of Science and Technology of China. Corresponding author: Shao Jing, graduated from the Multimedia Laboratory MMLab of the Chinese University of Hong Kong with a Ph.D., and is currently the head of the large model security team of Pujiang National Laboratory, leading the research on large models.
- AI 483 2024-06-14 14:05:04
-
- 'AI+physics prior knowledge', Zhejiang University and Chinese Academy of Sciences general protein-ligand interaction scoring method published in Nature sub-journal
- Editor | X protein is like a precision lock in the body, and the drug molecule is the key. Only the key that fits perfectly can unlock the door to treatment. Scientists have been looking for efficient ways to predict the fit between these "keys" and "locks," or protein-ligand interactions. However, traditional data-driven methods often fall into "rote learning", memorizing ligand and protein training data instead of truly learning the interactions between them. Recently, a research team from Zhejiang University and the Chinese Academy of Sciences proposed a new scoring method called EquiScore, which uses heterogeneous graph neural networks to integrate physical prior knowledge and characterize protein-ligand interactions in the equation transformation space. EquiScore is trained on a new data set
- AI 1133 2024-06-14 11:40:36
-
- The model behind Apple's intelligence is announced: the 3B model is better than Gemma-7B, and the server model is comparable to GPT-3.5-Turbo
- At the just-concluded Worldwide Developers Conference, Apple announced Appleintelligence, a new personalized intelligent system deeply integrated with iOS18, iPadOS18 and macOSSequoia. Apple+ Intelligence consists of a variety of highly intelligent generative models designed for users' daily tasks. In Apple's just-updated blog, they detailed two of the models. A device-side language model with about 3 billion parameters; a larger server-based language model that runs on Apple servers via private cloud computing. These two base models are part of Apple’s generative model family, and Apple says they will share more about this model family in the near future.
- AI 706 2024-06-13 20:44:13
-
- The dark blue G318 is equipped with great value, a little more comfortable than a hardcore car, and a little more hardcore than an urban SUV.
- It was finally time to wait until the price of Deep Blue G318 was released. On June 13, a "new visitor" came down from the Himalayan iceberg. Deep Blue moved the G318 launch conference to Lhasa. At the same time, Deep Blue G318 configuration is released! There is simply no way for friends and businessmen to survive. Another "super king" has arrived in the hardcore SUV circle. The top ten standard configurations of all series are full of sincerity. The 300,000-class hard-core SUV is tough and comfortable. Judging from the released configuration, the Deep Blue G318 is full of sincerity. If the price is less than 300,000 yuan, it will definitely be the leading SUV in the market. Taking the top ten standard equipment of all series as an example, the dark blue G318 seems to declare that I want them all. Super extended range 2.0 allows one liter of oil to generate 3.63 kilowatt-hours of electricity, and the golden hook battery and 348mm battery pack, which are the highest in its class, ensure ground clearance.
- AI 994 2024-06-13 19:23:04
-
- Open up the entire process of 'self-evolution' of intelligent agents! Fudan launches AgentGym, a general-purpose intelligent body platform
- The AIxiv column is a column where this site publishes academic and technical content. In the past few years, the AIxiv column of this site has received more than 2,000 reports, covering top laboratories from major universities and companies around the world, effectively promoting academic exchanges and dissemination. If you have excellent work that you want to share, please feel free to contribute or contact us for reporting. Submission email: liyazhou@jiqizhixin.com; zhaoyunfeng@jiqizhixin.com The self-evolution ability of AI general intelligence is not out of reach. LLM-basedAgent no longer needs the help of human supervisors and begins to realize "self-evolution"! After learning the expert trajectory, this agent has acquired basic general capabilities.
- AI 501 2024-06-13 18:25:27
-
- 70B model generates 1,000 tokens in seconds, code rewriting surpasses GPT-4o, from the Cursor team, a code artifact invested by OpenAI
- 70B model, 1000 tokens can be generated in seconds, which translates into nearly 4000 characters! The researchers fine-tuned Llama3 and introduced an acceleration algorithm. Compared with the native version, the speed is 13 times faster! Not only is it fast, its performance on code rewriting tasks even surpasses GPT-4o. This achievement comes from anysphere, the team behind the popular AI programming artifact Cursor, and OpenAI also participated in the investment. You must know that on Groq, a well-known fast inference acceleration framework, the inference speed of 70BLlama3 is only more than 300 tokens per second. With the speed of Cursor, it can be said that it achieves near-instant complete code file editing. Some people call it a good guy, if you put Curs
- AI 760 2024-06-13 15:47:02
-
- After one day of training on a single GPU, Transformer can achieve 99% accuracy in adding 100-digit numbers.
- Multiplication and sorting also work. Since it was proposed in 2017, Transformer has become the mainstream architecture for large AI models and has been firmly in the C position. However, what all researchers have to admit is that the Transformer performs extremely poorly on arithmetic tasks, albeit addition, a flaw that largely stems from the Transformer's inability to track the exact position of each digit in a large range of numbers. In order to solve this problem, researchers from the University of Maryland, CMU and other institutions have launched a challenge. They solved this problem by adding an embedding to each number that encodes the position of the number relative to the beginning. The study found that it took just one day to train 20 digits on a single GPU
- AI 866 2024-06-13 14:06:09
-
- No OpenAI data required, join the list of large code models! UIUC releases StarCoder-15B-Instruct
- At the forefront of software technology, UIUC Zhang Lingming's group, together with researchers from the BigCode organization, recently announced the StarCoder2-15B-Instruct large code model. This innovative achievement achieved a significant breakthrough in code generation tasks, successfully surpassing CodeLlama-70B-Instruct and reaching the top of the code generation performance list. The unique feature of StarCoder2-15B-Instruct is its pure self-alignment strategy. The entire training process is open, transparent, and completely autonomous and controllable. The model generates thousands of instructions via StarCoder2-15B in response to fine-tuning the StarCoder-15B base model without relying on expensive manual annotation.
- AI 1463 2024-06-13 13:59:56
-
- Google Gemini 1.5 technical report: Easily prove Mathematical Olympiad questions, the Flash version is 5 times faster than GPT-4 Turbo
- In February this year, Google launched the multi-modal large model Gemini 1.5, which greatly improved performance and speed through engineering and infrastructure optimization, MoE architecture and other strategies. With longer context, stronger reasoning capabilities, and better handling of cross-modal content. This Friday, Google DeepMind officially released the technical report of Gemini 1.5, which covers the Flash version and other recent upgrades. The document is 153 pages long. Technical report link: https://storage.googleapis.com/deepmind-media/gemini/gemini_v1_5_report.pdf In this report, Google introduces Gemini1
- AI 1251 2024-06-13 13:52:27
-
- Advanced practice of industrial knowledge graph
- 1. Background Introduction First, let’s introduce the development history of Yunwen Technology. Yunwen Technology Company...2023 is the period when large models are prevalent. Many companies believe that the importance of graphs has been greatly reduced after large models, and the preset information systems studied previously are no longer important. However, with the promotion of RAG and the prevalence of data governance, we have found that more efficient data governance and high-quality data are important prerequisites for improving the effectiveness of privatized large models. Therefore, more and more companies are beginning to pay attention to knowledge construction related content. This also promotes the construction and processing of knowledge to a higher level, where there are many techniques and methods that can be explored. It can be seen that the emergence of a new technology does not necessarily defeat all old technologies. It is also possible that the new technology and the old technology will be integrated with each other.
- AI 782 2024-06-13 11:59:28
-
- LeCun's new work: layered world model, data-driven humanoid robot control
- With large models as intelligence blessings, humanoid robots have become a new trend. The robots in science fiction movies that "can tell that I'm not a human being" seem to be getting closer. However, thinking and acting like humans is still a difficult engineering problem for robots, especially humanoid robots. Taking a simple learning to walk as an example, using reinforcement learning to train may evolve into the following: There is no problem in principle (following the reward mechanism), and the goal of going up the stairs has been achieved. Except that the process is relatively abstract and different from most human beings. Behavioral patterns may be different. The reason why robots have difficulty acting "naturally" like humans is due to the high-dimensional nature of the space of observation and action, as well as the inherent instability of the bipedal form. In this regard, LeCun participated in
- AI 1090 2024-06-13 11:37:17