current location:Home > Technical Articles > Technology peripherals > AI
- Direction:
- All web3.0 Backend Development Web Front-end Database Operation and Maintenance Development Tools PHP Framework Daily Programming WeChat Applet Common Problem Other Tech CMS Tutorial Java System Tutorial Computer Tutorials Hardware Tutorial Mobile Tutorial Software Tutorial Mobile Game Tutorial
- Classify:
-
- iFlytek announced that the new version of iFlytek 2.0 is officially launched
- iFlytek iFlytek Smart Text 2.0 launched iFlytek Spark V4.0 large model base PPT text generation large model AIPPT arrangement creation engine PPT online editing module real-time network search long text analysis AIWord AI reading and writing function document creation knowledge management online time: 2023 Number of documents generated in November: over 10 million
- AI 701 2024-08-12 13:01:30
-
- The latest guest agenda of the 2024AGI Business Trend Conference is announced, meet in Shenzhen to witness the shining moment of AI
- In 2024, AI will be everywhere. Outside of the large model arena of giants, a wider range of entrepreneurs, developers and individual players are increasingly paying attention to using AI to solve specific business demands and improve their own workflows. As the "next Internet" that the industry is looking forward to, and even "a platform opportunity beyond the mobile Internet", entrepreneurs are actively exploring the combination of intelligence with specific industries and precise scenarios such as marketing, design, hardware, finance, education, and medical care. From text to images, music and videos; from knowledge questions and answers to shopping guides and customer service, code assistance, emotional companionship, and even medical guidance and triage, legal consultation, scientific research and development, the prediction of using AI to "remake applications" seems to be
- AI 726 2024-08-12 13:01:02
-
- The black box has been opened! Transformer visual explanation tool that can be played, runs GPT-2 locally, and can also perform real-time reasoning
- It's 2024, is there anyone who still doesn't understand how Transformer works? Come and try this interactive tool. In 2017, Google proposed Transformer in the paper "Attentionisallyouneed", which became a major breakthrough in the field of deep learning. The number of citations of this paper has reached nearly 130,000. All subsequent models of the GPT family are also based on the Transformer architecture, which shows its wide influence. As a neural network architecture, Transformer is popular in a variety of tasks from text to vision, especially in the currently hot field of AI chatbots. However, for many non-professionals, the contents of Transformer are
- AI 921 2024-08-11 16:03:00
-
- Add fast and slow eyes to the video model, Apple's new training-free method surpasses everything SOTA in seconds
- Since the release of Sora, the field of AI video generation has become more "busy". In the past few months, we have witnessed Jimeng, RunwayGen-3, LumaAI, and Kuaishou Keling taking turns to explode. Unlike in the past, where one could tell at a glance that the models were generated by AI, this batch of large video models may be the “best” we have ever seen. However, the amazing performance of video large language models (LLM) is inseparable from a large and finely annotated video data set, which requires a very high cost. Recently, a number of innovative methods have emerged in the research field that do not require additional training: using trained image large language models to directly process video tasks, thus bypassing the "expensive" training process. In addition, most existing video LLMs
- AI 463 2024-08-11 16:02:31
-
- Hybrid experts are more assertive and can perceive multiple modalities and act according to the situation. Meta proposes modality-aware expert hybrids
- Mixed experts also have specializations in their profession. For current mixed-modality basic models, a common architectural design is to fuse encoders or decoders of specific modalities. However, this method has limitations: it cannot integrate information from different modalities, and it is difficult to output content containing multiple modalities. . In order to overcome this limitation, MetaFAIR's Chameleon team proposed a new single Transformer architecture in the recent paper "Chameleon: Mixed-modalearly-fusionfoundationmodels", which can transform discrete images and text based on the prediction target of the next token. The mixed modal sequence composed of tokens is modeled, so that in different modes
- AI 388 2024-08-11 13:02:22
-
- Open and closed source model 'Chaos': Let's see which agent can best glimpse human beings' true intentions
- The AIxiv column is a column where this site publishes academic and technical content. In the past few years, the AIxiv column of this site has received more than 2,000 reports, covering top laboratories from major universities and companies around the world, effectively promoting academic exchanges and dissemination. If you have excellent work that you want to share, please feel free to contribute or contact us for reporting. Submission email: liyazhou@jiqizhixin.com; zhaoyunfeng@jiqizhixin.com The first authors of this article are Qian Cheng and He Bingxiang, undergraduate students in the Department of Computer Science at Tsinghua University. Both are members of THUNLP. Qian Cheng's main research interests are tool learning and large model-driven agents. He is about to study for a PhD at UIUC. He Bingxiang’s main research interests are large model alignment and
- AI 1204 2024-08-11 13:02:02
-
- The first artificial intelligence party was successfully held
- In 1905, the first film "Dingjun Mountain" shot by the Chinese themselves was born in Beijing Fengtai Photo Studio, ending the history of no domestic films in China. More than a hundred years later, on August 28, 2023, in the "Original Heart" Hall of the People's Daily New Media Building, a song "New Dingjun Mountain" was performed in "cross-dimensional cooperation" by college students from the Chinese Academy of Opera and digital person Mu Lan , triggering bursts of applause from the audience present. The ancient Peking Opera art, iterative light and shadow technology and new digital technology pay tribute to the classics while breaking through the times and showing a new charm. This is one of the major programs launched by the first Artificial Intelligence (AI) Gala hosted by People's Daily Online. Through a series of cultural festivals full of technology, intelligence and future, the party
- AI 504 2024-08-10 22:23:02
-
- 2024 World Artificial Intelligence Conference TOP20 Young Excellent Papers Released
- In March 2024, the "Notice on Recommending Participating Papers for the "2024 World Artificial Intelligence Conference Youth Outstanding Paper Award"" was released, and a collection activity for young outstanding papers in the field of artificial intelligence was launched for universities, scientific research institutes, and enterprises around the world. As of the deadline for submissions, a total of 159 review papers have been received from domestic and foreign countries, including internationally renowned universities, scientific research institutions, and enterprises. After preliminary evaluation and re-evaluation, the TOP20 outstanding youth papers in 2024 are now released. (In alphabetical order by paper field and name) The 2024 World Artificial Intelligence Conference Youth Outstanding Paper Award will be announced at the World Artificial Intelligence Conference in July, so stay tuned!
- AI 980 2024-08-10 16:06:11
-
- Forum Preview | 'Knowledge Changes the World - Vertical Large Model Redefines Knowledge Management' forum is about to open
- Introduction: Large model technology has great potential in accelerating knowledge acquisition, integration, application and innovation, and can bring transformative impacts to many industries such as education, finance, and government affairs. At present, the application of vertical large models in enterprise scenarios in my country is still in the exploratory stage. Many enterprises are still seeking the best implementation plan for large models. Among them, deep integration with the enterprise's internal knowledge base is regarded as the enterprise-level scenario for the implementation of vertical large models. important direction of application. To this end, during the 2024 World Artificial Intelligence Conference and the High-level Conference on Artificial Intelligence Global Governance, Daguan Data will join hands with Fudan-Daguan Financial Vertical Application Large Model School-Enterprise Joint Research Center and Pudong Youth Federation to jointly host "Knowledge Changes the World-Vertical Big Forum on the topic "Model Redefines Knowledge Management". The forum will be held in July 2024
- AI 958 2024-08-10 11:20:06
-
- Selected into ICDE, the top international database conference, Tencent Cloud database technology innovation has been recognized by the authority
- I learned today that two papers from Tencent Cloud Database were selected into ICDE, the top international database conference. Its distributed transaction protocol - Lion and memory-disk spanning index design framework have solved common problems in the database field, and its technological innovation has been recognized by international authorities. Tencent Cloud Database Team shared technology at the ICDE venue. ICDE (IEEE International Conference on Data Engineering) is an international conference in the field of database research. Together with SIGMOD and VLDB, it is also known as the three top database conferences. Distributed database transaction processing Distributed transaction processing involves multiple rounds of cross-node communication and is slow. past migration
- AI 639 2024-08-10 07:10:06
-
- AI Helps Me With My Schoolwork, But Doesn\'t Do It for Me (ChatGPT, NotebookLM)
- I am a straight-A graduate student with mixed feelings on AI tools. Many people use AI to replace their own efforts in an academic setting. But you can use AI to assist your learning, add efficiency to research, and improve your studies. Start In-De
- AI 739 2024-08-10 06:43:32
-
- Tongyi Qianwen open sourced Qwen2-Math, becoming the most advanced mathematics-specific model
- According to news on August 9, Alibaba Tongyi team open sourced a new generation mathematical model Qwen2-Math, which includes a basic model and an instruction fine-tuning model with three parameters of 1.5B, 7B, and 72B. Qwen2-Math is developed based on the Tongyi Qianwen open source large language model Qwen2. The flagship model Qwen2-Math-72B-Instruct scores higher than GPT-4o, Claude-3.5-Sonnet, Gemini-1.5-Pro, and Llama on the authoritative evaluation set MATH. -3.1-405B, etc., handle a variety of mathematical problems such as algebra, geometry, counting and probability, number theory, etc. with an accuracy of 84%, becoming the most advanced mathematics-specific model. NOTE: In MATH benchmark evaluation
- AI 779 2024-08-09 19:44:14
-
- Google I/O Developer Conference China Station: Large model technology enters the entire ecosystem
- On Wednesday, Google held its China Developers Conference in Beijing. In addition to the I/O Conference in Mountain View, California in May, this year's I/OConnect global tour started from Berlin, Germany, to Bangalore, India, and has also arrived in Beijing, China. At this event, experts from many different fields shared in-depth Google's latest technological progress, development tools, and global service platforms in AI, Web, mobile, cloud technology and other fields for domestic developers. "China is one of the largest developer markets in the world, and our developers have always been an indispensable pioneering force on the global stage," Chen Junting, president of Google Greater China, said in the opening speech. "In the past year, 25 development teams from China have produced a total of 31 games and applications in different regions around the world.
- AI 599 2024-08-09 19:42:02
-
- Complete 1 year MD calculation in 2.5 days? DeepMind team's new calculation method based on Euclidean Transformer
- Editor | Radish Skin In recent years, great progress has been made in the development of machine learning force fields (MLFF) based on ab initio reference calculations. Although low test errors are achieved, the reliability of MLFF in molecular dynamics (MD) simulations is facing increasing scrutiny due to concerns about instability over longer simulation time scales. Research has shown a potential link between robustness to cumulative inaccuracies and the use of equivariant representations in MLFF, but the computational costs associated with these representations may limit this advantage in practice. To solve this problem, researchers from Google DeepMind and Technical University of Berlin (TUBerlin) proposed a transformer framework called SO3krates.
- AI 697 2024-08-09 19:40:30
-
- Karpathy's views are controversial: RLHF is not real reinforcement learning, and Google and Meta are opposed to it
- It seems that everyone still has different opinions on whether RLHF and RL can be classified into the same category. AI guru Karpathy is here again to popularize the concept of artificial intelligence. Yesterday, he tweeted, "Reinforcement learning based on human feedback (RLHF) is just reinforcement learning (RL)." Karpathy's full text is explained as follows: RLHF is the third (and also the third) method for training large language models (LLM). The last) main stage, the first two stages are pre-training and supervised fine-tuning (SFT). I think RLHF is just barely considered RL and it is not widely recognized. RL is powerful, but RLHF is not. Let's look at the example of AlphaGo, which was trained using real RL. computer games
- AI 1015 2024-08-09 16:40:09