current location:Home > Technical Articles > Technology peripherals > AI
- Direction:
- All web3.0 Backend Development Web Front-end Database Operation and Maintenance Development Tools PHP Framework Daily Programming WeChat Applet Common Problem Other Tech CMS Tutorial Java System Tutorial Computer Tutorials Hardware Tutorial Mobile Tutorial Software Tutorial Mobile Game Tutorial
- Classify:
-
- The latest progress of the Ant Bailing large model: it already has native multi-modal capabilities
- On July 5, at the "Trusted Large Models Help Industrial Innovation and Development" forum at the 2024 World Artificial Intelligence Conference, Ant Group announced the latest research and development progress of its self-developed Bailing model: the Bailing model has the ability to "see" and "know". The native multi-modal capabilities of "listening", "speaking" and "drawing" can directly understand and train multi-modal data such as audio, video, pictures, text and so on. Native multimodality is considered to be the only way to AGI. In China, only a few large model manufacturers have achieved this capability. The reporter saw from the demonstration at the conference that multi-modal technology can make large models feel more like humans.
- AI 491 2024-07-10 15:06:57
-
- In just a few seconds, protein dynamics information can be accurately inferred. AI models such as Shandong University and Beijing Institute of Technology RMSF-net are published in Nature sub-journals.
- Editor | The dynamics of the KX protein are critical to understanding its mechanism. However, computationally predicting protein kinetic information is challenging. Here, a research team from Shandong University, BioMap, Beijing Institute of Technology, Hubei Medical College, Ningxia Medical University and King Abdullah University of Science and Technology (KAUST) proposed a neural network model RMSF-net , which outperforms previous methods and produces state-of-the-art results in large-scale protein dynamics data sets; the model can accurately infer a protein's dynamics information in seconds. By efficiently learning from the integration of experimental protein structure data and cryo-EM data, the method is able to accurately identify cryo-EM images and PDBs
- AI 780 2024-07-10 14:55:00
-
- Integrating multi-omics data, BGI team's graph neural network model SpatialGlue was published in Nature sub-journal
- Editor: KX Spatial Transcriptomics and Multi-omics Data Integration Spatial transcriptomics is a major development after single-cell transcriptomics, making the integration of multi-omics data crucial. SpatialGlue: A graph neural network model with dual attention mechanism. Research teams from the Singapore Agency for Science, Technology and Research (A*STAR), BGI and Renji Hospital Affiliated to Shanghai Jiao Tong University School of Medicine proposed a graph neural network called SpatialGlue. model, which integrates multi-omics data through a dual attention mechanism to reveal the histologically relevant structure of tissue samples in a spatially aware manner. Advantages of SpatialGlue SpatialGlue is able to combine multiple data modalities with their respective spatial contexts. Compared with other methods
- AI 534 2024-07-03 20:32:35
-
- Neural networks may no longer need activation functions? Layer Normalization also has non-linear expression!
- The AIxiv column is a column where this site publishes academic and technical content. In the past few years, the AIxiv column of this site has received more than 2,000 reports, covering top laboratories from major universities and companies around the world, effectively promoting academic exchanges and dissemination. If you have excellent work that you want to share, please feel free to contribute or contact us for reporting. Submission email: liyazhou@jiqizhixin.com; zhaoyunfeng@jiqizhixin.com The authors of this article are all from the team of Associate Professor Huang Lei, School of Artificial Intelligence, Beihang University and National Key Laboratory of Complex Critical Software Environment. The first author, Ni Yunhao, is a first-year graduate student, the second author, Guo Yuxin, is a third-year graduate student, the third author, Jia Junlong, is a second-year graduate student, and the corresponding author is Associate Professor Huang Lei
- AI 843 2024-07-03 14:11:33
-
- Harbin Institute of Technology proposes an innovative iterative reasoning framework DPE-MNER: giving full play to the potential of multi-modal representation
- The AIxiv column is a column where this site publishes academic and technical content. In the past few years, the AIxiv column of this site has received more than 2,000 reports, covering top laboratories from major universities and companies around the world, effectively promoting academic exchanges and dissemination. If you have excellent work that you want to share, please feel free to contribute or contact us for reporting. Submission email: liyazhou@jiqizhixin.com; zhaoyunfeng@jiqizhixin.com The author team of this article comes from the Social Computing and Information Retrieval Research Center of Harbin Institute of Technology. The author team consists of: Zheng Zihao, Zhang Zihan, Wang Zexin, Fu Ruiji, Liu Ming, Wang Zhongyuan, Qin soldiers. Multimodal representation of multimodal named entity recognition as a basis for constructing multimodal knowledge graphs
- AI 468 2024-07-03 10:44:16
-
- Runway and Luma are fighting again! Yann LeCun bombards: No matter how good you are, you are not a 'world model'
- Editor of the Machine Power Report: Yang Wen The wave of artificial intelligence represented by large models and AIGC has been quietly changing the way we live and work, but most people still don’t know how to use it. Therefore, we have launched the "AI in Use" column to introduce in detail how to use AI through intuitive, interesting and concise artificial intelligence use cases and stimulate everyone's thinking. We also welcome readers to submit innovative, hands-on use cases. The AI video industry is "fighting" again! On June 29, the well-known generative AI platform Runway announced that its latest model Gen-3Alpha has started testing for some users. On the same day, Luma launched a new keyframe feature and made it available to all users for free. It can be said that "you have a good plan, I have a wall ladder"
- AI 1016 2024-07-03 09:13:06
-
- Published in the Nature sub-journal, the topological Transformer model predicts multi-scale protein-ligand interactions to assist drug development
- Editor | Radish Skin A new artificial intelligence application will help researchers improve their drug development capabilities. The project, called TopoFormer, was developed by an interdisciplinary team led by Professor Guowei Wei of the Department of Mathematics at Michigan State University. TopoFormer converts three-dimensional information about molecules into data that can be used by typical AI-based drug interaction models, expanding the ability of these models to predict drug effectiveness. “With artificial intelligence, you can make drug development faster, more efficient, and cheaper,” said Wei, who holds appointments in both the Department of Biochemistry and Molecular Biology and the Department of Electrical and Computer Engineering. Professor Wei explained that in the United States,
- AI 1063 2024-07-02 15:23:21
-
- Can't wait for OpenAI's Q*, Huawei Noah's secret weapon MindStar to explore LLM reasoning is here first
- The AIxiv column is a column where this site publishes academic and technical content. In the past few years, the AIxiv column of this site has received more than 2,000 reports, covering top laboratories from major universities and companies around the world, effectively promoting academic exchanges and dissemination. If you have excellent work that you want to share, please feel free to contribute or contact us for reporting. Submission email: liyazhou@jiqizhixin.com; zhaoyunfeng@jiqizhixin.com The authors of this paper are from Kang Jikun, Li Xinze, Chen Xi, Amirreza Kazemi, and Chen Boxing from Huawei's Noah's Ark Laboratory in Montreal. Artificial intelligence (AI) has made great progress in the past decade, especially in natural language processing and computer vision.
- AI 531 2024-07-02 05:01:41
-
- Come quickly! Luchen Open-Sora can collect wool, and you can easily start video generation for 10 yuan.
- Recently, the field of video generation models is booming, with Vincent videos and Tu videos emerging in endless ways. However, even though there are many models on the market, most people still cannot experience them because they are not qualified for internal testing and can only look at the "models" and sigh. Not long ago, we reported on Luchen Technology’s Open-Sora model. As the world’s first open source Sora-like model, it not only performs well on multiple types of videos, but is also low-cost and available to everyone. Does it work? how to use? Let’s take a look at the review of this site. Recently, the new open source version 1.2 of Open-Sora can generate 720p high-definition videos up to 16 seconds. The official video effect is as follows: The generated effect is really amazing. It is no wonder that so many readers in the background want to get started and experience it. Compared
- AI 845 2024-07-02 04:22:00
-
- Amazon Cloud Innovation 'Neural Sparse Retrieval”: Only text matching is needed to achieve semantic search
- The AIxiv column is a column where this site publishes academic and technical content. In the past few years, the AIxiv column of this site has received more than 2,000 reports, covering top laboratories from major universities and companies around the world, effectively promoting academic exchanges and dissemination. If you have excellent work that you want to share, please feel free to contribute or contact us for reporting. Submission email: liyazhou@jiqizhixin.com; zhaoyunfeng@jiqizhixin.com The authors of this article are Dr. Yang Yang, head of machine learning, and machine learning engineers Geng Zhichao and Guan Cong from the OpenSearch China R&D team. OpenSearch is a pure open source search and real-time analysis engine project initiated by Amazon Cloud Technology
- AI 964 2024-07-02 02:55:57
-
- Compose a graphics program just by looking at a hand-drawn sketch. Berkeley, California, teaches diffusion models new skills
- It turns out that diffusion models can be used not only to generate images and videos, but also to synthesize new programs. Suppose we give the model a hand-drawn "5"-shaped graphic, it can modify the program through continuous mutations, and finally get a program that can output the target graphic. The model comes from a research team at the University of California, Berkeley, who proposed a new method of program synthesis that uses a neural diffusion model to directly manipulate syntax trees. Thesis 1 is Shreyas Kapur, a doctoral student at the school, whose supervisor is Stuart Russell, professor of computer science at the school. Paper title: DiffusionOnSyntaxTreesForProgramSynthesis Paper address: https://arxiv.
- AI 997 2024-07-02 01:14:04
-
- Defeating 25 molecular design algorithms, Georgia Tech, University of Toronto, and Cornell proposed large language model MOLLEO
- Author | Editor Wang Haorui, Georgia Institute of Technology | ScienceAI Molecular discovery, as an optimization problem, poses significant computational challenges because its optimization goals may not be differentiable. Evolutionary algorithms (EAs) are commonly used to optimize black-box targets in molecular discovery by traversing chemical space through random mutation and crossover, but this results in extensive and expensive target evaluation. In this work, researchers from the Georgia Institute of Technology, the University of Toronto, and Cornell University collaborated to propose Molecular Language Enhanced Evolutionary Optimization (MOLLEO) by integrating pre-trained large language models (LLMs) with chemical knowledge into evolutionary algorithms. , significantly improving the molecular optimization capabilities of evolutionary algorithms. The study is titled "EfficientEvolutionarySearc"
- AI 1213 2024-07-02 01:07:36
-
- ICML 2024| Large language model helps CLIP-based out-of-distribution detection tasks
- Machine learning models can show superior performance when the distributions of the training and test data sets are the same. However, in an open world environment, models often encounter out-of-distribution (OOD) samples. OOD samples may cause the model to behave unpredictable, and the consequences of errors may be fatal, especially in high-risk scenarios such as autonomous driving [1,2]. Therefore, OOD detection is crucial to ensure the reliability of machine learning models in actual deployment. Most OOD detection methods [1, 3] can effectively detect OOD samples based on well-trained in-distribution (ID) classifiers. Ran
- AI 550 2024-07-01 23:29:18
-
- ICML 2024 Spotlight | Realignment in decoding makes language models less hallucinatory and more consistent with human preferences
- The AIxiv column is a column where this site publishes academic and technical content. In the past few years, the AIxiv column of this site has received more than 2,000 reports, covering top laboratories from major universities and companies around the world, effectively promoting academic exchanges and dissemination. If you have excellent work that you want to share, please feel free to contribute or contact us for reporting. Submission email: liyazhou@jiqizhixin.com; zhaoyunfeng@jiqizhixin.com This article introduces a paper on language model alignment research, which was completed by doctoral students from three universities in Switzerland, the United Kingdom, and France, and researchers from Google DeepMind and Google Research. Among them, the corresponding author Ti
- AI 500 2024-07-01 22:09:56
-
- Developers are ecstatic! Meta's latest release of LLM Compiler achieves 77% automatic tuning efficiency
- Meta has developed an awesome LLMCompiler to help programmers write code more efficiently. Yesterday, the three major AI giants OpenAI, Google, and Meta teamed up to release the latest research results of their own large models - OpenAI launched CriticGPT, a new model specially designed to find bugs based on GPT-4 training, Google open sourced the 9B and 27B versions of Gemma2, and Meta came up with Developed a latest artificial intelligence breakthrough - LLMCompiler. This is a powerful set of open source models designed to optimize code and revolutionize compiler design. This innovation has the potential to change the way developers approach code optimization, making it faster, more efficient, and more economical
- AI 1246 2024-07-01 18:16:39