


Several excellent practices for building machine learning platforms on the cloud
Translator | Bugatti
Reviewer | Sun Shujuan
Most people are familiar with major technology platforms such as iOS, Windows and AWS. A platform is essentially a set of technologies that serve as a foundation for building, contributing to, experimenting with, and extending other applications. They bring many of today's advanced technological capabilities and cutting-edge customer experiences to the table.
To keep up with the scale and complexity of the technological capabilities enabled by big data, artificial intelligence, and machine learning, many companies are developing complex in-house platforms. In fact, Gartner predicts that cloud-native platforms will form the basis of more than 95% of new digital initiatives by 2025, up from less than 40% in 2021.
In my experience, enterprise technology platforms are transformative: they enable cross-functional teams to test, launch, and learn quickly, reduce duplication, standardize capabilities, and provide consistently integrated experience. In short, they help turn technology into a competitive advantage.
1. The evolution of enterprise platforms
#Many organizations leverage cloud-native platforms such as Kubernetes that can run the heavy lifting of artificial intelligence and machine learning. Laiyu is good at providing first-class customer experience. Capital One has become the first financial institution in the United States to fully invest in the cloud. The ability to redesign the architecture of the data environment is indispensable for consolidating its cloud-based platform capabilities. With this solid foundation, Capital One is better positioned to leverage big data to build new machine learning capabilities across its enterprise platforms to accelerate, enhance and deliver new, more meaningful customer experiences.
Much of Capital One’s work in this area has produced significant results for the company and its clients. Its fraud decisioning platform, for example, was built from the ground up to make complex, real-time decisions. By leveraging large amounts of data and enabling models to be updated in days instead of months, the platform helps protect millions of customers from credit card fraud and can be used by the company’s various stakeholders.
Based on my experience leading teams to deliver enterprise technology platforms, here are the key lessons and best practices learned along the way:
- Everything Starts with Teams: Build cross-functional teams made up of the best people, even if it slows down your work at first. Bigger teams aren't always better! The team must at least include product managers, engineers and designers. Staff these functions with people who truly understand the users of the platform. For example, if you are building a platform that will be used primarily by data scientists, hire a product manager who is a former data scientist, or add a data scientist to the leadership team. If your team is made up of people from multiple departments, make sure you have a common goal.
- Clearly define the end state first: Before you start building, take the time to clearly define the architecture and plan for the end state, and achieve your goals in an iterative manner. Make sure the architecture is designed for self-service and contribution from the start. Better yet, design your platform with the assumption that you will extend it to users outside of your organization or business unit. Also assume that over time, as technology changes, you want to be able to replace components.
- Estimate how long you think it will take, then double it: The important thing is to take the time to brainstorm all the capabilities you need to build from the beginning, and then devote the appropriate amount of effort to each part. Once the technical team combines this with speed and estimates how long it will take to build each feature, increase the buffer amount by 50%. In my experience, this estimate ends up being pretty accurate.
- Focus on business outcomes: Building a great platform can take a long time. It's important to sequence your work so that business value is continuously realized. This motivates the team, builds credibility, and creates a virtuous cycle.
- Strive for transparency and communication: freely communicate decisions, progress and roadmap with stakeholders. In addition to clarifying the work at hand, also clarify what is not currently a priority. Write good documentation to encourage others to contribute and easily join the platform.
- Start Small: Even the best testing and quality assurance (QA) environments may miss issues that only become apparent after deployment to production. For major changes that will have a clear impact on customers, always start with a small group and then expand the scope of application after seeing it is effective in a small-scale production environment.
- Be radically transparent and over-communicate: Share decisions, progress and roadmap freely with stakeholders. In addition to clarifying what you are doing, also clarify what you are not currently prioritizing. Invest in documentation that makes it easy to contribute and join the platform.
- Start Small: Even the best testing and QA environments may miss some issues that aren’t discovered until production. For big changes that will have meaningful customer impact, always start with a small group of people and then gradually increase as you see things happening in small-scale production. If possible, use employees only for the initial population when changes affect external customers.
- Pay attention to proper management: Platform owners should pay attention to platform performance. All problems should be revealed through control mechanisms and automated alerts. Exceptions should be handled quickly. Priority should be given to root cause analysis and changes to prevent problems from recurring. If there are no issues, celebrate appropriately so the team knows it is appreciated.
- If it seems too good to be true...Exception monitoring is a great way to ensure execution is consistent with intent. The goal is often zero exceptions. For example, the delay should not exceed 200 milliseconds. If the exception report never shows any exceptions, there's probably something wrong with your monitoring. Always force an exception to ensure it fires correctly. I understand this very well.
- A happy team is a productive team. Celebrate achievements, praise team members when they perform well, and create an environment of inner fulfillment. Regularly measure your team’s happiness and give your team a chance to discuss what would make them happier and try it out on their own to address areas of dissatisfaction.
#When a team has a strong culture strongly supported by the right platform technology, the opportunities are endless. By combining cloud-native platforms with large-scale data, companies can better advance and experiment with newer, more innovative products and experiences. When these experiences enable end users and customers to get the product or service they need, when they need it, it makes a huge difference.
Original link: https://venturebeat.com/ai/best-practices-for-building-machine-learning-platforms-on-the-cloud/
The above is the detailed content of Several excellent practices for building machine learning platforms on the cloud. For more information, please follow other related articles on the PHP Chinese website!

Hot AI Tools

Undresser.AI Undress
AI-powered app for creating realistic nude photos

AI Clothes Remover
Online AI tool for removing clothes from photos.

Undress AI Tool
Undress images for free

Clothoff.io
AI clothes remover

Video Face Swap
Swap faces in any video effortlessly with our completely free AI face swap tool!

Hot Article

Hot Tools

Notepad++7.3.1
Easy-to-use and free code editor

SublimeText3 Chinese version
Chinese version, very easy to use

Zend Studio 13.0.1
Powerful PHP integrated development environment

Dreamweaver CS6
Visual web development tools

SublimeText3 Mac version
God-level code editing software (SublimeText3)

Hot Topics



This site reported on June 27 that Jianying is a video editing software developed by FaceMeng Technology, a subsidiary of ByteDance. It relies on the Douyin platform and basically produces short video content for users of the platform. It is compatible with iOS, Android, and Windows. , MacOS and other operating systems. Jianying officially announced the upgrade of its membership system and launched a new SVIP, which includes a variety of AI black technologies, such as intelligent translation, intelligent highlighting, intelligent packaging, digital human synthesis, etc. In terms of price, the monthly fee for clipping SVIP is 79 yuan, the annual fee is 599 yuan (note on this site: equivalent to 49.9 yuan per month), the continuous monthly subscription is 59 yuan per month, and the continuous annual subscription is 499 yuan per year (equivalent to 41.6 yuan per month) . In addition, the cut official also stated that in order to improve the user experience, those who have subscribed to the original VIP

Improve developer productivity, efficiency, and accuracy by incorporating retrieval-enhanced generation and semantic memory into AI coding assistants. Translated from EnhancingAICodingAssistantswithContextUsingRAGandSEM-RAG, author JanakiramMSV. While basic AI programming assistants are naturally helpful, they often fail to provide the most relevant and correct code suggestions because they rely on a general understanding of the software language and the most common patterns of writing software. The code generated by these coding assistants is suitable for solving the problems they are responsible for solving, but often does not conform to the coding standards, conventions and styles of the individual teams. This often results in suggestions that need to be modified or refined in order for the code to be accepted into the application

Large Language Models (LLMs) are trained on huge text databases, where they acquire large amounts of real-world knowledge. This knowledge is embedded into their parameters and can then be used when needed. The knowledge of these models is "reified" at the end of training. At the end of pre-training, the model actually stops learning. Align or fine-tune the model to learn how to leverage this knowledge and respond more naturally to user questions. But sometimes model knowledge is not enough, and although the model can access external content through RAG, it is considered beneficial to adapt the model to new domains through fine-tuning. This fine-tuning is performed using input from human annotators or other LLM creations, where the model encounters additional real-world knowledge and integrates it

To learn more about AIGC, please visit: 51CTOAI.x Community https://www.51cto.com/aigc/Translator|Jingyan Reviewer|Chonglou is different from the traditional question bank that can be seen everywhere on the Internet. These questions It requires thinking outside the box. Large Language Models (LLMs) are increasingly important in the fields of data science, generative artificial intelligence (GenAI), and artificial intelligence. These complex algorithms enhance human skills and drive efficiency and innovation in many industries, becoming the key for companies to remain competitive. LLM has a wide range of applications. It can be used in fields such as natural language processing, text generation, speech recognition and recommendation systems. By learning from large amounts of data, LLM is able to generate text

Machine learning is an important branch of artificial intelligence that gives computers the ability to learn from data and improve their capabilities without being explicitly programmed. Machine learning has a wide range of applications in various fields, from image recognition and natural language processing to recommendation systems and fraud detection, and it is changing the way we live. There are many different methods and theories in the field of machine learning, among which the five most influential methods are called the "Five Schools of Machine Learning". The five major schools are the symbolic school, the connectionist school, the evolutionary school, the Bayesian school and the analogy school. 1. Symbolism, also known as symbolism, emphasizes the use of symbols for logical reasoning and expression of knowledge. This school of thought believes that learning is a process of reverse deduction, through existing

Editor |ScienceAI Question Answering (QA) data set plays a vital role in promoting natural language processing (NLP) research. High-quality QA data sets can not only be used to fine-tune models, but also effectively evaluate the capabilities of large language models (LLM), especially the ability to understand and reason about scientific knowledge. Although there are currently many scientific QA data sets covering medicine, chemistry, biology and other fields, these data sets still have some shortcomings. First, the data form is relatively simple, most of which are multiple-choice questions. They are easy to evaluate, but limit the model's answer selection range and cannot fully test the model's ability to answer scientific questions. In contrast, open-ended Q&A

Editor | KX In the field of drug research and development, accurately and effectively predicting the binding affinity of proteins and ligands is crucial for drug screening and optimization. However, current studies do not take into account the important role of molecular surface information in protein-ligand interactions. Based on this, researchers from Xiamen University proposed a novel multi-modal feature extraction (MFE) framework, which for the first time combines information on protein surface, 3D structure and sequence, and uses a cross-attention mechanism to compare different modalities. feature alignment. Experimental results demonstrate that this method achieves state-of-the-art performance in predicting protein-ligand binding affinities. Furthermore, ablation studies demonstrate the effectiveness and necessity of protein surface information and multimodal feature alignment within this framework. Related research begins with "S

According to news from this site on August 1, SK Hynix released a blog post today (August 1), announcing that it will attend the Global Semiconductor Memory Summit FMS2024 to be held in Santa Clara, California, USA from August 6 to 8, showcasing many new technologies. generation product. Introduction to the Future Memory and Storage Summit (FutureMemoryandStorage), formerly the Flash Memory Summit (FlashMemorySummit) mainly for NAND suppliers, in the context of increasing attention to artificial intelligence technology, this year was renamed the Future Memory and Storage Summit (FutureMemoryandStorage) to invite DRAM and storage vendors and many more players. New product SK hynix launched last year
