In recent years, many major breakthroughs have been made in the field of machine learning, and human resource management service products driven by artificial intelligence technology also have a huge and dynamic market. More and more companies and government agencies are gradually thinking about applying machine learning technology to human resource management, making effective decisions through neural networks, and accurately predicting the results of human resource management.
This article introduces four aspects of applying machine learning to human resource management research, mainly including technical difficulties, introduction to human resources management decision-making systems, system design methods and system security. It is hoped that readers can have a preliminary understanding of related research. .
In 2019, CEOs of 20 large companies in the United States conducted relevant seminars. The results showed that the application of machine learning technology faces unique challenges in the field of human resources management. . Developing valuable HRM decision-making systems presents not only technical challenges, but also barriers to vectoring the inherent complexity of HRM outcomes, as well as difficult-to-address data, ethical, legal constraints and concerns about affected employees or other stakeholders. a selection process that some believe to be controversial. HR management decisions need to avoid selection procedures that are susceptible to legal challenges or that are considered controversial by employees or other stakeholders.
The summary includes the following aspects:
Introduction to the human resources management decision-making system
The implementation of the human resources management decision-making system faces the following challenges:
This framework demonstrates the principles of a machine learning system for conceptualizing design and human resource management. The idea behind the framework is that system design is inseparable from the system's highest priority goals. HRM objectives help designers choose from the many possible ways in which machine learning can support the HRM decision-making process. The design of the implementation in turn affects how the system is evaluated. For example, systems that automate decision-making may be evaluated based on their accuracy or other important criteria, and systems that provide inputs must be judged based on the accuracy of the inputs and how they affect the overall decision outcome. If the system fails to meet security standards, the implementation design must be modified until the designer is able to obtain a system that is valuable for human resource management objectives and is capable of meeting security parameters.
Figure Machine Learning System Framework
In the early stages of development, there are many design options to integrate machine learning-based Inputs are integrated into decision making. Designs vary in terms of timing (e.g., before or after humans make decisions) and degree of influence (e.g., recommending an option versus directing attention to important features). Five major design implementations of machine learning decision-making systems will be highlighted here:
At decision time, machine learning systems score human resource management records and make decisions automatically without the involvement of human decision-makers .
2. Recommended. Machine learning systems provide recommendations to human decision-makers as additional input.
3. Score. Machine learning systems provide scores to humans as additional input.
4. Summary. Machine learning systems automatically summarize for human decision-makers.
5. Audit. Machine learning systems flag abnormalities for review by human decision-makers as part of the audit process.
The design process begins with determining the priority goals of the machine learning system. Different combinations of goals require different design implementations, as shown in the table.
Picture
These objectives also indicate potential measures of the effectiveness of the evaluation process. For example, if the goal is to reduce workload, the system should reduce the number of human decision-makers or the time they spend recording ratings; if the goal is to improve human decision-making, the system should help improve the quality of decisions by measuring them with evidence. Better contribute to important HR management results.
Machine learning systems that automatically summarize narrative records can serve as a model for decision support. Most of a person's HR records are divided into two types of free-form text and person attributes. Free-form text such as task lists, descriptions of responsibilities, and summaries of key accomplishments. Personnel attributes are pre-quantified, interpretable, management-useful data, such as years of experience, order of merit, or promotion test scores. While the latter type of information is easier to process and use in models or visualizations, the former type of information is also needed to make fully informed human resource management decisions.
Management processing decisions require thoughtful review of records and a manual review or scoring process by experienced personnel. Of the various design implementations considered to support manual review, "summary" is the most versatile. This is the only design that is moderately or highly consistent with all human resource management goals. Automated summaries are useful for providing feedback, increasing transparency, and improving the accuracy of human decision-making, and they are at least somewhat useful for standardizing and reducing manual workload. At the same time, the summary implementation maintains a high degree of manual control over the decision-making process, so it is more likely than other designs to meet security standards. In fact, the summary highlights the elements of the text that the system deems important and, therefore, is an explanation of the system's decisions. Therefore, summaries can serve as a useful aid in helping managers understand model outputs in other design implementations.
Human resources management decisions are a vital force that affects the future of the enterprise. Therefore, the principle of “first do no harm” must be adopted when making significant changes to the decision-making process. As investment in machine learning increases, a wealth of research and policy documents aim to provide normative guidance for the responsible and ethical use of machine learning (and artificial intelligence more generally).
For example, existing rules and frameworks to protect member privacy will continue to apply to any development projects. During development and deployment, three principles are particularly relevant for testing systems that require machine learning systems to be accurate, fair, and interpretable:
Accuracy means that a machine learning system or the model it contains is correct with a high probability predict the outcome of interest.
Fairness means that the machine learning system treats subgroups equally.
Explainability means humans can understand the factors and relationships that lead to the results of a machine learning system.
These security standards sometimes conflict with each other. To increase fairness, designers may impose limitations on the system that reduce its accuracy or interpretability. To increase interpretability, system designers may use more interpretable (but less flexible) modeling methods, which may impact accuracy and fairness. Testing must include balancing accuracy, fairness, and interpretability to arrive at a design that meets human resource management objectives as well as legal and ethical constraints.
Regarding fairness, it is important to note that there is no single definition of fairness, and it is often impossible to satisfy competitive types of fairness. Therefore, agencies must choose a definition to move forward with testing. A distinction is made here between procedural fairness, which ensures that an HR management process or algorithm treats members of different subgroups equally, and outcome fairness, which checks whether a model or process outcome is biased.
Finally, explainability is critical to achieving human resource management goals because people may ignore or abuse the system if they don't understand how it contributes to better decision-making. Furthermore, defining explainability is inseparable from the target audience, as different types of users require different levels of explanation. Designers can consider using inherently interpretable models to increase interpretability, and they can also conduct human-in-the-loop testing to assess how well people understand the system's functionality.
This article mainly introduces the research of machine learning in the field of human resources management from four aspects: technical difficulties, introduction to human resources management decision-making systems, system design methods and system security. We hope it can be helpful to readers who want to have a preliminary understanding of this research.
The above is the detailed content of Let's talk about the collision between machine learning and human resources management?. For more information, please follow other related articles on the PHP Chinese website!