CN109034398B - Gradient lifting tree model construction method and device based on federal training and storage medium - Google Patents

Gradient lifting tree model construction method and device based on federal training and storage medium Download PDF

Info

Publication number
CN109034398B
CN109034398B CN201810918867.3A CN201810918867A CN109034398B CN 109034398 B CN109034398 B CN 109034398B CN 201810918867 A CN201810918867 A CN 201810918867A CN 109034398 B CN109034398 B CN 109034398B
Authority
CN
China
Prior art keywords
round
node
splitting
sample
training
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201810918867.3A
Other languages
Chinese (zh)
Other versions
CN109034398A (en
Inventor
成柯葳
范涛
刘洋
陈天健
杨强
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
WeBank Co Ltd
Original Assignee
WeBank Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by WeBank Co Ltd filed Critical WeBank Co Ltd
Priority to CN201810918867.3A priority Critical patent/CN109034398B/en
Publication of CN109034398A publication Critical patent/CN109034398A/en
Application granted granted Critical
Publication of CN109034398B publication Critical patent/CN109034398B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Landscapes

  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The invention discloses a feature selection method based on federal training, which comprises the following steps: performing federal training on two aligned training samples by adopting an XGboost algorithm to construct a gradient lifting tree model, wherein the gradient lifting tree model comprises a plurality of regression trees, and one split node of each regression tree corresponds to one feature of the training sample; counting average profit values of split nodes corresponding to the same feature in the gradient lifting tree model, and taking the average profit values as scores of the corresponding features; and carrying out feature sorting based on the scores of the features, and outputting sorting results for feature selection, wherein if the features which do not correspond to the split nodes exist in the training sample, the default scores are used for the features. The invention also discloses a feature selection device based on federal training and a computer readable storage medium. The invention realizes federal training modeling by using training samples of different data parties, thereby realizing the feature selection of multi-party sample data.

Description

Gradient lifting tree model construction method and device based on federal training and storage medium
Technical Field
The invention relates to the technical field of machine learning, in particular to a gradient lifting tree model construction method and device based on federal training and a computer readable storage medium.
Background
In the current information age, some behaviors of people can be shown through data, such as consumption behaviors, so that big data analysis is derived, a corresponding behavior analysis model is built through machine learning, and then the behaviors of people can be classified or predicted based on the behavior characteristics of users.
In the existing machine learning technology, a party is usually used for independently training sample data, namely unilateral modeling. Meanwhile, based on the established mathematical model, the characteristics with relatively high importance degree in the sample characteristic set can be determined. However, in many cross-domain big data analysis scenarios, such as where the user has both consumption behavior and loan behavior, the user consumption behavior data is generated at the consumption service provider, and the user loan behavior data is generated at the financial service provider, if the financial service provider needs to predict the user's loan behavior based on the user's consumption behavior characteristics, then the consumption behavior data of the consumption service provider needs to be used and machine-learned together with the loan behavior data of the present party to construct a prediction model.
Therefore, aiming at the application scene, a new modeling mode is needed to realize the joint training of sample data of different data providers, and further realize that both parties participate in modeling together.
Disclosure of Invention
The invention mainly aims to provide a gradient lifting tree model construction method and device based on federal training and a computer readable storage medium, and aims to solve the technical problem that the prior art cannot realize joint training of sample data of different data providers and further cannot realize joint participation modeling of both parties.
In order to achieve the above object, the present invention provides a gradient lifting tree model construction method based on federal training, which includes the following steps:
performing federal training on two aligned training samples by adopting an XGboost algorithm to construct a gradient lifting tree model, wherein the gradient lifting tree model comprises a plurality of regression trees, and one split node of each regression tree corresponds to one feature of the training sample;
counting average profit values of split nodes corresponding to the same feature in the gradient lifting tree model, and taking the average profit values as scores of the corresponding features;
And carrying out feature sorting based on the scores of the features, and outputting sorting results for feature selection, wherein if the features which do not correspond to the split nodes exist in the training sample, the default scores are used for the features.
Optionally, the two aligned training samples are a first training sample and a second training sample, respectively;
the first training sample attribute comprises a sample ID and a part of sample characteristics, and the second training sample attribute comprises a sample ID, another part of sample characteristics and a data tag;
the first training sample is provided by and stored locally to the first data party and the second training sample is provided by and stored locally to the second data party.
Optionally, the performing federal training on the two aligned training samples using the XGboost algorithm to construct the gradient-lifting tree model includes:
on the second data side, acquiring a first-order gradient and a second-order gradient of each training sample in a sample set corresponding to the node splitting of the round;
if the first-round node splitting is the first-round node splitting for constructing a regression tree, encrypting the first-order gradient and the second-order gradient, and then sending the first-order gradient and the second-order gradient to the first data side together with the sample ID of the sample set so as to calculate the benefit value of the split node of the training sample locally corresponding to the sample ID in each splitting mode on the first data side based on the encrypted first-order gradient and the second-order gradient;
If the node splitting of the round is the non-first round node splitting for constructing the regression tree, a sample ID of the sample set is sent to the first data side so as to be used for calculating the benefit value of the split node of the training sample which locally corresponds to the sample ID in each splitting mode along the first-order gradient and the second-order gradient used for the first round node splitting at the first data side;
the second data party receives the encryption gain values of all the split nodes returned by the first data party and decrypts the encryption gain values;
on the second data side, calculating the benefit value of splitting nodes of a training sample corresponding to the sample ID locally in each splitting mode based on the first-order gradient and the second-order gradient;
based on the profit values of all the split nodes calculated by the two parties, determining the global optimal split node for splitting the node of the round;
based on the global optimal splitting node of the current round of node splitting, splitting is carried out on a sample set corresponding to the current node, and a new node is generated to construct a regression tree of the gradient lifting tree model.
Optionally, before the step of obtaining the first-order gradient and the second-order gradient of each training sample in the sample set corresponding to the node splitting of the present round at the second data side, the method further includes:
When node splitting is carried out, judging whether the node splitting of the round corresponds to constructing a first regression tree;
if the node splitting of the current round corresponds to the first regression tree construction, judging whether the node splitting of the current round is the first node splitting of the first regression tree construction;
if the node splitting of the round is the first round node splitting for constructing the first regression tree, initializing a first-order gradient and a second-order gradient of each training sample in the sample set corresponding to the node splitting of the round at the second data side; if the node splitting of the first round is the node splitting of the first round which is not the first round for constructing the regression tree, the first-order gradient and the second-order gradient used by the node splitting of the first round are used;
if the node splitting of the round corresponds to the construction of the non-first regression tree, judging whether the node splitting of the round is the first round node splitting of the construction of the non-first regression tree;
if the node splitting of the round is the first round node splitting for constructing a non-first regression tree, updating a first-order gradient and a second-order gradient according to the last round of federal training; if the node splitting of the first round is the node splitting of the first round, which constructs a regression tree of the first round, the first-order gradient and the second-order gradient used by the node splitting of the first round are used.
Optionally, the gradient lifting tree model building method based on federal training further comprises:
When generating new nodes to construct a regression tree of the gradient lifting tree model, judging whether the depth of the regression tree of the round reaches a preset depth threshold value or not at the second data side;
if the depth of the regression tree of the current round reaches the preset depth threshold, stopping node splitting to obtain a regression tree of the gradient lifting tree model, and if not, continuing the node splitting of the next round.
Optionally, the gradient lifting tree model building method based on federal training further comprises:
when node splitting is stopped, judging whether the total number of the round of regression trees reaches a preset number threshold value at the second data side;
if the total number of the regression trees in the round reaches the preset number threshold, stopping the federal training, otherwise, continuing the next round of federal training.
Optionally, the gradient lifting tree model building method based on federal training further comprises:
on the second data side, recording the related information of the global optimal splitting node determined by each round of node splitting;
wherein the related information includes: a provider of the corresponding sample data, a signature encoding of the corresponding sample data, and a benefit value.
Optionally, the counting the average benefit value of the split nodes corresponding to the same feature in the gradient-lifted tree model includes:
And on the second data side, taking each global optimal splitting node as a splitting node of each regression tree in the gradient lifting tree model, and counting the average benefit value of the splitting nodes corresponding to the same feature codes.
Further, in order to achieve the above object, the present invention further provides a gradient lifting tree model building device based on federal training, where the gradient lifting tree model building device based on federal training includes a memory, a processor, and a gradient lifting tree model building program stored on the memory and capable of running on the processor, where the gradient lifting tree model building program, when executed by the processor, implements the steps of the gradient lifting tree model building method based on federal training as set forth in any one of the above.
Further, to achieve the above object, the present invention further provides a computer readable storage medium having stored thereon a gradient-lifting tree-model building program which, when executed by a processor, implements the steps of the federally-trained gradient-lifting tree-model building method according to any one of the above.
The method adopts an XGboost algorithm to perform federal training on two aligned training samples to construct a gradient lifting tree model, wherein the gradient lifting tree model is a regression tree set and comprises a plurality of regression trees, and a split node of each regression tree corresponds to one feature of the training sample; the average profit value of the split nodes corresponding to the same feature in the tree model is promoted through statistical gradient, the average profit value is used as the score of the corresponding feature, and further the scoring of the features of the two training sample data is achieved; and finally, carrying out feature sorting based on the scores of the features, and outputting sorting results for feature selection, wherein the higher the score is, the higher the importance of the features is. The invention realizes federal training modeling by using training samples of different data parties, thereby realizing the feature selection of multi-party sample data.
Drawings
FIG. 1 is a schematic diagram of a hardware operating environment involved in an embodiment of a federally trained gradient-lifting tree model building apparatus of the present invention;
FIG. 2 is a schematic flow chart of an embodiment of a federally trained gradient-lifted tree model construction method according to the present invention;
FIG. 3 is a detailed flowchart of the embodiment of step S10 in FIG. 2;
FIG. 4 is a schematic diagram of training results of an embodiment of a federally trained gradient-lifted tree model construction method according to the present invention.
The achievement of the objects, functional features and advantages of the present invention will be further described with reference to the accompanying drawings, in conjunction with the embodiments.
Detailed Description
It should be understood that the specific embodiments described herein are for purposes of illustration only and are not intended to limit the scope of the invention.
The invention provides a gradient lifting tree model building device based on federal training.
As shown in fig. 1, fig. 1 is a schematic structural diagram of a hardware operation environment related to an embodiment scheme of a gradient lifting tree model building device based on federal training.
The gradient lifting tree model building device based on federal training can be a personal computer or a server and other equipment with calculation processing capacity.
As shown in fig. 1, the gradient-lifting tree model building apparatus based on federal training may include: a processor 1001, such as a CPU, a network interface 1004, a user interface 1003, a memory 1005, a communication bus 1002. Wherein the communication bus 1002 is used to enable connected communication between these components. The user interface 1003 may include a Display, an input unit such as a Keyboard (Keyboard), and the optional user interface 1003 may further include a standard wired interface, a wireless interface. The network interface 1004 may optionally include a standard wired interface, a wireless interface (e.g., WI-FI interface). The memory 1005 may be a high-speed RAM memory or a stable memory (non-volatile memory), such as a disk memory. The memory 1005 may also optionally be a storage device separate from the processor 1001 described above.
Those skilled in the art will appreciate that the federally trained gradient-lifting tree model building apparatus structure shown in fig. 1 is not limiting of the apparatus and may include more or fewer components than shown, or may combine certain components, or a different arrangement of components.
As shown in fig. 1, an operating system, a network communication module, a user interface module, and a file copy program may be included in the memory 1005, which is a type of computer storage medium.
In the gradient lifting tree model building device based on federal training shown in fig. 1, the network interface 1004 is mainly used for connecting a background server and performing data communication with the background server; the user interface 1003 is mainly used for connecting a client (user side) and performing data communication with the client; and the processor 1001 may be configured to call a gradient-lifting tree model building program stored in the memory 1005 and perform the following operations:
performing federal training on two aligned training samples by adopting an XGboost algorithm to construct a gradient lifting tree model, wherein the gradient lifting tree model comprises a plurality of regression trees, and one split node of each regression tree corresponds to one feature of the training sample;
counting average profit values of split nodes corresponding to the same feature in the gradient lifting tree model, and taking the average profit values as scores of the corresponding features;
and carrying out feature sorting based on the scores of the features, and outputting sorting results for feature selection, wherein if the features which do not correspond to the split nodes exist in the training sample, the default scores are used for the features.
Further, the two aligned training samples are a first training sample and a second training sample, respectively; the first training sample attribute comprises a sample ID and a part of sample characteristics, and the second training sample attribute comprises a sample ID, another part of sample characteristics and a data tag; the first training sample is provided by a first data party and is stored locally to the first data party, and the second training sample is provided by a second data party and is stored locally to the second data party; the processor 1001 calls the gradient-lifting tree-model building program stored in the memory 1005 to further perform the following operations:
On the second data side, acquiring a first-order gradient and a second-order gradient of each training sample in a sample set corresponding to the node splitting of the round;
if the first-round node splitting is the first-round node splitting for constructing a regression tree, encrypting the first-order gradient and the second-order gradient, and then sending the first-order gradient and the second-order gradient to the first data side together with the sample ID of the sample set so as to calculate the benefit value of the split node of the training sample locally corresponding to the sample ID in each splitting mode on the first data side based on the encrypted first-order gradient and the second-order gradient;
if the node splitting of the round is the non-first round node splitting for constructing the regression tree, a sample ID of the sample set is sent to the first data side so as to be used for calculating the benefit value of the split node of the training sample which locally corresponds to the sample ID in each splitting mode along the first-order gradient and the second-order gradient used for the first round node splitting at the first data side;
the second data party receives the encryption gain values of all the split nodes returned by the first data party and decrypts the encryption gain values;
on the second data side, calculating the benefit value of splitting nodes of a training sample corresponding to the sample ID locally in each splitting mode based on the first-order gradient and the second-order gradient;
Based on the profit values of all the split nodes calculated by the two parties, determining the global optimal split node for splitting the node of the round;
based on the global optimal splitting node of the current round of node splitting, splitting is carried out on a sample set corresponding to the current node, and a new node is generated to construct a regression tree of the gradient lifting tree model.
Further, the processor 1001 calls the gradient-lifting tree model building program stored in the memory 1005 to further perform the following operations:
when node splitting is carried out, judging whether the node splitting of the round corresponds to constructing a first regression tree;
if the node splitting of the current round corresponds to the first regression tree construction, judging whether the node splitting of the current round is the first node splitting of the first regression tree construction;
if the node splitting of the round is the first round node splitting for constructing the first regression tree, initializing a first-order gradient and a second-order gradient of each training sample in the sample set corresponding to the node splitting of the round at the second data side; if the node splitting of the first round is the node splitting of the first round which is not the first round for constructing the regression tree, the first-order gradient and the second-order gradient used by the node splitting of the first round are used;
if the node splitting of the round corresponds to the construction of the non-first regression tree, judging whether the node splitting of the round is the first round node splitting of the construction of the non-first regression tree;
If the node splitting of the round is the first round node splitting for constructing a non-first regression tree, updating a first-order gradient and a second-order gradient according to the last round of federal training; if the node splitting of the first round is the node splitting of the first round, which constructs a regression tree of the first round, the first-order gradient and the second-order gradient used by the node splitting of the first round are used.
Further, the processor 1001 calls the gradient-lifting tree model building program stored in the memory 1005 to further perform the following operations:
on the first data side, calculating the benefit value of splitting nodes of a training sample locally corresponding to the sample ID in each splitting mode based on the encrypted first-order gradient and second-order gradient;
or on the first data side, calculating the benefit value of splitting nodes of a training sample corresponding to the sample ID locally in each splitting mode along with a first-order gradient and a second-order gradient used for splitting the first-round nodes;
encrypting the benefit values of all the split nodes and then sending the encrypted benefit values to the second data party.
Further, the processor 1001 calls the gradient-lifting tree model building program stored in the memory 1005 to further perform the following operations:
when generating new nodes to construct a regression tree of the gradient lifting tree model, judging whether the depth of the regression tree of the round reaches a preset depth threshold value or not at the second data side;
If the depth of the regression tree of the current round reaches the preset depth threshold, stopping node splitting to obtain a regression tree of the gradient lifting tree model, and if not, continuing the node splitting of the next round.
Further, the processor 1001 calls the gradient-lifting tree model building program stored in the memory 1005 to further perform the following operations:
when node splitting is stopped, judging whether the total number of the round of regression trees reaches a preset number threshold value at the second data side;
if the total number of the regression trees in the round reaches the preset number threshold, stopping the federal training, otherwise, continuing the next round of federal training.
Further, the processor 1001 calls the gradient-lifting tree model building program stored in the memory 1005 to further perform the following operations:
on the second data side, recording the related information of the global optimal splitting node determined by each round of node splitting;
wherein the related information includes: a provider of the corresponding sample data, a signature encoding of the corresponding sample data, and a benefit value.
Further, the processor 1001 calls the gradient-lifting tree model building program stored in the memory 1005 to further perform the following operations:
and on the second data side, taking each global optimal splitting node as a splitting node of each regression tree in the gradient lifting tree model, and counting the average benefit value of the splitting nodes corresponding to the same feature codes.
Based on the hardware operation environment related to the embodiment scheme of the gradient lifting tree model building device based on the federal training, the following embodiments of the gradient lifting tree model building method based on the federal training are provided.
Referring to fig. 2, fig. 2 is a flowchart illustrating an embodiment of a federally trained gradient-lifted tree model construction method according to the present invention. In this embodiment, the method for constructing the gradient lifting tree model based on federal training includes the following steps:
step S10, performing federal training on two aligned training samples by adopting an XGboost algorithm to construct a gradient lifting tree model, wherein the gradient lifting tree model comprises a plurality of regression trees, and one split node of each regression tree corresponds to one feature of the training sample;
the XGboost (eXtreme Gradient Boosting) algorithm is an improvement on the Boosting algorithm based on the GBDT (Gradient Boosting Decision Tree, gradient lifting tree) algorithm, wherein the internal decision tree is a regression tree, the algorithm output is a set of regression trees and comprises a plurality of regression trees, the basic idea of training and learning is to traverse all segmentation methods of all features of a training sample (namely, a node splitting mode), select a segmentation method with the minimum loss, obtain two leaves (namely, splitting nodes to generate new nodes), and then continue traversing until:
(1) If the splitting stopping condition is met, outputting a regression tree;
(2) And if the iteration stopping condition is met, outputting a regression tree set.
In this embodiment, the training samples used by the XGboost algorithm are two independent training samples, that is, each training sample belongs to a different data party. If two training samples are regarded as one integral training sample, the two training samples belong to different data parties, so that the integral training sample can be regarded as being segmented, and the training samples are different features (sample longitudinal cutting) of the same sample.
Furthermore, it is possible to provide a device for the treatment of a disease. Because the two training samples belong to different data parties respectively, in order to realize federal training modeling, sample alignment needs to be carried out on original sample data provided by the two parties.
In this embodiment, federal training means that the sample training process is completed by cooperation of two data parties, and the final training results in a regression tree contained in the gradient lifting tree model, where the split nodes of the regression tree correspond to the features of the training samples of both parties.
Step S20, counting average profit values of split nodes corresponding to the same feature in the gradient lifting tree model, and taking the average profit values as scores of the corresponding features;
In the XGboost algorithm, when all segmentation methods of all features of a training sample are traversed, the advantages and disadvantages of the segmentation methods are evaluated through the benefit value, and the segmentation method with the minimum loss is selected for each split node. Therefore, the profit value of the split node can be used as the evaluation basis of the feature importance, and the bigger the profit value of the split node is, the smaller the node segmentation loss is, and the bigger the importance of the feature corresponding to the split node is.
In this embodiment, since the gradient lifting tree model obtained by training includes multiple regression trees, and different regression trees may use the same feature to perform node segmentation, it is necessary to count average benefit values of split nodes corresponding to the same feature in all regression trees included in the gradient lifting tree model, and use the average benefit values as scores of the corresponding features.
And step S30, feature sorting is performed based on the scores of the features, and sorting results are output for feature selection, wherein if the features of the non-corresponding split nodes exist in the training sample, default scores are used for the features.
In this embodiment, the degree of importance of the feature is represented by the degree of importance of the feature, and after the score of each feature is obtained, the feature is ranked and the ranking result is output, for example, from high to low, so that the importance of the feature ranked in front is higher than that of the feature ranked in back. Thus, by feature selection, features in the sample that are not related to sample prediction or classification can be eliminated. For example, the student samples include: gender, school achievements, attendance rate and number of times of surfacing, if the classification target is a three-good student and a non-three-good student, the characteristic gender is obviously irrelevant to or not relevant to the three-good student, so that the students can be removed.
In the embodiment, two aligned training samples are subjected to federal training by adopting an XGboost algorithm to construct a gradient lifting tree model, wherein the gradient lifting tree model is a regression tree set and comprises a plurality of regression trees, and one splitting node of each regression tree corresponds to one characteristic of the training sample; the average profit value of the split nodes corresponding to the same feature in the tree model is promoted through statistical gradient, the average profit value is used as the score of the corresponding feature, and further the scoring of the features of the two training sample data is achieved; and finally, carrying out feature sorting based on the scores of the features, and outputting sorting results for feature selection, wherein the higher the score is, the higher the importance of the features is. The embodiment realizes federal training modeling by using training samples of different data parties, and further realizes feature selection of multi-party sample data.
Further, to facilitate description of a specific implementation of the joint training of the present invention, this embodiment is specifically illustrated with two independent training samples.
In this embodiment, the first data party provides a first training sample, where the first training sample attribute includes a sample ID and a portion of sample features; the second data party provides a second training sample, the second training sample attribute including a sample ID, another portion of sample characteristics, and a data tag.
Where a sample characteristic refers to a characteristic that the sample represents or has, such as the sample is a person, the corresponding sample characteristic may be age, gender, income, academic, etc. The data tag is used for classifying a plurality of different samples, and the classification result is obtained by judging according to the characteristics of the samples.
The federal training modeling method has the main significance of realizing the bidirectional privacy protection of the sample data of both parties. Thus, during the federal training process, the first training sample is stored locally on the first data party and the second training sample is stored locally on the second data party, e.g., the data in Table 1 below is provided by the first data party and stored locally on the first data party, and the data in Table 2 below is provided by the second data party and stored locally on the second data party.
TABLE 1
As shown in table 1 above, the first training sample attributes include sample IDs (X1 to X5), age feature, gender feature, and Amount of given credit feature.
TABLE 2
Sample ID Bill Payment Education Lable
X1 3102 2 24
X2 17250 3 14
X3 14027 2 16
X4 6787 1 10
X5 280 1 26
As shown in Table 2 above, the second training sample attribute includes sample IDs (X1-X5), bill Payment feature, reduce feature, and data tag Lable.
Further, referring to fig. 3, fig. 3 is a schematic diagram of a refinement process of an embodiment of step S10 in fig. 2. Based on the above embodiment, in this embodiment, the step S10 specifically includes:
Step S101, at the side of the second data side, acquiring a first-order gradient and a second-order gradient of each training sample in a sample set corresponding to the node splitting of the round;
the XGboost algorithm is a machine learning modeling method that requires the use of a classifier (i.e., a classification function) to map sample data to one of a given class, and thus can be applied to data prediction. In learning classification rules with a classifier, a loss function is required to determine the magnitude of a fitting error of machine learning.
In this embodiment, when node splitting is performed each time, a first-order gradient and a second-order gradient of each training sample in the sample set corresponding to the node splitting of the present round are obtained at the second data side.
The gradient lifting tree model needs to be subjected to multiple rounds of federal training, each round of federal training correspondingly generates a regression tree, and the generation of the regression tree needs to be subjected to multiple node splitting.
Therefore, in each round of federal training, the first node splitting uses the training sample stored initially, the next node splitting uses the training sample of the new node corresponding sample set generated by the last node splitting, and in the same round of federal training, each round of node splitting uses the first-order gradient and the second-order gradient used by the first-round node splitting. The next federal training will update the first and second gradients used in the previous federal training round using the results of the previous federal training round.
The XGboost algorithm supports a self-defined loss function, and uses the self-defined loss function to calculate a first-order partial derivative and a second-order partial derivative of the objective function, so as to correspondingly obtain a first-order gradient and a second-order gradient of sample data to be trained locally.
Based on the description of the XGboost algorithm and the gradient-lifting tree model in the above embodiments, the split nodes need to be determined for constructing the regression tree, and the split nodes can be determined by the benefit values. The calculation formula of the gain value gain is as follows:
wherein I is L An inclusive sample set representing a left child node after a current node split, I R An inclusive sample set, g, representing the right child node after the current node split i Represents the first order gradient of sample i, h i The second order gradient of sample i is represented, and λ and γ are constants.
Because the sample data to be trained respectively have the first data party and the second data party, the profit value of the splitting node of the respective sample data under each splitting mode needs to be calculated on the first data party side and the second data party side respectively.
In this embodiment, since the first data side and the second data side are aligned in advance, both sides have the same gradient characteristics, and since the data tag exists in the sample data of the second data side, the benefit value of splitting the node in each splitting mode of both sides sample data is calculated based on the first-order gradient and the second-order gradient of the sample data of the second data side.
Step S102, if the round of node splitting is the first round of node splitting for constructing a regression tree, encrypting the first-order gradient and the second-order gradient, and then sending the first-order gradient and the second-order gradient to the first data side together with the sample ID of the sample set so as to calculate the benefit value of the split node of the training sample locally corresponding to the sample ID in each splitting mode based on the encrypted first-order gradient and the second-order gradient on the first data side;
in this embodiment, in order to realize bidirectional privacy protection of sample data of both parties in the federal training process, if the node splitting of the present round is first node splitting of constructing a regression tree, the first-order gradient and the second-order gradient of the sample data are obtained by calculation on the second data party side, then encrypted, and then sent to the first data party.
On the first data side, the profit value of the split node of the local sample data of the first data side is calculated based on the first-order gradient and the second-order gradient of the received sample data and the calculation formula of the profit value gain, and the profit value obtained by calculation is also an encryption value because the first-order gradient and the second-order gradient are encrypted, so that encryption of the profit value is not needed.
After the profit value of the split node under various splitting modes of the sample data is calculated, the split node can be split to generate a new node so as to construct a regression tree. The embodiment preferably uses the second data party with the sample data with the data label to master and construct the regression tree of the gradient lifting tree model. Thus, the benefit value of the splitting node in each splitting mode needs to be sent to the second data party by the first data party local sample data calculated at the first data party side.
Step S103, if the node splitting of the present round is a non-first round node splitting of constructing a regression tree, a sample ID of the sample set is sent to the first data side so as to be used for using a first-order gradient and a second-order gradient used for the first round node splitting on the first data side, and a benefit value of splitting nodes of a training sample locally corresponding to the sample ID in each splitting mode is calculated;
in this embodiment, if the present round of node splitting is a non-first round of node splitting for constructing the regression tree, only the sample ID of the sample set corresponding to the present round of node splitting is required to be sent to the first data party, and the first data party continues to use the first-order gradient and the second-order gradient used during the first round of node splitting, so as to calculate the benefit value of splitting the node of the training sample corresponding to the received sample ID locally in each splitting mode.
Step S104, the second data party receives the encryption income values of all the split nodes returned by the first data party and decrypts the encryption income values;
step S105, on the second data side, calculating a benefit value of splitting nodes of the training sample corresponding to the sample ID locally in each splitting mode based on the first-order gradient and the second-order gradient;
on the second data side, based on the first-order gradient and the second-order gradient of the sample data obtained by calculation and the calculation formula of the gain value gain, the gain value of the splitting node of the sample data to be trained locally on the second data side is calculated in each splitting mode.
Step S106, determining the global optimal splitting node of the splitting of the round of nodes based on the profit values of all the splitting nodes calculated by the two parties respectively;
since the initial sample data of the two parties are aligned by the samples, the profit value of all the split nodes calculated by the two parties can be regarded as the profit value of the split nodes in each split mode for the whole data sample of the two parties, and therefore, the split node with the largest profit value is taken as the global optimal split node for splitting the nodes of the round by comparing the sizes of the profit values.
It should be noted that, the sample feature corresponding to the global best split node may belong to the training sample of the first data party or the training sample of the second data party.
Optionally, since the regression tree construction of the gradient lifting tree model is dominated by the second data party, on the second data party side, the relevant information of the global optimal splitting node determined by each round of node splitting needs to be recorded; the related information includes: a provider of the corresponding sample data, a signature encoding of the corresponding sample data, and a benefit value.
For example, if the data party a holds the feature f corresponding to the global optimal partition point i This record is then (Site A, E A (f i ) Gain). Otherwise, if the data party B holds the feature f corresponding to the global optimal partition point i Then this record is (Site B, E B (f i ) Gain). Wherein E is A (f i ) Representing the characteristics f of the data party A i Coding E B (f i ) Representing the characteristic f of the data side B i Coding is carried out, by which the feature f can be marked i Without revealing its original characteristic data.
Optionally, when selecting the features in the foregoing embodiment, it is preferable to use each global optimal splitting node as a splitting node of each regression tree in the gradient lifting tree model, and calculate an average benefit value of the splitting nodes corresponding to the same feature code.
Step S107, splitting a sample set corresponding to the current node based on the global optimal splitting node of the current node splitting, and generating a new node to construct a regression tree of the gradient lifting tree model.
If the sample feature corresponding to the global optimal splitting node of the current round of node splitting belongs to the training sample of the first data party, the sample data corresponding to the current node of the current round of splitting belongs to the first data party. Correspondingly, if the sample feature corresponding to the global optimal splitting node of the current round of node splitting belongs to the training sample of the second data party, the sample data corresponding to the current node of the current round of splitting belongs to the second data party.
By node splitting, new nodes (left child node and right child node) can be generated, thereby constructing a regression tree. Through multiple rounds of node splitting, new nodes can be continuously generated, so that a regression tree with deeper tree depth is obtained, and if node splitting is stopped, a regression tree of the gradient lifting tree model can be obtained.
In this embodiment, since the data of the calculation communication of both parties are encrypted data of the model intermediate result, the training process will not leak the original feature data. And meanwhile, an encryption algorithm is used in the whole training process so as to ensure the privacy of data. Preferably, a partial homomorphic encryption algorithm is used to support the addition homomorphism.
Further, in an embodiment, based on the difference of the node splitting conditions, the first-order gradient and the second-order gradient of the training sample for node splitting are obtained specifically by:
1. the round of node splitting corresponds to the construction of a first regression tree
1.1, if the node splitting of the round is the first round node splitting for constructing a first regression tree, initializing a first-order gradient and a second-order gradient of each training sample in a sample set corresponding to the node splitting of the round at a second data side;
1.2, if the node splitting of the first round is the node splitting of the first round which is not the first round for constructing the regression tree, the first-order gradient and the second-order gradient used by the node splitting of the first round are used.
2. Non-initial regression tree is correspondingly constructed through node splitting of the round
2.1, if the node splitting of the round corresponds to the node splitting of the first round of constructing a non-first regression tree, updating a first-order gradient and a second-order gradient according to the federal training of the previous round;
2.2, if the node splitting of the round is the node splitting of the non-first round for constructing the non-first regression tree, the first-order gradient and the second-order gradient used by the node splitting of the first round are used.
Further, in one embodiment, to reduce the complexity of the regression tree, a depth threshold of the regression tree is preset for node splitting limitation.
In this embodiment, when each round of generating new nodes to construct a regression tree of the gradient lifting tree model, determining, on the second data side, whether the depth of the regression tree of the round reaches a preset depth threshold;
if the depth of the regression tree of the current round reaches a preset depth threshold, stopping node splitting, further obtaining a regression tree of the gradient lifting tree model, and if not, continuing the node splitting of the next round.
It should be noted that, the condition for limiting node splitting may also be stopping node splitting when the node cannot continue splitting, for example, if the node cannot continue splitting if the node corresponds to a sample corresponding to the current node.
Further, in another embodiment, to avoid overfitting during training, a number threshold of regression trees is preset to limit the number of regression trees generated.
In this embodiment, when node splitting is stopped, on the second data side, it is determined whether the total number of the present round of regression trees reaches a preset number threshold;
if the total number of the regression trees in the round reaches a preset number threshold, stopping the federal training, otherwise, continuing the next round of federal training.
It should be noted that the condition for limiting the number of regression trees to be generated may be to stop constructing the regression tree when the node cannot continue to split.
For a better understanding of the present invention, the federal training and modeling process of the present invention is illustrated below based on the sample data in tables 1 and 2 in the above examples.
First pass federal training: training a first regression tree
(1) First round node splitting
1.1 on the second data side, a first order gradient (g) of the sample data in Table 2 was calculated i ) And second order gradient (h i ) The method comprises the steps of carrying out a first treatment on the surface of the For g i And h i After encryption, the encrypted data is sent to a first data party;
1.2 on the first data side, based on g i And h i Calculating gain values gain of split nodes in all possible split modes of the sample data in the table 1; sending the benefit value gain to the second party;
since the Age feature in table 1 has 5 sample data division modes, the Gender feature has 2 sample data division modes, and the Amount of given credit feature has 5 sample data division modes, the sample data in table 1 has 12 split modes in total, that is, the profit value of the split node corresponding to the 12 split modes needs to be calculated.
1.3, on the second data side, calculating the gain value gain of the split node in all possible splitting modes of the sample data in the table 2;
because the Bill Payment feature in Table 2 has 5 sample data dividing modes and the effect feature has 3 sample data dividing modes, the sample data in Table 2 has 8 splitting modes in total, that is, the profit value of the splitting node corresponding to the 8 dividing modes needs to be calculated.
1.4, selecting a feature corresponding to the maximum profit value from profit values of split nodes corresponding to 12 division modes calculated from a first data side and profit values of split nodes corresponding to 8 division modes calculated from a second data side as a global optimal split node for splitting the node of the round;
and 1.5, splitting the sample data corresponding to the current node based on the global optimal splitting node of the current node splitting, and generating a new node to construct a regression tree of the gradient lifting tree model.
1.6, judging whether the depth of the round of regression tree reaches a preset depth threshold value; if the depth of the regression tree of the current round reaches a preset depth threshold, stopping node splitting, further obtaining a regression tree of the gradient lifting tree model, and if not, continuing the node splitting of the next round;
1.7, judging whether the total number of the round of regression trees reaches a preset number threshold value; if the total number of the regression trees in the round reaches a preset number threshold, stopping the federal training, otherwise, continuing the next round of federal training.
(2) Second, third-round node splitting
2.1, assuming that the feature corresponding to the splitting node of the previous round is Bill Payment less than or equal to 3102, the feature is taken as a splitting node (corresponding samples are X1, X2, X3, X4 and X5), two new splitting nodes are generated, wherein the left node corresponds to a sample set (X1 and X5) which is less than or equal to 3102, the right node corresponds to a sample set (X2, X3 and X4) which is greater than 3102, and the sample set (X1 and X5) and the sample set (X2, X3 and X4) are taken as new sample sets respectively to continue splitting of the second and third rounds of nodes so as to split the two new nodes respectively, and generate new nodes.
2.2 since the second and third rounds of node splitting belong to the same round of federal training, the sample gradient values used for the first round of node splitting continue to be used. Assuming that a split node of the present round corresponds to a feature Amount of given credit which is less than or equal to 200, the feature is taken as a split node (corresponding samples are X1 and X5), two new split nodes are generated, wherein a left node corresponds to a sample X5 which is less than or equal to 200, and a right node corresponds to a sample X1 which is greater than 200; likewise, another split node of the present round corresponds to a feature with Age less than or equal to 35, and that feature is taken as a split node (corresponding samples X2, X3, X4), resulting in two new split nodes, with the left node corresponding to samples X2, X3 less than or equal to 35 and the right node corresponding to sample X4 greater than 35. The specific implementation flow refers to the first round of node splitting process.
Second round of federal training: training a second regression tree
And 3.1, as the node splitting of the round belongs to the next round of federal training, updating the first-order gradient and the second-order gradient used in the previous round of federal training by the result of the previous round of federal training, continuing the node splitting of the second round of federal training to generate a new node to construct the next regression tree, and particularly realizing the construction process of the previous regression tree with reference to the flow.
3.2 as shown in fig. 4, in the above embodiment, after two rounds of federal training, the sample data in tables 1 and 2 generate two regression trees, where the first regression tree includes three split nodes, respectively: bill Payment is less than or equal to 3102, amount of given credit is less than or equal to 200, and Age is less than or equal to 35; the second regression tree includes two split nodes, one for each: bill Payment is less than or equal to 6787, gender+=1.
3.3, based on two regression trees of the gradient lifting tree model as shown in fig. 4, the average benefit value corresponding to the characteristics of the sample data: bill Payment is (gain1+gain4)/2; the reduction is 0; age is gain3; gender is gain5; amount of given credit is gain2.
The invention also provides a computer readable storage medium.
The computer readable storage medium of the present invention stores thereon a gradient-lifting tree-model building program which, when executed by a processor, implements the steps of the federally-trained gradient-lifting tree-model building method as described in any one of the embodiments above.
From the above description of the embodiments, it will be clear to those skilled in the art that the above-described embodiment method may be implemented by means of software plus a necessary general hardware platform, but of course may also be implemented by means of hardware, but in many cases the former is a preferred embodiment. Based on such understanding, the technical solution of the present invention may be embodied essentially or in a part contributing to the prior art in the form of a software product stored in a storage medium (e.g. ROM/RAM), comprising instructions for causing a terminal (which may be a mobile phone, a computer, a server or a network device, etc.) to perform the method according to the embodiments of the present invention.
While the embodiments of the present invention have been described above with reference to the drawings, the present invention is not limited to the above-described embodiments, which are merely illustrative and not restrictive, and many modifications may be made thereto by those of ordinary skill in the art without departing from the spirit of the present invention and the scope of the appended claims, which are to be accorded the full scope of the present invention as defined by the following description and drawings, or by any equivalent structures or equivalent flow changes, or by direct or indirect application to other relevant technical fields.

Claims (9)

1. A feature selection method based on federal training, characterized in that the feature selection method based on federal training comprises the steps of:
performing federal training on two aligned training samples by adopting an XGboost algorithm to construct a gradient lifting tree model, wherein the two aligned training samples are a first training sample and a second training sample respectively, the first training sample is provided by a first data party and stored in the first data party, the second training sample is provided by a second data party and stored in the second data party, the gradient lifting tree model comprises a plurality of regression trees, and one split node of each regression tree corresponds to one feature of the training sample;
Counting average profit values of split nodes corresponding to the same feature in the gradient lifting tree model, and taking the average profit values as scores of the corresponding features;
feature sorting is carried out based on the scores of the features, and sorting results are output for feature selection, wherein if features which do not correspond to the split nodes exist in the training sample, default scores are used for the features;
wherein, the performing federal training on the two aligned training samples by using the XGboost algorithm to construct the gradient lifting tree model includes:
on the second data side, acquiring a first-order gradient and a second-order gradient of each training sample in a sample set corresponding to the node splitting of the round;
if the first-round node splitting is the first-round node splitting for constructing a regression tree, encrypting the first-order gradient and the second-order gradient, and then sending the first-order gradient and the second-order gradient to the first data side together with the sample ID of the sample set so as to calculate the benefit value of the split node of the training sample locally corresponding to the sample ID in each splitting mode on the first data side based on the encrypted first-order gradient and the second-order gradient;
if the node splitting of the round is the non-first round node splitting for constructing the regression tree, a sample ID of the sample set is sent to the first data side so as to be used for calculating the benefit value of the split node of the training sample which locally corresponds to the sample ID in each splitting mode along the first-order gradient and the second-order gradient used for the first round node splitting at the first data side;
The second data party receives the encryption gain values of all the split nodes returned by the first data party and decrypts the encryption gain values;
on the second data side, calculating the benefit value of splitting nodes of a training sample corresponding to the sample ID locally in each splitting mode based on the first-order gradient and the second-order gradient;
based on the profit values of all the split nodes calculated by the two parties, determining the global optimal split node for splitting the node of the round;
based on the global optimal splitting node of the current round of node splitting, splitting is carried out on a sample set corresponding to the current node, and a new node is generated to construct a regression tree of the gradient lifting tree model.
2. The federally trained based feature selection method according to claim 1, wherein the first training sample attribute comprises a sample ID and a portion of sample features and the second training sample attribute comprises a sample ID, another portion of sample features and a data tag.
3. The federal training-based feature selection method according to claim 1, wherein before the step of obtaining, on the second data side, a first-order gradient and a second-order gradient of each training sample in the sample set corresponding to the node splitting of the present round, the method further comprises:
When node splitting is carried out, judging whether the node splitting of the round corresponds to constructing a first regression tree;
if the node splitting of the current round corresponds to the first regression tree construction, judging whether the node splitting of the current round is the first node splitting of the first regression tree construction;
if the node splitting of the round is the first round node splitting for constructing the first regression tree, initializing a first-order gradient and a second-order gradient of each training sample in the sample set corresponding to the node splitting of the round at the second data side; if the node splitting of the first round is the node splitting of the first round which is not the first round for constructing the regression tree, the first-order gradient and the second-order gradient used by the node splitting of the first round are used;
if the node splitting of the round corresponds to the construction of the non-first regression tree, judging whether the node splitting of the round is the first round node splitting of the construction of the non-first regression tree;
if the node splitting of the round is the first round node splitting for constructing a non-first regression tree, updating a first-order gradient and a second-order gradient according to the last round of federal training; if the node splitting of the first round is the node splitting of the first round, which constructs a regression tree of the first round, the first-order gradient and the second-order gradient used by the node splitting of the first round are used.
4. The federally trained feature selection method according to claim 1, wherein the federally trained feature selection method further comprises:
When generating new nodes to construct a regression tree of the gradient lifting tree model, judging whether the depth of the regression tree of the round reaches a preset depth threshold value or not at the second data side;
if the depth of the regression tree of the current round reaches the preset depth threshold, stopping node splitting to obtain a regression tree of the gradient lifting tree model, and if not, continuing the node splitting of the next round.
5. The federally trained feature selection method according to claim 4, wherein the federally trained feature selection method further comprises:
when node splitting is stopped, judging whether the total number of the round of regression trees reaches a preset number threshold value at the second data side;
if the total number of the regression trees in the round reaches the preset number threshold, stopping the federal training, otherwise, continuing the next round of federal training.
6. A federally training based feature selection method according to any of claims 1-5, further comprising:
on the second data side, recording the related information of the global optimal splitting node determined by each round of node splitting;
wherein the related information includes: a provider of the corresponding sample data, a signature encoding of the corresponding sample data, and a benefit value.
7. The federally trained feature selection method according to claim 6, wherein said counting average benefit values for split nodes corresponding to the same feature in the gradient-lifted tree model comprises:
and on the second data side, taking each global optimal splitting node as a splitting node of each regression tree in the gradient lifting tree model, and counting the average benefit value of the splitting nodes corresponding to the same feature codes.
8. A federally trained feature selection apparatus comprising a memory, a processor, and a feature selection program stored on the memory and executable on the processor, the feature selection program when executed by the processor implementing the steps of the federally trained feature selection method according to any of claims 1-7.
9. A computer readable storage medium, having stored thereon a feature selection program, which when executed by a processor, implements the steps of the federally trained feature selection method according to any of claims 1-7.
CN201810918867.3A 2018-08-10 2018-08-10 Gradient lifting tree model construction method and device based on federal training and storage medium Active CN109034398B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810918867.3A CN109034398B (en) 2018-08-10 2018-08-10 Gradient lifting tree model construction method and device based on federal training and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810918867.3A CN109034398B (en) 2018-08-10 2018-08-10 Gradient lifting tree model construction method and device based on federal training and storage medium

Publications (2)

Publication Number Publication Date
CN109034398A CN109034398A (en) 2018-12-18
CN109034398B true CN109034398B (en) 2023-09-12

Family

ID=64633061

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810918867.3A Active CN109034398B (en) 2018-08-10 2018-08-10 Gradient lifting tree model construction method and device based on federal training and storage medium

Country Status (1)

Country Link
CN (1) CN109034398B (en)

Families Citing this family (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109711556B (en) * 2018-12-24 2020-11-03 中国南方电网有限责任公司 Machine patrol data processing method and device, network-level server and provincial-level server
CN109492420B (en) 2018-12-28 2021-07-20 深圳前海微众银行股份有限公司 Model parameter training method, terminal, system and medium based on federal learning
CN109934179B (en) * 2019-03-18 2022-08-02 中南大学 Human body action recognition method based on automatic feature selection and integrated learning algorithm
CN110427969B (en) * 2019-07-01 2020-11-27 创新先进技术有限公司 Data processing method and device and electronic equipment
CN110297848B (en) * 2019-07-09 2024-02-23 深圳前海微众银行股份有限公司 Recommendation model training method, terminal and storage medium based on federal learning
CN110795603B (en) * 2019-10-29 2021-02-19 支付宝(杭州)信息技术有限公司 Prediction method and device based on tree model
CN110851786B (en) * 2019-11-14 2023-06-06 深圳前海微众银行股份有限公司 Inter-enterprise data interaction method, device, equipment and storage medium based on longitudinal federal learning
CN110990829B (en) * 2019-11-21 2021-09-28 支付宝(杭州)信息技术有限公司 Method, device and equipment for training GBDT model in trusted execution environment
CN111079939B (en) * 2019-11-28 2021-04-20 支付宝(杭州)信息技术有限公司 Machine learning model feature screening method and device based on data privacy protection
CN110941963A (en) * 2019-11-29 2020-03-31 福州大学 Text attribute viewpoint abstract generation method and system based on sentence emotion attributes
CN111178538B (en) * 2019-12-17 2023-08-15 杭州睿信数据科技有限公司 Federal learning method and device for vertical data
CN111178408B (en) * 2019-12-19 2023-06-20 中国科学院计算技术研究所 Health monitoring model construction method and system based on federal random forest learning
CN110968886B (en) * 2019-12-20 2022-12-02 支付宝(杭州)信息技术有限公司 Method and system for screening training samples of machine learning model
CN111368901A (en) * 2020-02-28 2020-07-03 深圳前海微众银行股份有限公司 Multi-party combined modeling method, device and medium based on federal learning
CN111340614B (en) * 2020-02-28 2021-05-18 深圳前海微众银行股份有限公司 Sample sampling method and device based on federal learning and readable storage medium
CN111507479B (en) * 2020-04-15 2021-08-10 深圳前海微众银行股份有限公司 Feature binning method, device, equipment and computer-readable storage medium
CN111523686B (en) * 2020-04-23 2021-08-03 支付宝(杭州)信息技术有限公司 Method and system for model joint training
CN111291417B (en) * 2020-05-09 2020-08-28 支付宝(杭州)信息技术有限公司 Method and device for protecting data privacy of multi-party combined training object recommendation model
CN111738359B (en) * 2020-07-24 2020-11-27 支付宝(杭州)信息技术有限公司 Two-party decision tree training method and system
CN113435537B (en) * 2021-07-16 2022-08-26 同盾控股有限公司 Cross-feature federated learning method and prediction method based on Soft GBDT
CN113723477B (en) * 2021-08-16 2024-04-30 同盾科技有限公司 Cross-feature federal abnormal data detection method based on isolated forest
CN113722987B (en) * 2021-08-16 2023-11-03 京东科技控股股份有限公司 Training method and device of federal learning model, electronic equipment and storage medium

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107704966A (en) * 2017-10-17 2018-02-16 华南理工大学 A kind of Energy Load forecasting system and method based on weather big data
CN107767183A (en) * 2017-10-31 2018-03-06 常州大学 Brand loyalty method of testing based on combination learning and profile point
CN107993139A (en) * 2017-11-15 2018-05-04 华融融通(北京)科技有限公司 A kind of anti-fake system of consumer finance based on dynamic regulation database and method
CN108021984A (en) * 2016-11-01 2018-05-11 第四范式(北京)技术有限公司 Determine the method and system of the feature importance of machine learning sample
CN108257105A (en) * 2018-01-29 2018-07-06 南华大学 A kind of light stream estimation for video image and denoising combination learning depth network model
CN108375808A (en) * 2018-03-12 2018-08-07 南京恩瑞特实业有限公司 Dense fog forecasting procedures of the NRIET based on machine learning

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20180089587A1 (en) * 2016-09-26 2018-03-29 Google Inc. Systems and Methods for Communication Efficient Distributed Mean Estimation

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108021984A (en) * 2016-11-01 2018-05-11 第四范式(北京)技术有限公司 Determine the method and system of the feature importance of machine learning sample
CN107704966A (en) * 2017-10-17 2018-02-16 华南理工大学 A kind of Energy Load forecasting system and method based on weather big data
CN107767183A (en) * 2017-10-31 2018-03-06 常州大学 Brand loyalty method of testing based on combination learning and profile point
CN107993139A (en) * 2017-11-15 2018-05-04 华融融通(北京)科技有限公司 A kind of anti-fake system of consumer finance based on dynamic regulation database and method
CN108257105A (en) * 2018-01-29 2018-07-06 南华大学 A kind of light stream estimation for video image and denoising combination learning depth network model
CN108375808A (en) * 2018-03-12 2018-08-07 南京恩瑞特实业有限公司 Dense fog forecasting procedures of the NRIET based on machine learning

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
Communication-efficient learning of deep networks from decentralized data;H. Brendan McMahan 等;《ARTIFICIAL INTELLIGENCE AND STATISTICS》;20171231;第54卷;第1-4节 *

Also Published As

Publication number Publication date
CN109034398A (en) 2018-12-18

Similar Documents

Publication Publication Date Title
CN109034398B (en) Gradient lifting tree model construction method and device based on federal training and storage medium
CN109165683B (en) Sample prediction method, device and storage medium based on federal training
CN109299728B (en) Sample joint prediction method, system and medium based on construction of gradient tree model
JP7095140B2 (en) Multi-model training methods and equipment based on feature extraction, electronic devices and media
EP3001332A1 (en) Target user determination method, device and network server
CN110309923B (en) Transverse federal learning method, device, equipment and computer storage medium
US20200294111A1 (en) Determining target user group
US8572019B2 (en) Reducing the dissimilarity between a first multivariate data set and a second multivariate data set
CN108763314A (en) A kind of interest recommends method, apparatus, server and storage medium
Verbraken et al. Profit optimizing customer churn prediction with Bayesian network classifiers
US9881345B2 (en) Evaluating an impact of a user's content utilized in a social network
CN103678672A (en) Method for recommending information
US9418119B2 (en) Method and system to determine a category score of a social network member
CN109446171B (en) Data processing method and device
CN111506823A (en) Information recommendation method and device and computer equipment
Celbiş A machine learning approach to rural entrepreneurship
CN111951052A (en) Method and device for acquiring potential customers based on knowledge graph
CN108932646B (en) User tag verification method and device based on operator and electronic equipment
US11468521B2 (en) Social media account filtering method and apparatus
US20190180193A1 (en) Accurate and interpretable rules for user segmentation
CN113535991A (en) Multimedia resource recommendation method and device, electronic equipment and storage medium
CN111353554B (en) Method and device for predicting missing user service attributes
CN109451334B (en) User portrait generation processing method and device and electronic equipment
CN112817563A (en) Target attribute configuration information determination method, computer device, and storage medium
CN110489660B (en) User economic condition portrait method of social media public data

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant