INTRODUCTION
As the Internet of Things evolves, more and more data is created at the edge nodes. This data may be gathered to construct various AI models and implement them in the actual world. Due to the fact that the classic distributed learning paradigm necessitates the collection of users' personal information for centralised computing, it might result in some privacy leakage issues. Communication is a difficulty that distributed learning must always overcome, with an excessive volume of data in transmission and a large number of communications being the most frequent issues.
To address the aforementioned issues, Google presented the federated learning (FL) framework with promising growth potential in 2016. A typical FL system consists of two components: a central server and training participants. Specifically, the users participating in FL and the model of the central service are the same models, and the central server will first broadcast the model parameters globally at the start of learning. When the communication time point is reached, all or part of the users transfer the model parameters (or gradients) to the central server for aggregation, then the updated model parameters are globally broadcasted again. The FL framework provides some protection for users' sensitive data.
Recent research has revealed, however, that sensitive user data can still be compromised through inference attacks on uploaded parameters [1]. Researchers have employed a number of privacy-preserving techniques, such as multiparty secure computation [2], homomorphic encryption [3], and differential privacy [4], to prevent hostile analysts from gaining access to sensitive user information. Differential privacy approaches are utilised more frequently in FL due to their lower computational and time requirements. Therefore, the analysis and scheme design in this paper are also based on federated learning with differential privacy.
Recent publications [5, 6] have utilised differential privacy to preserve FL and parallel SGD approaches for generating FL models. However, a more practical difficulty, the communication problem, is overlooked. Local SGD is a method described in refs. [7, 8] for reducing the number of communications, and it is theoretically proven that Local SGD has the same convergence speed as parallel SGD. In addition, the study in ref. [8] is based on IID data, whereas the analysis in this paper is based on Non-IID data. In addition, the FedAVG model is the central aggregation model employed in this paper, and it is frequently used because of its superior aggregation outcomes.
The data of each FL user can be regarded as a sample of the overall data, and the assured involvement of users can be regarded as a sufficient quantity of data. However, the aggregation effect of FL can make the model parameters different from local ones, resulting in the generalisation ability of the global model that is not necessarily suitable for local users, which can reduce the idea of user participation in federated learning. Moreover, users themselves incur costs by participating in federated learning (e.g. computation costs, communication costs, data costs etc. [9]) therefore, it is a reasonable idea to compensate users for the costs. To this goal, academics have created a variety of incentive systems, including game theory [10], auction theory [11], contract theory [12], and prospect theory, although the majority of these compensate users for the losses incurred by local processing and communication uploads. However, few studies discuss the costs involved with the potential privacy leaks that may result from inference attacks. Users have varying sensitivities to privacy leakage, which can be reflected in incentives as individual preferences for privacy protection vigour. Since the level of privacy protection is inversely related to the quantity of privacy leakage ɛ, users with more privacy leakage should be compensated more.
In this paper, we use a reverse auction model (also known as a purchasing model) to achieve incentives for users. When FL is viewed as a procurement problem, that is, when users are viewed as suppliers and central servers as purchasers, reverse auctions can effectively solve the problem of selecting the winner. In the context of reverse auctions, [13] used single-attribute auctions, but numerous theoretical and experimental studies [14–16] have demonstrated that multi-attribute auctions generate greater utility for purchasers than single-attribute auctions. Multi-attribute auctions allow suppliers to place bids based on attributes other than price, allowing purchasers to obtain high-quality goods or services at reduced costs from participating suppliers.
Negotiation is another answer to the procurement issue. In previous studies, negotiations and auctions have always been studied separately, for example, comparing economic efficiency, allocating efficiency and pareto optimality. Table 1 briefly describes each negotiation and auction have their own advantages and disadvantages. However, recent research [17] shows that combining auctions and negotiations will improve social welfare. Also, this is in line with the idea of the privacy paradox, where more private information can be obtained by providing a small incentive.
TABLE 1 Advantages and disadvantages of reverse auctions and negotiations.
Auction | Advantages | Disadvantages |
Auction | Greater selection of suppliers; | Not conducive to communication |
Effective avoidance of human | Between supply and demand; complicated | |
Factors affecting transactions | Design and difficult to implement. | |
Negotiations | Interactivity and efficiency; | More subject to human factors. |
Facilitates more knowledge | ||
About suppliers |
In summary, this paper presents the federated learning reverse auction and negotiation model that satisfies differential privacy requirements (FLRNDP).
This article contributes the following:
-
Theoretical analysis: This paper provides the privacy analysis, convergence analysis, and local iteration count analysis of federated learning in the context of combining differential privacy and Local SGD. The analysis provides a more intuitive demonstration of the impact of differential privacy on the performance of federated learning.
-
Auction model: In this paper, we propose a reverse auction model to motivate and select users considering the amount of data and the strength of privacy protection. Experiments show that the model we design has higher accuracy and model quality as well as lower loss compared to state-of-the-art algorithms.
-
Post-auction negotiations: We have conducted some analysis of post-auction negotiations, and the theory demonstrates that post-auction negotiations can improve social welfare, disguised as increased revenue for the central server from the central party's perspective.
The rest of the paper is organised as follows. Section 2 contains the literature review. Section 3 contains the basic theory. Section 4 contains the system framework. Section 5 contains some theoretical analysis to introduce Section 6. Section 6 contains our reverse auction. Section 7 contains post-auction negotiation model. Section 8 contains experimental analysis. Section 9 contains the conclusion.
RELATED WORK
Federated learning and differential privacy
Since the 2016 launch of the FL framework, scholars have done extensive FL-related research. McMahan et al. [18] proposed two aggregation methods for federated learning: FedAVG and FedSGD, among which FedAVG is widely used due to its simple expressions and better robustness to non-IID data. On this basis, Yu et al. [19] give a proof of convergence of federated learning in FedAVG as well as in the parallel SGD framework as a way to show why FedAVG is effective. However, since parallel SGD can impose excessive communication overhead on federated learning, Stich [8] proposed the use of Local SGD as an alternative to parallel SGD and theoretically proved that both have a convergence rate of , but their theory is only applicable under IID datasets. Li et al. [20] give a proof of Local SGD convergence in the case of non-IID dataset and shows that Local SGD still performs well in terms of convergence rate even in the non-IID case.
However, none of these articles consider the issue of privacy leakage during data transmission [1]. In 2020, Wei et al. [4] introduced differential privacy to federal learning to combat advanced inference attacks for the first time and performed a theoretical as well as numerical analysis of its performance. However, due to its lack of generalisability by assuming that all users use the same strength of privacy protection, Hu et al. [6] and Pandey et al. [10] used federated learning for personalised privacy protection. However, both of them are too lenient for privacy leakage measurements, which leads to over-measurement of privacy leakage. At the same time, they all use parallel SGD, which imposes a significant burden on communication.
In order to measure privacy leakage accurately, [21] proposed a generalised method for measuring privacy leakage caused during machine learning, MA, which measures privacy leakage during training by means of higher-order moments. Mironov [22] implemented a tighter upper bound than MA, Rényi differential privacy, by introducing Rényi entropy on the idea of Abadi [21]. Currently, RDP has generally replaced MA with more machine learning privacy preserving.
The federated learning framework in this paper uses a combination of Local SGD and RDP to both reduce the communication overhead and provide a more accurate measure of privacy leakage.
Federated learning and incentives
Inside FL, although users can get better models by participating in FL, we believe that users are rational and may be reluctant to participate in FL in the face of cost and resource consumption. Thus, the study of FL and incentive mechanism arises. Shapley value [23] is a measure of user contribution in cooperative games, and it is widely used in FL to calculate contribution, such as in refs. [24–26]. However, its excessive time complexity leads to the need to choose a certain approximation in practice, Zelei et al. [27] heuristically eliminate the lower Shapley values, thus reducing the computing time, but this method is too simple and brutal. Zhenan et al. [28] can effectively estimate the Shapley value by using a compressed perception method. Due to the high time complexity of the direct calculation method, [29, 30] used the reputation value method to indirectly measure user contributions and combined the reputation value with the blockchain to effectively achieve reputation value protection.
Since users are always assumed to be rational in the incentive mechanism, it is possible for them to choose to misrepresent costs or parameters as a way to increase their revenue. Pejó et al. [31] used Nash equilibrium by modelling FL as a noncooperative game to ensure that the data submitted by users do not shift NE points. However, the actual scenario is always in a non-complete information environment and Weng et al. [32] solves the non-cooperative game in FL by using BNE. In addition to using traditional game theory, Wei et al. [33] also used contract theory to address the selection of users involved in FL.
However, none of the aforementioned articles have gone into the discussion of compensating for privacy breaches (due to differential privacy protection), Sun et al. [34] using contract theory for privacy breaches, which not only ensures that users do not misreport, but also ensures that higher gains are obtained for larger degrees of privacy breaches. However, not all users have individualised privacy protection strengths. Zhang et al. [35] used the VCG mechanism to select users with different privacy-preserving strengths, and also explored the difference between local and global models in a non-IID environment. However, both assume the same amount of user data, our framework relaxes this assumption and the privacy protection strength of users is personalised, and more importantly, we use multiple attributes. In addition, we propose post-auction negotiation and perform a theoretical analysis.
PRELIMINARIES
In this section, we introduce FL, DP, Multi-attitude Reverse Auction, and Negotiation basics.
Federated learning
Suppose there exists a set , where is the total number of users in the FL system. Where each user has his own private training dataset containing N
k
samples, with and respectively denoting the feature vector and the corresponding label of the n − th training sample at worker k. In theory, the FL problem is usually described as the problem of how to minimise the expected risk, that is, minimising the following expression.
However, the actual situation is predominantly characterised by Non-IID user data, and this data heterogeneity causes a number of problems, such as users having different P K (.) and having different prediction goals and result preferences [35]. This renders the exact calculation of expected risk minimisation impossible, and because the data distribution of each user is unknown, as well as for computational convenience, empirical risk minimisation is now the most common method for risk minimisation.
Empirical risk minimisation
Given that P
K
(.) is unknown for each user, Equation (1) is approximated by summing and averaging the losses obtained after each mini-batch. Thus, FL seeks to solve subordinate issues in the real situations.
To solve the problem in Equation (3), Zheng et al. [36] use the mini-batch parallel SGD method, where the user is trained locally once to communicate with the server, to achieve the desired learning accuracy. This creates a significant communication issue, which is partially resolved by the Local SGD method described in this paper. While maintaining convergence efficiency comparable to that of parallel mini-batch SGD.
Specifically, we assume that the set of communication times between the client and the server is , where E represents the user's local Epoch count and nE represents the nE times the user has run the Epoch at the nth round of communication, it can also be called global synchronisation steps, and all times outside of communication are Local SGD. Let be the t − th user, and if , that is, this moment is the global update moment, then FedAVG will collect the model parameters of the user and aggregate them at the central server, then the FedAVG update at this point will be described as
At the same time, since the data analysed in this paper is Non-IID data, Non-IID should be measured. The measurement used in this paper is described in ref. [20]:
Quantifying the degree of non-IID (heterogeneity)
Let and be the minimum values of and respectively. We use the term for quantifying the degree of Non-IID. If the data are IID, then Γ obviously goes to zero as the number of samples grows. If the data are Non-IID, then Γ is non-zero, and its magnitude reflects the heterogeneity of the data distribution.
Differential privacy
Differential privacy is a straightforward and effective mathematical tool that defends against advanced inference attacks by employing a unique noise addition mechanism to make the newly generated perturbed data statistically equivalent to the original dataset. In this paper's differential privacy FL model, noise is added to the gradients if to ensure that the final training results satisfy differential privacy. If , the user uploads the model parameters to the central server for model aggregation. In this paper, we measure privacy leakage using Rényi differential privacy.
Definition
((α, ɛ)-differential privacy [22]) A randomised mechanism is said to have ɛ-Rényi differential privacy of order α, or (α, ɛ)-RDP for short, if for any adjacent it holds that
RDP is a relaxed version of classic (ɛ, 0)-differential privacy. The gap between two different distributions is measured by using the Rényi divergence. When α = 1, is the Kullback–Leibler divergence, that is, .
To ensure that the process of machine learning satisfies differential privacy in general, the usual way used is to add Gaussian noise to the gradient, that is, , where n ∼ N (0, σ 2 I d ). According to Proposition 7 in ref. [22], the adjacent dataset with Gaussian noise added is measured using the Rayleigh divergence, resulting in , where μ is the sensitivity, It is calculated as , where f (.) is the query function.
Reverse auction
Unlike single-source reverse auctions, FL should be viewed as multi-source, that is, multiple users need to be identified to participate in the training task of FL. Thus the task can be transformed into a winner determination problem. For price submission, the classical sealed submission is used and all users are assumed to be rational. Also, the auctions are designed to satisfy the following definitions to ensure that users are willing to participate in the auction and that the data submitted is real.
Definition
(Budget Feasibility [BF]) A Multi-Attribute Reverse Auction is budget feasible, if and only if the sum of payments received by each user does not exceed the total budget B, that is,
Definition
(Individual Rationality (IR)) In a reverse auction, the user receives a payment that should be non-negative after subtracting the cost, that is,
In order to ensure that the information submitted by the user is truthful, we need to introduce the Myerson Lemma.
lemma
(Myerson Lemma [13]): If the auction is truthful, the following two conditions need to be satisfied:
-
Selection rules are monotonous: If a candidate wins by bid price C, then it can also win when he bids C′ < C.
-
The payment to participants is the critical value: If the candidate's bid price is greater than this critical value, it is unlikely that it will win the bid.
Negotiation
In this paper, we convert the negotiation model to an ultimatum game, in part because of its strategic simplicity, using the Rubinstein bargaining model.
Definition
(purchaser and Supplier Fractions of Supply Chain Surplus [37]:) Suppose that player 1 and player 2 in the bargaining game have fixed discount factors δ 1, δ 2 and complete information is available to both purchasers and suppliers, then there is a unique Nash equilibrium. In equilibrium, the player 1 to offer a negotiated price and quality pair will gain a fraction of the supply chain surplus of . The player 2 to make an offer will gain a fraction of the surplus of .
SYSTEM FRAMEWORK
The FL system analysed in this paper consists of a trustworthy but inquisitive central server, which wants to collect user parameters in order to create a global model, and a set of users . Assume that each worker possesses a private training dataset D k consisting of N k data samples, with each data sample containing the feature vector and its corresponding label. The server's objective is to recruit a specific number of users for FL training (where successful participants must be selected through reverse auctions to ensure high accuracy of each training) without requiring their private information. The FLRNDP framework is depicted in Figure 1. The following describes the specific workflow for each round of communication. The training concludes when a predetermined number of communication rounds or a global model has converged to a particular level.
-
The central server will send the parameter w 0 to each user who is eligible to participate in the training.
-
The central server provides the auctioneer with its own budget B and evaluation function S(.).
-
Users submit their cost C i and non-economic attributes Q i to the auctioneer using a seal.
-
The auctioneer uses a reverse auction to determine the user k who participated in the FL and the corresponding payment P k .
-
The central server delegates multiple agents to negotiate with the users after identifying those who will participate in the training.
-
The user performs local training using the re-agreed and uploads the training parameters to the central server at the time of communication following the negotiation.
-
The central server aggregates the parameters of user uploads as w t+1.
-
The new parameters are distributed to all users by the central server.
[IMAGE OMITTED. SEE PDF]
THEORETICAL ANALYSIS: PRIVACY AND CONVERGENCE ANALYSIS
In this section, we will analyse the convergence of FL after adding differential privacy. We will show the convergence when the data is Non-IID. When the data is IID, we only need to set Γ = 0.
Privacy analysis
In this part, we leverage the notion of RDP to quantify each user's volume of privacy leakage. To ensure that the training results are all satisfy the differential privacy protection, the current practice is to add Gaussian noise to the gradient. The form is
In this paper, we assume that the amount of privacy leakage is unique to each user, whereas the amount of Gaussian noise added per round is the same for all users. Since the essence of differential privacy is the privacy leakage brought by the large size of the measurement noise, a larger σ will make the Gaussian distribution tend to be uniformly distributed, making it more difficult to distinguish the noise-added distribution from the original distribution, thereby reducing the privacy leakage ɛ.
Combining Proposition 7 from ref. [22] with our description of the amount of privacy leakage, we provide a theorem for the additional noise for a given amount of privacy leakage.
Theorem
In each round of FL, user k perturbs the gradient using Gaussian noise to satisfy Rényi differential privacy with , and satisfies .
Where ɛ k is the intensity of privacy protection for each user, and c is the gradient clipping threshold, we write it as such to distinguish it from the user cost of the auction phase. The proof for Theorem 1 is included in Appendix A.
From the combination of Equation (7) as well as Theorem 1, a smaller privacy leak results in a larger gradient perturbation effect, where n
k
is a random variable, that is, . Notice that the aggregation model used for our FL is FedAVG, that is, , which is converted into gradient descent form as
By abbreviating to p
k
, the aggregated noise is
Combining Equations (9) and (10), it is clear that excessive noise will cause the model parameters to deviate significantly from their initial values. As the training continues and the model gradually converges, the excessive noise will cause the model to deviate too far from the direction of convergence, resulting in a reduction in accuracy.
Convergence analysis
This subsection will provide convergence proofs for the differential privacy federation learning framework presented in this paper when the data is non-IID. And convergence is demonstrated when all devices participate, that is, when devices are involved.
First, we make the following assumptions about loss function F 1, F 2, …, F N for each user:
Assumption
F 1, F 2, …, F N are all L-smooth: for all y and x, .
Assumption
F 1, F 2, …, F N are all μ -strongly convex: for all y and x, .
The following two assumptions have been made by the works [19].
Assumption
let be sampled from the t − th device's local data uniformly at random. The variance of stochastic gradients in each device is bound: , for .
Assumption
The expected squared norm of stochastic gradients is uniformly bounded, that is, , for all and t = 0, …, T − 1.
Before proving convergence, we need some additional descriptions as well as lemmas. Since noise is added to the gradient during training, therefore, we define and , where . In the meantime, we make and , where .
lemma
(Bound of one step SGD). If we have
It is clear from Lemma 1 that the inclusion of differential privacy increases the upper bound.
lemma
(Bounding the variance). The variance of the aggregated gradient obtained by random sampling of the data and full sampling are as follows:
lemma
(Bounding the divergence of ) In the case where η t is non-increasing and η t ≤ 2η t + E for all t ≥ 0.
It follows that
It can be seen from lemma 3, if E is large, the difference between the local model parameters and the aggregated model parameters will become excessively large, resulting in a poorer fit of the model parameters to the local data after aggregation. But E is small, the aggregated model parameter will be skewed like a user with a larger amount of data |D k |, eventually forming a local optimal solution. This is illogical in the case of Non-IID data, as F 1, …, F N may be vastly distinct from F.
With lemma 1, 2, and 3, we can describe the convergence of the differential privacy FL framework used in this paper.
Theorem
Let Assumptions 1 to 4 hold and choose , γ = max{8
κ, E} and the learning rate . The expected upper bound of the difference between the loss after T-wheel and the optimal loss is:
From Theorem 2, first of all, we can see that the convergence rate of Local SGD after adding differential privacy is the same as that of parallel SGD, both being . Secondly, since each user adds different noise to the gradient during training, the aggregated noise . From the analysis, it can be seen that reducing noise will effectively increase the convergence rate as well as the accuracy.
Since the exact size of E cannot be given, we try to give the relationship between T and E to illustrate the theoretical value of E. Assume that FL stops when , let T
ɛ* be the moment when convergence is satisfied. Then
In the next section, we will improve the model accuracy by reducing the aggregation noise with the incentive model we designed. At the same time, users will be willing to participate in the training of the FL when they get the benefit, so as to ensure the participation.
MULTI-ATTRIBUTE REVERSE AUCTION DESIGN
In this section, we describe in detail the FLRNDP framework's multi-attribute reverse auction model. First, we will describe the design goals, followed by a formal description of the user selection process. As this section represents the auction portion of the overall model, we abbreviate it as FLRNDP-A.
Design goals
According to the summary in the previous section, in order to encourage users to participate in learning, it is necessary to select a privacy budget of ɛ and a data volume of |D| in order to improve the accuracy of the final training. We view the FL issue as a procurement issue, for which reverse auctions are currently more prevalent.
To minimise the variance of the distribution of the random variable while selecting the maximum number of users. We can transform the procurement problem into an optimisation problem and solve it using mathematical programming, as outlined below,
Of which Equation (17b) is to satisfy Definition 2, where to select and compensate users who participate in federal learning on a limited budget Equation (17c) is to satisfy Definition 3, and Equation (17d) is to guarantee the assignment. When x
k
= 1, this user is selected, and vice versa, it is not selected. After we use this function to solve the optimisation problem, we can find that maybe more than one minimum value point can reach the minimum value:
In this function, is a set when there are more than one minimum value point and a minimum value point when there is only one point. As this function indicate, we can get the set of minimum value point which consist of that satisfy the object function to reach the minimum value.
However, since FL still requires a large amount of data to prevent overfitting, we need to maximise the cardinality of , so we use the next minimum problem to get the optimisation
But so far, this programming expression is still not concise enough. First, the expression of is . It can be seen that and ɛ is the most straightforward to obtain. Secondly, it can be seen from Equation (17a) that is a fixed value for the selected user in the case of user selection, and the magnitude of the user's share in the overall aggregated noise is only determined by the amount of the user's own data. So, we can abbreviate as . Therefore, .
After the above simplification, the expressions for aggregation noise of Equation (17a) can be transformed to
In this subsection, we have described the goals by going through the mathematical language, but the solution to this programming is difficult to obtain, and we will make some relaxations to the original problem in the next subsection and make some changes in order to satisfy the truthfulness.
Multi-attribute reverse auction model
In the previous section, we described design goals, but this is only the ideal state. Since our auction is based on FL, we also need to consider the amount of data, and the only way to increase the amount of data is to increase the number of users. Therefore, it is a contradictory problem and we cannot get an optimal solution. And to solve this problem, we put Equation (19) into the original problem as a penalty term, that is,
Think of the central server as the purchaser and the users involved in the learning as the suppliers. He faces K bidders, expressed as a set . At the beginning of the auction, the purchaser will specify the attributes to be submitted by the bidder, which include cost C, attribute q i (i = 1, 2, …, m). The purchaser will convert the attributes into a scoring function S(.): R m → R. From the analysis in the previous subsection, our reverse auction model has two attributes, so the supplier submits the attribute they need to submit to the purchaser by submitting a sealed bid. Let , denoting the set of all supplier submitted attribute tuples received by the purchaser.
For the winner determination problem (WDP) in multi-attribute auctions, by converting multiple attributes into weighted product (WP) [38]. Therefore our first idea is to treat Equation (18) as the evaluation function needed for the WDP solution, that is,
However, this evaluation function design solution is problematic. If there is a user k whose |D k | and ɛ k are greater, it is not guaranteed to be chosen because S(.) is not monotonic. In addition, since both data size |D| and privacy leakage ɛ affect training accuracy, it is a less correct idea to simply consider the ratio size, and it is a better solution to adjust the weight of data size and privacy leakage and to achieve monotonicity of S(.) by using WP. Following is an explanation of how to construct the evaluation function S(.).
First, the tuple of attributes uploaded by the user is constructed as a matrix, that is,
The changes in the aforementioned attributes and the determination of the scoring function are primarily due to the fact that |D| is significantly greater than ɛ, causing the determinants of the evaluation function to be excessively skewed towards |D|. Also, α and β are weight values that indicate which of ɛ and |D|‘s attributes is somewhat more significant in FL. The optimal amounts of α and β will be determined through experimentation.
From the above description, the purpose of our reverse auction is to obtain the maximum with the minimum payment. Therefore, we can convert Equation (21) into
however, up to now, we have not determined the value of P, but according to the constraint P
k
− C
k
≥ 0, P must be at least greater than the cost C, so we can replace Equation (27a) to
The reason for using here is to prevent different cost functions from having an impact on the follow-up. Now, our optimisation becomes both simple and feasible. Let , we call ρ′ the bid price per unit of mass, sorting incrementally to get
After sorting, we will select candidates from users to participate in the FL, where satisfies,
Algorithm
FLRNDP-A model algorithm
Therefore, the final selected supplier receives payment of
After giving the user selection methods described above, comparing Equations (27a) and (30) to see that there are cases where the number of users selected exceeds the optimal value, for which we still recommend selecting more users as a way to reduce the possibility of underfitting due to sample imbalance.
After giving a full description of the auction, we'll use a few theorems to show that our method for choosing the winner is IR, CR, and truthful.
Theorem
FLRNDP-A is individually rational.
Theorem
FLRNDP-A has budget feasibility.
Theorem
FLRNDP-A is truthful.
The proof procedure of the above theorem is given in Appendix C.
Algorithm 1 describes the process of user selection, while, for a more visual representation of the flow of our framework, we also give Figure 2.
In the next section, we negotiate the users selected from the auction with an incentive to improve the data quality as a way to further improve the accuracy of the aggregation model.
[IMAGE OMITTED. SEE PDF]
POST-AUCTION NEGOTIATION MODEL
As shown in Table 1, since the auction model does not permit purchasers and suppliers to communicate effectively as in negotiations, and with a predetermined amount of user data, we believe that the quality of the data that can be negotiated is related to the endogenous factor ɛ, which is proportional to the number of bonuses that the purchaser can offer. This paper's post-auction negotiation model focuses on reducing the privacy protection intensity of users through additional incentives, that is, by increasing ɛ k . Figure 3 represents the general process of Post-Auction Negotiation Model.
[IMAGE OMITTED. SEE PDF]
We believe the amount of change in privacy leakage as a result of bonuses should be as follows:
In this section, we treat the purchaser as having agents to negotiate with the purchaser separately. Thus the negotiation can be considered as a two-player game.
First, we express the utility functions of the purchaser and the supplier in the auction phase.
We assume that the utility of the purchaser in reverse auction is
Since the change in the intensity of supplier privacy protection is personal private information, the bonus to be paid to the purchaser during the phase of negotiation should be an expectation bonus. After incentivising users to increase the amount of privacy leakage, the purchaser's utility changes to
In addition, the suppliers' compromised privacy to lessen the intensity of privacy protection can be viewed as an additional cost to the suppliers' utility.
Given the utility of purchasers and suppliers, we use social welfare to measure the overall utility
Bargaining phase
After describing some of the basics, the negotiations are described below.
During this phase, the purchaser and the winning supplier k can negotiate the quality requirement, , and the corresponding payment to the supplier, . Consider δ
1 and δ
2 to be the discount factors for the purchaser and the suppliers respectively. The longer the duration of negotiations, the more likely it is that the discount rate between the parties will diminish the value of the deal. According to ref. [17], the profit following multiple rounds of negotiation can be expressed as
With the above equation, purchaser-supplier negotiations eventually reach a unique Nash equilibrium [37], and Theorem 6 describes the benefits to both parties as well as the payments after the Nash equilibrium is reached:
Theorem
(Unique Equilibrium for Price after Negotiation) If the purchaser and suppliers decide to negotiate after the auction, the purchaser will first make a new quality demand, then the parties will reach a unique transaction price P Nash equilibrium in a round of negotiation with the value:
In the next subsection, we will describe the bonuses that the purchaser will pay to the supplier.
Bonus to supplier
The after-negotiation bonus may be viewed as a cost to the purchaser in order to incentivise the supplier to alter quality. In a manner resembling the auction phase, purchasers always seek to minimise expenditures in order to maximise return. In contrast, an expectation-related expression is obtained when ɛ′ ∼ P (ɛ∣bonus) holds true. Still using user k as an example, we can express the bonuses as follows,
The first formula in the constraint conditions employs the incentive compatibility theorem to validate the user's relaxed privacy protection intensity value. The second formulation ensures the user's rationality so that their participation in the negotiation is profitable.
As you can see from Equation (38), bonus should be:
At the end of this section, we propose Theorem 7. And the proof of this theorem is presented in Appendix C.
Theorem
(Social Welfare Improvement) Negotiations between the two players after an auction improve social welfare relative to auction only.
EXPERIMENT
Our experiments consider an FL system consisting of a central server and K = 100 users. We assume a random selection of privacy leakages for K = 100 users from (3, 10). Before starting FL, we assume that the server has a fixed budget B to motivate workers to perform local model updates. Our budget is B = 100,000.
For the learning configuration, unless otherwise stated, a simultaneous FL scheme is used in this paper. The mini-batch for each user is set to 32 mini-batch data samples. For privacy considerations, the noise variance is determined by the degree of personal privacy disclosure. We use SGD as the optimiser and set the initial learning rate η = 0.1, while setting a simple learning rate decrease (10 rounds of communication). Also, our cost function is set to C(θ; |D|, ɛ) = θ(|D| + ɛ), which θ is users' private information. In the experiment, we randomly draw the values of user θ from a uniform distribution U (0.5, 1). We do not restrict the choice of the cost function, because our algorithm already normalises the cost. Next, we describe the dataset and model used in the experiments, and the baseline for performance comparison.
-
Dataset: We utilise the MNIST and Fashion-MNIST datasets. In this paper, only the non-IID case is considered. Specifically, for the non-IID scenario, we use Fedlab [39] to perform an unbalanced partitioning of the data, resulting in a different amount of training data for each user, whose classes are not necessarily identical. Figure 4 demonstrates that we used Fedlab to divide the dataset for 100 users and used the data of 20 users to create the image, demonstrating that unbalanced division in Fedlab can ensure that the amount and type of data for each user is distinct.
-
Model: MLP: We used the most basic MLP model, that is, a three-layer structure with only one hidden layer, where the genus-out function of the hidden layer uses the sigmoid function.
CNN: It consists of two 5 × 5 convolutional layers (the first 16 channels and the second 32 channels, each 2 × 2 max pooling), both Re-Lu activation layers, and a 10-cell fully connected layer.
Comparison algorithm and performance metrics
We compared the algorithm with two 21-year, [34, 35]. Where Sun et al. [34] use contract theory for user selection and Zhang et al. [35] use VCG auctions for user selection while ensuring truthfulness.
In the performance metrics, we give the accuracy comparison, loss comparison, F1-score of our algorithm FLRNDP-A with DP-FFL [35], Pain-FL-C [34]. And additionally, we give the confusion matrix of our algorithm to visualise the accuracy of our FLRNDP-A (Figures 4 and 5).
[IMAGE OMITTED. SEE PDF]
[IMAGE OMITTED. SEE PDF]
Choice of E
As seen in Section 5.2, the magnitude of the E value affects the FL's precision, and exact values cannot be obtained. In this subsection, we will determine experimentally the optimal value of E. We conducted federation learning with the same configuration at E = 20, 30, 40, and 50, where the privacy leakage of user k was controlled under identical conditions to the initial random selection, and is involved in training at the same time. Figure 5 demonstrates that the model accuracy is greatest when E = 30 for two distinct models and data sets respectively. In subsequent comparison experiments, E is fixed at 30.
Choice of α
As can be seen from Section 6.2, S(.) corresponding to each user is controlled by ɛ and |D|. These two parameters control the amount of privacy leakage and the importance of the amount of data. Since FL with differential privacy requires both higher privacy leakage and higher data volume, the value of α cannot be obtained as an explicit expression, so we use an experimental approach to try to find a parameter value that makes the learning model more accurate, in which we use a budget B = 100,000 while fixing selected users, and the results show that when α = 0.6 and β = 0.4, the highest accuracy is achieved (Figure 6).
Comparison
As can be seen from Figure 7, our algorithm always fluctuates more than FP-FFL and Pain-FL-C. We think this is a normal situation because FP-FFL and Pain-FL-C are mainly selected for ɛ, because the magnitude of C is proportional to the perturbation effect on the model, and our algorithm considers not only ɛ, but also |D|. It can be seen that FLRNDP-A often fluctuates in the early stage, which we consider as a good performance because a slightly higher ɛ will help the model jump out of the local optimum and thus bring better accuracy performance, which, of course, brings fluctuations in the late stage. Although there were fluctuations in the early period, it can be seen that our algorithm finally outperformed the comparison algorithm due to our relatively higher amount of data in the selection of users as a way to reduce the risk of overfitting and to learn more features in the case of the non-IID dataset.
[IMAGE OMITTED. SEE PDF]
[IMAGE OMITTED. SEE PDF]
It can also be seen from Figure 8 that our algorithm has some jitteriness in the pre-training period, which we attribute to the inclusion of differential privacy that makes the model slightly more volatile in the pre-training period. In the late training period, the loss of our algorithm is generally lower than that of the comparison algorithm.
[IMAGE OMITTED. SEE PDF]
In addition to accuracy and loss, we introduced F1-score to evaluate the model quality. Although accuracy is often used as a measure of the model, it has the limitation that it does not reasonably reflect the predictive power of the model when the sample is unbalanced. Therefore, we introduced the F1-score. The F1-score is a weighted summed average of Precision and Recall, which is written in the binary classification problem as , where P is Precision, R is Recall. As can be seen in Figure 9, the F1-score values of our model are slightly higher than those of the two comparison algorithms. And we find that our model has fewer fluctuations in F1-score, which indicates that the quality of our model is gradually improved during the training process.
[IMAGE OMITTED. SEE PDF]
Finally, we give the multiclassification confusion matrix obtained by our algorithm using the MLP model under MNIST and FashionMNIST datasets to visually represent the accuracy of our model, Figure 10, where the main diagonal line represents the correct classification.
[IMAGE OMITTED. SEE PDF]
CONCLUSION
In this paper, we analyse the privacy and convergence of FL with differential privacy using Local SGD, concluding that its convergence is related to aggregation noise. For user selection, we employ a multi-attribute reverse auction method to reduce aggregation noise. After the user is selected, further privacy leakage is negotiated with the user, and it is demonstrated that post-auction negotiation increases social welfare. We conducted experiments on the number of local epochs E and selected the better E, followed by experiments on the selection of α. Experiments reveal that the model is most accurate when E = 30 and α = 0.6. On this basis, we compare it with two state-of-the-art algorithms. Experiments show that our method outperforms the other two algorithms in terms of accuracy and model quality.
Our future work intends to focus on the combination of fairness and incentives, because FL is easily brought to another local optimum by users with high contributions without adding fairness. The research on combining fairness and incentives is currently in its infancy, and there will be a lot of interesting work to investigate.
ACKNOWLEDGEMENTS
National Natural Science Foundation of China, Grant Number: 62062020; National Natural Science Foundation of China, Grant Number: 72161005; Technology Foundation of Guizhou Province, Grant Number: QianKeHeJiChu-ZK [2022]-General184.
CONFLICT OF INTEREST STATEMENT
The author declares that there is no conflict of interest.
DATA AVAILABILITY STATEMENT
The data that support the findings of this study are available from the corresponding author, Hongqin Lyu, upon reasonable request.
APPENDIX - A
Assuming that ξ
k
and are adjacent mini-batch data that differ only at the j − th sample, the global sensitivity of the gradient is
APPENDIX - B
Proof Of Lemma 1
Since , therefore EB = 0. Then, we now shift our target to the boundary of A, expand A as follows,
The above inequality can be transformed from ‖x + y‖2 ≤ ‖x + y‖2 + ‖x − y‖2 = 2‖x‖2 + 2‖y‖2 to
From the L-smoothness of F
k
(.), it follows that
Therefore, Equation (B2) is
Now, we bound A
1:
By Cauchy-Schwarz inequality and AM-GM inequality to A
11, we have
Notice the μ-strong convexity of F
k
(.), we bound A
12 and A
13
Combining Equations (B3)–(B8) gives
We next aim to bound C, We define . Since . C can be expressed as
Noting that
Therefore,
-
η t L − 1 ≤ 0
-
-
Combining Equations (B9) and (B14), we get that the final bound of A is
Notice that since , in Lemma 1 is erased.
Proof Of Lemma 2
The reason that the noise is for both the sample and the full sample in the above expression is that for the same user, the noise values are from the same Gaussian distribution, and the tensor shape composed of the gradient is fixed, so can be used to represent it.
Thus, Equation (B16) reduces to
Proof of Lemma 3
Since we use a model that performs FedAVG aggregation every E steps. Therefore, for any t ≥ 0 there exists a t
0 ≤ t, such that t − t
0 ≤ E − 1 and . Also, we use that η
t
is non-increasing and . Thus,
Proof Of Theorem 2
Since the noise addition process is performed during training, either or , there is always . We let and, combining Lemma 1, 2 and 3, get
For a diminishing stepsize, we let . Then, we will prove .
Using mathematical induction, assume the conclusion holds for some t, it follows that
Due to the strong convexity of F (.),
Specifically, if we choose and denote , then
End of proof.
APPENDIX - C
Proof Of Theorem 3
If user k of the candidate set is selected, then its utility is . If user k is unchecked, its utility is 0. Now we focus only on the selected user, assumed to be the k − th. From Equation (28), the payment received by user k is .
According to
Proof Of Theorem 4
From Equation (27), is computed for each user selected by our auction model. The selection stops when the total payment exceeds B, so our auction payment is
Proof Of Theorem 5
According lemma 1, we know what conditions need to be met if the auction is to be truthful. We know what conditions need to be met if the auction is to be truthful.
First, we prove that the selection rule is monotonic. Suppose user k is selected, then his . If user k decreases his bid, he will likely be ahead of the previous one, at least unchanged, if the other users' bids remain the same.
Second, we come to prove that participants receive a critical value of the bonus. Suppose that the first k users are selected to participate in FL through an auction, which contains a user j whose received payment is P j . In the following, we will describe two possible changes:
-
If user j's bid , then its ranking position will remain at least the same, while the revenue will not change.
-
If user j's bid , then its position will move backwards, resulting in not being selected as a user to participate in the training, resulting in 0 gain for itself.
In summary, the payment P is the critical value. By the lemma 1, it is clear that FLRNDP-A has truthfulness.
APPENDIX - D
Proof Of Theorem 6
Based on Definition 4, the purchaser's profit will be of the social welfare and the supplier's profit will be of the social welfare.
From Equation (34), the gain of the purchaser is
For the sake of notational unity, we will write as , therefore, the payment is
End of proof.
Proof Of Theorem 7
Here, we only analyse one purchaser and one supplier. During the auction phase, the social welfare guided by both parties is:
After negotiation, the social welfare are:
According to S ≥ P ≥ C,
End of proof.
Kairouz, P. , et al.: Advances and open problems in federated learning. Found. Trends® Mach. Learn. 14(1–2), 1–210 (2021). [DOI: https://dx.doi.org/10.1561/2200000083]
Zeng, D. , et al.: FedLab: A Flexible Federated Learning Framework (2021). arXiv preprint arXiv:210711621
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2023. This work is published under http://creativecommons.org/licenses/by-nc-nd/4.0/ (the "License"). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
The incentive mechanism of federated learning has been a hot topic, but little research has been done on the compensation of privacy loss. To this end, this study uses the Local SGD federal learning framework and gives a theoretical analysis under the use of differential privacy protection. Based on the analysis, a multi‐attribute reverse auction model is proposed to be used for user selection as well as payment calculation for participation in federal learning. The model uses a mixture of economic and non‐economic attributes in making choices for users and is transformed into an optimisation equation to solve the user choice problem. In addition, a post‐auction negotiation model that uses the Rubinstein bargaining model as well as optimisation equations to describe the negotiation process and theoretically demonstrate the improvement of social welfare is proposed. In the experimental part, the authors find that their algorithm improves both the model accuracy and the F1‐score values relative to the comparison algorithms to varying degrees.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer