1. Introduction and Preliminary
The theory of vector optimization is at the crossroads of many subjects. The terms “minimum,” “maximum,” and “optimum” are in line with a mathematical tradition, while words such as “efficient” or “non-dominated” find larger use in business-related topics. Historically, linear programs were the focus in the optimization community, and initially, it was thought that the major divide was between linear and nonlinear optimization problems; later, people discovered that some nonlinear problems were much harder than others, and the “right” divide was between convex and nonconvex problems. The author has determined that affineness and generalized affinenesses are also very useful for the subject “optimization”.
Suppose X, Y are real linear topological spaces [1].
A subset is called a linear set if B is a nonempty vector subspace of X.
A subset is called an affine set if the line passing through any two points of B is entirely contained in B (i.e., whenever and );
A subset is called a convex set if any segment with endpoints in B is contained in B (i.e., whenever and ).
Each linear set is affine, and each affine set is convex. Moreover, any translation of an affine (convex, respectively) set is affine (convex, resp.). It is known that a set B is linear if and only if B is affine and contains the zero point of X; a set B is affine if and only if B is a translation of a linear set.
A subset Y+ of Y is said to be a cone if for all and . We denote by the zero element in the topological vector space Y and simply by 0 if there is no confusion. A convex cone is one for which for all and . A pointed cone is one for which . Let Y be a real topological vector space with pointed convex cone Y+. We denote the partial order induced by Y+ as follows:
where intY+ denotes the topological interior of a set Y+.A function f: is said to be linear if
whenever and ; f is said to be affine if whenever ; and f is said to be convex if whenever .In the next section, we generalize the definition of affine function, prove that our generalized affine functions have some similar properties with generalized convex functions, and present some examples which show that our generalized affinenesses are not equivalent to one another.
In Section 3, we recall some existing definitions of generalized convexities, which are very comparable with the definitions of generalized affinenesses introduced in this article.
Section 4 works with optimization problems that are defined and taking values in linear topological spaces, devotes to the study of constraint qualifications, and derives some optimality conditions as well as a strong duality theorem.
2. Generalized Affinenesses
A function f: is said to be affine on D if , there holds
We introduce here the following definitions of generalized affine functions.
A function f: is said to be affinelike on D if such that
A function f: is said to be preaffinelike on D if such that
In the following Definitions 3 and 4, we assume that is any given linear set.
A function f: is said to be B-subaffinelike on D if , such that
A function f: is said to be B-presubaffinelike on D if , such that
For any linear set B, since , we may take u = 0. So, affinelikeness implies subaffinelikeness, and preaffinelikeness implies presubaffinelikeness.
It is obvious that affineness implies preaffineness, and the following Example 1 shows that the converse is not true.
An example of an affinelike function which is not an affine function.
It is known that a function is an affine function if and only it is in the form of ; therefore
is not an affine function.However, f is affinelike. taking
thenSimilarly, affinelikeness implies preaffinelikeness (), and presubaffinelikeness implies subaffinelikeness. The following Example 2 shows that a preaffinelike function is not necessary to be an affinelike function.
An example of a preaffinelike function which is not an affinelike function.
Consider the function .
Take , then ; but
thereforeSo f is not affinelike.
But f is an preaffinelike function. For taking if , if , then
where .An example of a subaffinelike function which is not an affinelike function.
Consider the function and the linear set .
taking then
therefore is B-subaffinelike on .is not affinelike on Actually, for one has , but
henceAn example of a presubaffinelike function which is not a preaffinelike function.
Actually, the function in Example 3 is subaffinelike, therefore it is presubaffinelike on D.
However, for one has
butHence
This shows that the function f is not preaffinelike on D.
An example of a presubaffinelike function which is not a subaffinelike function.
Consider the function .
Take the 2-dimensional linear set .
Take , then
Either or must be negative; but , ; therefore
And so, is not B-subaffinelike.
However, is B-presubaffinelike.
Case 1. If both of are positive, we take , , , then
Case 2. If both of are negative, we take , , , then
Case 3. If one of is negative, and the other is non-negative, we take
Then
And so are both non-negative or both negative; taking or , respectively, one has
whereTherefore, is B-presubaffinelike.
An example of a subaffinelike function which is not a preaffinelike function.
Consider the function
Take the 2-dimensional linear set
Take , then
In the above inequality, we note that either or , .
Therefore, is not preaffinelike.
However, is B-subaffinelike.
In fact, we may choose with x large enough such that
Then,
whereAn example of a preaffinelike function which is not a subaffinelike function.
Consider the function .
Take the 2-dimensional linear set .
Take , then
So, ,
However, for ,
(1)
Actually, if x = 0, it is obvious that ; if , the right side of (1) implies that , and the left side of (1) is . This proves that the inequality (1) must be true. Consequently,
So is not B-subaffinelike.
On the other hand, we may take if or if , then
where .Therefore, is preaffinelike.
So far, we have showed the following relationships (where subaffinelikeness and presubaffinelikeness are related to “a given linear set B”):
The following Proposition 1 is very similar to the corresponding results for generalized convexities (see Proposition 2).
Suppose f: is a function, a given linear set, and t is any real scalar.
- (a)
f is affinelike on D if and only if f (D) is an affine set;
- (b)
f is preaffinelike on D if and only if is an affine set;
- (c)
f is B-subaffinelike on D if and only if f (D) + B is an affine set;
- (d)
f is B-presubaffinelike on D if and only if + B is an affine set.
(a) If f is affinelike on D, , such that
Therefore, f (D) is an affine set.
On the other hand, assume that f (D) is an affine set. we have
Therefore, such that
And hence f is affinelike on D.
(b) Assume f is a preaffinelike function.
for such that
Since f is preaffinelike, such that
Therefore
where . Consequently, is an affine set.On the other hand, suppose that is an affine set. Then, since ,
Therefore, such that
Then, f is an affinelike function.
(c) Assume that f is B-subaffinelike.
, , such that and . The subaffinelikeness of f implies that , and such that
i.e.,Therefore
whereThen, f (D) + B is an affine set.
On the other hand, assume that f (D) + B is an affine set.
, such that
i.e., where . And hence f is B-subaffinelike.(d) Suppose f is a B-presubaffinelike function.
, similar to the proof of (b), , , for which and
where . This proves that + B is an affine set.On the other hand, assume that + B is an affine set.
, since , , such that
Therefore,
i.e., where . And so f is B-presubaffinelike. □The presubaffineness is the weakest one in the series of the generalized affinenesses introduced here. The following example shows that our definition of presubaffinelikeness is not trivial.
An example of non-presubaffinelike function.
Consider the function .
Take the linear set .
Take , then
Either or must be negative, but hold for ; therefore, for any scalar
(Actually, , one has ; and either or , then, either or ).
And so, is not B-presubaffinelike.
3. Generalized Convexities
In this section, we recall some existing definitions of generalized convexities, which are very comparable with the definitions of generalized affinenesses introduced in this article.
Let Y be a topological vector space, be a nonempty set, and Y+ be a convex cone in Y and .
It is known that a function f: is said to be Y+-convex on D if, for all , , there holds
The following Definition 5 was introduced in Fan [2].
A function f: is said to be Y+-convexlike on D if , , such that
We may define Y+-preconvexlike functions as follows.
A function f: is said to be Y+-preconvexlike on D if , , , such that
Definition 7 was introduced by Jeyakumar [3].
A function f: is said to be Y+-subconvexlike on D if , , , such that
In fact, in Jeyakumar [3], the definition of subconvexlike was introduced as the following form Definition 8.
A function f: is said to be Y+-subconvexlike on D if , , , , such that
Li and Wang ([4]) proved that: A function f: is Y+-subconvexlike on D by Definition 8 if and only if , , , such that
From the definitions above, one may introduce the following definition of presubconvexlike functions.
A function f: is said to be Y+-presubconvexlike on D if , , , , such that
And, similar to ([4]), one can prove that a function f: is Y+-presubconvexlike on D if and only if , , , , , such that
Our Definitions 7 and 9 are more comparable with our definitions of generalized affineness.
Similar to the proof of the above Proposition 1, we present the following Proposition 2.
Some examples of generalized convexities were given in [5,6].
Let f: X be function, and t > 0 be any positive scalar, then
(a) f is Y+-convexlike on D if and only if is convex;
(b) f is Y+-subconvexlike on D if and only if is convex;
(c) f is Y+-preconvexlike on D if and only if is convex;
(d) f is Y+-presubconvexlike on D if and only if is convex.
4. Constraint Qualifications
Consider the following vector optimization problem:
where f: , , , Y+, Zi+ are closed convex cones in Y and Zi, respectively, and D is a nonempty subset of X.Throughout this paper, the following assumptions will be used ( are real scalars).
such that
We note that the condition (A1) says that f and are presubconvexlike, and (j = 1, 2, …, n) are preaffinelike.
Let F be the feasible set of (VP), i.e.,
The following is the well-known definition of a weakly efficient solution.
A point is said to be a weakly efficient solution of (VP) with a weakly efficient value if for every there exists no satisfying .
We first introduce the following constraint qualification which is similar to the constraint qualification in the differentiate form from nonlinear programming.
Let . We say that (VP) satisfies the No Nonzero Abnormal Multiplier Constraint Qualification (NNAMCQ) at if there is no nonzero vector satisfying the system
where is some neighborhood of .It is obvious that NNAMCQ holds at with being the whole space X if and only if for all satisfying , there exists such that
Hence, NNAMCQ is weaker than ([7], (CQ1)) (in [7], CQ1 was for set-valued optimization problems) in the constraint , which means that only the binding constraints are considered. Under the NNAMCQ, the following KuhnTucker type necessary optimality condition holds.
Assume that the generalized convexity assumption (A1) is satisfied and either (A2) or (A3) holds. If is a weakly efficient solution of (VP) with , then exists a vector with such that
(2)
for a neighborhood of .Since is a weakly efficient solution of (VP) with there exists a nonzero vector such that (2) holds. Since NNAMCQ holds at , must be nonzero. Otherwise if = 0 then must be a nonzero solution of
But this is impossible, since the NNAMCQ holds at . □
Similar to ([7], (CQ2)) which is slightly stronger than ([7], (CQ1)), we define the following constraint qualification which is stronger than the NNAMCQ.
(SNNAMCQ) Let . We say that (VP) satisfies the No Nonzero Abnormal Multiplier Constraint Qualification (NNAMCQ) at provided that
- (i)
satisfying,
- (ii)
, , s.t. for all .
We now quote the Slater condition introduced in ([7], (CQ3)).
(Slater Condition CQ). Let . We say that (VP) satisfies the Slater condition at if the following conditions hold:
- (i)
, s.t. ;
- (ii)
for all j.
Similar to ([7], Proposition 2) (again, in [7], discussions are made for set-valued optimization problems), we have the following relationship between the constraint qualifications.
The following statements are true:
(i) Slater CQ SNNAMCQ NNAMCQ with being the whole space X;
(ii) Assume that (A1) and (A2) (or (A1) and (A3)) hold and the NNAMCQ with being the whole space X without the restriction of at . Then, the Slater condition (CQ) holds.
The proof of (i) is similar to ([7], Proposition 2). Now we prove (ii). By the assumption (A1), the following sets C1 and C2 are convex:
Suppose to the contrary that the Slater condition does not hold. Then, or . If the former holds, then by the separation theorem [1], there exists a nonzero vector such that
for all . Since are convex cones, consequently we have(3)
for all and take in (3), we have which contradicts the NNAMCQ. Similarly if the latter holds then there exists such that , which contradicts NNAMCQ. □(Calmness Condition). Let . Let and . We say that (VP) satisfies the calmness condition at provided that there exist , a neighborhood of , and a map with such that for each
Satisfying
there is no , such thatAssume that (A1) is satisfied and either (A2) or (A3) holds. If is a weakly efficient solution of (VP) with , and the calmness condition holds at , then there exists , a neighborhood of , and a vector with such that
(4)
It is easy to see that under the calmness condition, being a weakly efficient solution of (VP) implies that is a weakly efficient solution of the perturbed problem: VP(p,q)
By assumption, the above optimization problem satisfies the generalized convexity assumption (A1). Now we prove that the NNAMCQ holds naturally at . Suppose that satisfies the system:
(5)
If , then there exists small enough such that . Since , , and there exists , which implies that , hence
which contradicts (5). Hence, and (5) becomesIf , then there exists p small enough such that . Let , then
and hence which is impossible. Consequently, as well. Hence, there exists with such that(6)
It is obvious that (6) implies (4) and hence the proof of the theorem is complete. □
Let be normed spaces. We say that (VP) satisfies the error bound constraint qualification at a feasible point if there exist positive constants , and such that
where BX is the unit ball of X, andNote that the error bound constraint qualification is satisfied at a feasible point if and only if the function is pseudo upper-Lipschitz continuous around in the terminology of ([8]) (which is referred to as being calm at in [9]). Hence, being either pseudo-Lipschitz continuous around in the terminology of [10] or upper-Lipschitz continuous at in the terminology of [11] implies that the error bound constraint qualification holds at . Recall that a function is called a polyhedral multifunction if its graph is a union of finitely many polyhedral convex sets. This class of function is closed under (finite) addition, scalar multiplication, and (finite) composition. By ([12], Proposition 1), a polyhedral multifunction is upper-Lipschitz. Hence, the following result provides a sufficient condition for the error bound constraint qualification.
Let X = Rn and W = Rm. Suppose that D is polyhedral and h is a polyhedral multifunction. Then, the error bound constraint qualification always holds at any feasible point .
Since D is polyhedral and h is a polyhedral multifunction, its inverse map is a polyhedral multifunction. That is, the graph of S is a union of polyhedral convex sets. Since
which is also a union of polyhedral convex sets, is also a polyhedral multifunction and hence upper-Lipschitz at any point of by ([12], Proposition 1). Therefore, the error bound constraint qualification holds at . □Let X be a normed space, be a function, and . f is said to be Lipschitz near if there exist , a neighborhood of , and a constant Lf > 0 such that for all ,
where BY is the unit ball of Y.Let X be a normed space, be a function and . f is said to be strongly Lipschitz on if there exist a constant Lf > 0 such that for all , and ,
The following result generalizes the exact penalization [13].
Let X be a normed space, be a function which is strongly Lipschitz of rank Lf on a set . Let and suppose that is a weakly efficient solution of
with . Then, for all , is a weakly efficient solution of the exact penalized optimization problemwhere .Let us prove the assertion by supposing the contrary. Then, there is a point , , and satisfying . Let and be a point such that . Then, for any ,
Since is arbitrary, it contradicts the fact that is a weakly efficient solution of
□Suppose is a normed space and f is strongly Lipschitz on D. If is a weakly efficient solution of (VP) and the error bound constraint qualification is satisfied at , then (VP) satisfies the calmness condition at .
By the exact penalization principle in Proposition 5 is a weakly efficient solution of the penalized problem
The results then follow from the definitions of the calmness and the error bound constraint qualification. □
Assume that the generalized convexity assumption (A1) is satisfied with f replaced by and either (A2) or (A3) holds. Suppose is a normed space and f is strongly Lipschitz on D. If is a weakly efficient solution of (VP) and the error bound constraint qualification is satisfied at , then there exist , a neighborhood of , and a vector with such that (4) holds.
Using Proposition 4, Theorem 3 has the following easy corollary.
Suppose Y is a normed space, X = Rn, W = Rm and D is polyhedral, and f is strongly Lipschitz on D. Assume that the generalized convexity assumption (A1) is satisfied with f replaced by and either (A2) or (A3) holds. If is a weakly efficient solution of (VP) without the inequality constraint , and h is a polyhedral multifunction, then there exist , a neighborhood of a vector with such that
Our last result Theorem 4 is a strong duality theorem, which generalizes a result in Fang, Li, and Ng [14].
For two topological vector spaces Z and Y, let B(Z; Y) be the set of continuous linear transformations from Z to Y and
The Lagrangian map for (VP) is the function
defined byGiven , consider the vector minimization problem induced by (VP):
and denote by the set of weakly efficient value of the problem (VPST). The Lagrange dual problem associated with the primal problem (VP) isThe following strong duality result holds which extends the strong duality theorem in ([7], Theorem 7) (which was for set-valued optimization problems), to allow weaker convexity assumptions. We omit the proof since it is similar to [7].
Assume that (A1) is satisfied, either (A2) or (A3) is satisfied, and a constraint qualification such as NNAMCQ is satisfied. If is a weakly efficient solution of (VP), then there exists
such that5. Conclusions
We introduce the following definitions of generalized affine functions: affinelikeness, preaffinelikeness, subaffinelikeness, and presubaffinelikeness. Examples 1 to 7 show that definitions of affine, affinelike, preaffinelike, subaffinelike, and presubaffinelike functions are all different. Example 8 is an example of non-presubaffinelike function; presubaffineness is the weakest one in the series. Proposition 1 demonstrates that our generalized affine functions have some similar properties with generalized convex functions.
And then, we work with vector optimization problems in real linear topological spaces, and obtain necessary conditions, sufficient conditions, or necessary and sufficient conditions for weakly efficient solutions, which generalize the corresponding classical results in [13,15] and some recent results in [7,9,16,17,18]. We note that the constraint qualifications in [13,17,18] are in the differentiation form. Compared with the results in [19] and ([20], p. 297) in discussions of convex constraints, we only required weakened convexities for constraint qualifications in this article. We note that [17] works with semi-definite programming. In [17], two groups of functions gi(x) ≥ 0, iI and hj(x) = 0, jJ can be just considered as two topological spaces (I and J do not have to be finite sets). We also note that f is supposed to be “proper convex” in [18]; and in [18], functions are required to be “quasiconvex”.
Generalized affine functions and generalized convex functions can be used for other discussions of optimization problems, e.g., dualities, scalarizations, as well as saddle points, etc.
Not applicable.
The author declares no conflict of interest.
Footnotes
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
References
1. Deimling, K. Nonlinear Functional Analysis; Springer: Berlin/Heidelberg, Germany, 1985.
2. Fan, K. Minimax Theorems. Proc. Natl. Acad. Sci. USA; 1953; 39, pp. 42-47. [DOI: https://dx.doi.org/10.1073/pnas.39.1.42] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/16589233]
3. Jeyakumar, V. Convexlike Alternative Theorems and Mathematical Programming. Optimization; 1985; 16, pp. 643-652. [DOI: https://dx.doi.org/10.1080/02331938508843061]
4. Li, Z.F.; Wang, S.Y. Lagrange Multipliers and Saddle Points in Multiobjective Programming. J. Optim. Theo. Appl.; 1994; 1, pp. 63-80. [DOI: https://dx.doi.org/10.1007/BF02191762]
5. Zeng, R. Generalized Gordan Alternative Theorem with Weakened Convexity and its Applications. Optimization; 2002; 51, pp. 709-717. [DOI: https://dx.doi.org/10.1080/0233193021000031615]
6. Zeng, R.; Caron, R.J. Generalized Motzkin Theorem of the Alternative and Vector Optimization Problems. J. Optim. Theo. Appl.; 2006; 131, pp. 281-299. [DOI: https://dx.doi.org/10.1007/s10957-006-9140-6]
7. Li, Z.-F.; Chen, G.-Y. Lagrangian Multipliers, Saddle Points and Duality in Vector Optimization of Set-Valued Maps. J. Math. Anal. Appl.; 1997; 215, pp. 297-315. [DOI: https://dx.doi.org/10.1006/jmaa.1997.5568]
8. Ye, J.J.; Ye, X.Y. Necessary Optimality Conditions for Optimization Problems with Variational Inequality Constriants. Math. Oper. Res.; 1997; 22, pp. 977-997. [DOI: https://dx.doi.org/10.1287/moor.22.4.977]
9. Rockafellar, R.T.; Wets, R.J.-B. Variational Analysis; Springer-Verlag: Berlin/Heidelberg, Germany, 1998.
10. Aubin, J.-P. Lipschitz Behavior of Solutions to Convex Minimization Problems. Math. Oper. Res.; 1984; 9, pp. 87-111. [DOI: https://dx.doi.org/10.1287/moor.9.1.87]
11. Robinson, S.M. Stability Theory for Systems of Inequalities. Part I: Linear Systems. SIAM J. Numer. Anal.; 1975; 12, pp. 754-769. [DOI: https://dx.doi.org/10.1137/0712056]
12. Robinson, S.M. Some Continuity Properties of Polyhedral Multifunctions. Math. Program. Stud.; 1981; 14, pp. 206-214.
13. Clarke, F.H. Optimization and Nonsmooth Analysis; Wiley-Interscience: New York, NY, USA, 1983.
14. Fang, D.H.; Li, C.; Ng, K.F. Constraint Qualifications for Optimality Conditions and Total Lagrange Dualities in Convex Infinite Programming. Nonlinear Anal.; 2010; 73, pp. 1143-1159. [DOI: https://dx.doi.org/10.1016/j.na.2010.04.020]
15. Luc, D.T. Theory of Vector Optimization; Springer: Berlin/Heidelberg, Germany, 1989.
16. Nguyen, M.-H.; Luu, D.V. On Constraint Qualifications with Generalized Convexity and Optimality Conditions; Cahiers de la Maison des Sciences Economiques Université Panthéon-Sorbonne (Paris 1): Paris, France, 2006; 20.
17. Kanzi, N.; Nobakhtian, S. Nonsmooth Semi-Infinite Programming Problems with Mixed Constraints. J. Math. Anal. Appl.; 2009; 351, pp. 170-181. [DOI: https://dx.doi.org/10.1016/j.jmaa.2008.10.009]
18. Zhao, X.P. Constraint Qualification for Quasiconvex Inequality System with Applications in Constraint Optimization. J. Nonlinear Convex. Anal.; 2016; 17, pp. 879-889.
19. Khazayel, B.; Farajzadeh, A. On the Optimality Conditions for DC Vector Optimization Problems. Optimization; 2022; 71, pp. 2033-2045. [DOI: https://dx.doi.org/10.1080/02331934.2020.1847109]
20. Ansari, Q.H.; Yao, J.-C. Recent Developments in Vector Optimization; Springer Link: Berlin/Heidelberg, Germany, 2012.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2023 by the author. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
In this paper, we introduce a series of definitions of generalized affine functions for vector-valued functions by use of “linear set”. We prove that our generalized affine functions have some similar properties to generalized convex functions. We present examples to show that our generalized affinenesses are different from one another, and also provide an example to show that our definition of presubaffinelikeness is non-trivial; presubaffinelikeness is the weakest generalized affineness introduced in this article. We work with optimization problems that are defined and taking values in linear topological spaces. We devote to the study of constraint qualifications, and derive some optimality conditions as well as a strong duality theorem. Our optimization problems have inequality constraints, equality constraints, and abstract constraints; our inequality constraints are generalized convex functions and equality constraints are generalized affine functions.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer