The notion of well-posedness is significant for several mathematical problems and it is closely related to the stability of an optimization problem: it plays, in fact, a crucial role in the theoretical and in the numerical aspects of optimization theory   . The study of well-posedness, used also in different areas as mathematical programming, calculus of variations and optimal control, becomes important mainly for problems in which, due to certain hypotheses, the optimization models are imprecise or when the existing algorithms in literature are sufficient enough to guarantee only the approximate solutions of such problems while the exact solution may not exist or may even be more difficult to compute. Under these hypotheses, the well-posedness of an optimization problem is fundamental, in the sense that it ensures the convergence of the sequence of approximate solutions, obtained through iterative techniques, to the exact solution of the problem.
Two different concepts of well-posedness are known in scalar optimization. The first, due to J. Hadamard, requires existence and uniqueness of the optimal solution and studies its continuous dependence from the data of the considered optimization problem. The second approach, introduced by A. N. Tykhonov, in 1966, requires, instead, besides the existence and the uniqueness of the optimal solution, the convergence of every minimizing sequence of approximate solutions to the unique minimum point. The links between Hadamard and Tykhonov well-posedness have been studied in    . There, besides uniqueness, additional structures are involved: in   , for example, basic ingredient is convexity.
The notion of well-posedness for a vector optimization problem is, instead, less developed, less advanced; there is no commonly accepted definition of well-posed problem, in vector optimization. Some attempts in this direction have been already done     and have been made some comparisons with their scalar counterparts. For instance,  gave a survey on various aspects on well-posedness of optimization problems.
The well-posedness was generalized also to other contexts: variational inequalities, Nash equilibria and saddle point problems, all special cases of an equilibrium problem. For instance,  investigated well-posedness for optimization problems with constraints defined by variational inequalities while Margiocco et al.    discussed Tykhonov well-posedness for Nash equilibria.  at last, gave a definition of well-posed for saddle point problems and related results.  introduced the notion of well-posedness for variational inequality problems based on the fact that an optimization problem can be formulated as a variational inequality problem involving the derivative of the objective function. In all these cases, the idea is an extension of the concept of minimizing sequences seen as approximate solutions.
2. Research Aims
The aim of this survey is twofold. The first aim is to recall some basic aspects of the mathematical theory of well-posedness in scalar optimization, to collect the two notions of well-posedness, Tykhonov well-posedness and Hadamard well-posedness, to give some strengthened versions of well-posedness and to show, in particular, some generalizations of the two types of well-posedness. The underlying idea is that sometimes the uniqueness of the solution could be dropped. Indeed, in different situations like in linear and quadratic programming, there is not always required the uniqueness of the solution; sometimes, namely, the uniqueness of the solution for a particular minimization problem is not of such an importance as its stability. The second aim is to present the notion of well-posedness in the vector optimization and, in particular, to verify if the well-posedness of the vector problem is equivalent to the well-posedness of the scalarized problem or better investigate the links between the well posedness of a vector optimization problem and of a vector variational inequality. Among the various vector well-posedness notions known in the literature, the attention is focused on the concept of pointwise well-posedness. After a review of well-posedness properties, the authors extend the study to a scalarizing procedure that preserve well-posedness of the notions listed, namely to a result, obtained with a special scalarizing function, which links the notion of pontwise well-posedness to the well-posedness of a suitable scalar variational inequality of a differential type.
The authors hope that the paper is very useful for stimulating the research and for providing fresh insights leading to new applications.
The paper is organized as follows. In Section 2, after the introduction, the research aims are analysed while in Section 3 some results on Tykhonov well-posedness and on Hadamard well-posedness and on their relations are analysed. In Section 4, some generalizations of the notion of well-posedness are investigate (in case in which there is not uniqueness of solutions) and some strenghthened versions of well-posedness (for istance well-posedness in the sense of Levitin and Polyak) while in Section 5 are studied some results of well-posedness of vector optimization problems and among the various vector well-posedness notions, known in the literature, the attention is focused on the concept of pointwise well-posedness, introduced in  , (in particular a type of pointwise well-posedness and strong pointwise well-posedness for vector optimization problems). Subsequently, always in the section, are established basic well-posedness results for a vector variational inequality. Section 6 is devoted to the main results of the paper obtained by means of a special scalarization function. The notion of pontwise well-posedness is linked to well-posedness of a suitable scalar variational inequality of a differential type whose construction represents an interesting application of the so-called “oriented distance function”, a special scalarizing function, which allows to establish a parallelism between the well-posedness of the original vector problem and the well-posedness of the associate scalar problem. Section 7, finally, contains a general discussion on directions for future research and provides a conclusion. In other words, the article ends with some concluding remarks while the last part of this article represents the reviewed references. The emphasis is layed on papers published in the last three decades.
3. Tykhonov and Hadamard Well-Posedness
In scalar optimization the different notions of well-posedness are based either on the behaviour of “appropriate” minimizing sequences (converging to a solution of the problem) or on the dependence of the optimal solutions on the data of optimization problem. This section is devoted exactly to a study of two different notions of well-posedness. In particular, in it the authors give, initially, a characterization of the Tykhonov well-posedness, and a characterization of the Hadamard well-posedness, for a problem of minimizing a function f on a closed and convex set K; subsequently, they show the links between the two definitions and also some extensions and summarize some known results.
1) Tykhonov well-posedness
The first notion of well-posedness of an optimization problem was introduced in 1966 by A.N. Tykhonov and later took his name.
Let be a real-valued function and let K be a nonempty subset of . Throughout this paper, the scalar optimization problem:
is denoted by and consists in finding such that
The set, possible empty, of the solutions of the optimization problem is denoted by argmin .
The optimization problem is said Tykhonov well-posed if it satisfies together the following properties:
a) existence of the solution (i.e. has a solution),
b) uniqueness of the solution (i.e. the solution set for is a singleton),
c) is a good approximation of the solution of , if is close to
The problem is said Tykhonov well-posed if there exist exactly a unique such that for all , and if for any sequence such that (i.e. ).
Recalling that a sequence is said minimizing sequence for problem when as , the previous definition can be rephrased in equivalent way, so  :
Definition 3.1: The problem is said Tykhonov well-posed if it has, on K, a unique global minimum point, , and, moreover, every minimizing sequence for converges to .
The definition 2.1 is motivated by the fact that, usually, every numerical method for solving provides iteratively some minimizing sequences for ; such sequences are also called sequences of approximate solutions for the problem and therefore it is important to be sure that the approximate solutions are not far from the (unique) minimum .
In other words, the Tykhonov well-posedness of the optimization problem requires existence and uniqueness of minimum point towards which every sequence of approximate solutions of the problem converges. More precisely, to consider well-posedness of Tykhonov type, it is introduced the notion of “approximating sequence” for the solutions of optimization problems and it is required convergence of such sequences to a solution of the problem. For more details see   .
When K is compact, the uniqueness of the solution of a minimization problem is enough to guarantee its well-posedness but there are however simple examples in which the uniqueness of the solution of is not enough to guarantee its Tykhonov well-posedness even for continuous functions.
A simple example of a problem with a unique solution but which is not Tykhonov well-posed is the following:
Obviously . has a unique solution at zero, namely the argmin , while , provides a minimizing sequence which does not converge to this unique solution. Hence is not Tykhonov well-posed. Therefore, for continuous functions the Tykhonov well-posedness of an optimization problem simply means that every minimizing sequence of is convergent.
Let . If , has a unique minimum but it is not Tykhonov well-posed, since the sequences is minimizing but it does not converges to .
If , then is Tykhonov well-posed.
For convex functions in finite dimensions the uniqueness of the solution is enough to guarantee its Tykhonov well-posedness while this is no longer valid in infinite dimensions  . It is, in fact, known the following result:
Proposition 3.1: (  ) Let be a convex function and let K be convex. If has a unique solution, then is Tykhonov well-posed.
Different characterizations of Tykhonov well-posedness for minimization problems determined by convex functions in Banach spaces can be found in  .
The next fundamental theorem  gives an alternative characterization of Tykhonov well-posed problems: it uses the set of ε-optimal solutions and states that Tykhonov well-posedness of can be characterized by behaviour of as .
Theorem 3.1: If the minimization problem is Tykhonov well-posed, then
is the set of ε-minimizers (approximate solutions) of f over K and diam denotes the diameter of given set.
Conversely, if f is lower semicontinuous and bounded from below on K,
implies Tykhonov well-posedness of .
When K is closed and f is lower semicontinuous and bounded, from below it is possible to use the sets:
to introduce the notion of well-posedness of :
Definition 3.3: Let K be closed and let be lower semicontinuous. The minimization problem is said to be well-posed if:
Of course, if to any of the notions of generalized well-posedness is added the uniqueness of the solution, it is obtained the corresponding non generalized notion. Different characterizations of Tykhonov well-posedness for minimization problems determined by convex functions in Banach spaces can be found in  .
2) Hadamard well-posedness
The second notion of well-posedness is inspired by the classical idea of J. Hadamard to the beginning of previous century: it requires existence and uniqueness of solution of the optimization problem together with continuous dependence of the optimal solution and optimal value on the data of the problems.
Definition 3.3: The minimization problem is said to be Hadamard well-posed if it has unique solution ( ) and depends continuously on the data of the problem.
This is the well-known condition of well-posedness considered in the study of differential equations, translated for minimum problems. The essence of this notion is that a “small” change of the data of the problem yields a “small” change of the solution.
In fact very often the mathematical model of a phenomenon is so complicated that it is necessary to simplify it and replace it by other model which is “near” the original and, at the same time, it is important to be sure that the new problem will have a solution which is “near” the original one. The well-known variational principle of Ekeland  , an important tool for nonlinear analysis and optimization, asserts just that a particular optimization problem can be replaced by other which is near the original and has a unique solution.
3) Relations between Hadamard and Tykhonow well-posedness
Almost all the literature deals with different notions of well-posedness, even if especially with Tykhonov well-posedness. Some researchers have investigated the relations between these notions of well-posedness but there is no general research to such relations. At first sight, the two notions seem to be independent but, at least in the convex case, there are some papers showing a connection between the two properties: for instance    . The two notions (Tykhonov and Hadamard well-posedness) are equivalent at least for continuous objective functions. The links between Hadamard and Tykhonov well-posedness have been studied in    . There, besides uniqueness, additional structures are involved: in   , for example, basic ingredient is convexity. The object of this section is to describe generally the relations between Hadamard and Tykhonov well-posedness: a central role is provided by the well-known Hausdorff convergence.
We remember the concept of Hausdorff convergence of sequences of sets.
Let D, E be subsets of and define
Definition 3.4: Let be a sequences of subsets of . We say that converges to in the sense of Hansdorff, and we write when .
The following theorems  show the relations between the Tykhonov and the Hadamard well-posedness:
Theorem 3.2: Let K be a closed convex subset of and let be a convex continuous function with one and only one minimum point on every closed and convex subset of K. If is
Hadamard well-posed, with respect to the well-known Hausdorff convergence, then is Tykhonov well-posed on every closed and convex subset of K.
Theorem 3.3: Let be a convex function uniformly continuous on every bounded set. If is Tykhonov well-posed on every closed and convex set, then is Hadamard well-posed, with respect to the Hausdorff convergence.
The Tykhonov well-posedness does not, in general, imply the Hadamard well-posedness if the objective function is only continuous.
4. Some Generalizations
In the above definitions it is required the existences and the uniqueness of solution towards which every minimizing sequence converges. The different notions of well-posedness, however, admit generalizations which do not require uniqueness of the solution. In other words, the uniqueness requirement can be relaxed and well-posed optimization problems with several solutions can be considered. Therefore, while the requirement of existence in the previous definitions is crucial, the uniqueness condition is more debatable. In fact, many problems in linear and quadratic programming or many multicriteria optimization problems are usually considered as well-posed problems, although uniqueness is usually not satisfied  .
More precisely, in scalar optimization problems it is difficult to guarantee the uniqueness of the optimal solutions, uniqueness that is critical to the solution stability and calculation.
In other words, the different notions of well-posedness admit generalizations which do not require uniqueness of the solution. In particular, the concept of Tykhonov well posedness can be extended to minimum problems without uniqueness of the optimal solutions. It becomes imperative, namely, to generalize the notion of well-posedness for a minimization problem, introduced by Tykhonov, based on the fact that every minimizing sequence converges towards the unique minimum solution and to discuss the well-posedness for problems having more than one solution.
This new definition requires existence, but not uniqueness, of solution of , and, for every minimizing sequences, the convergence of some subsequence of the minimizing sequence towards some optimal solution.
Definition 4.1: The problem is called Tykhonov well-posed in the generalized sense if every minimizing sequence for has some subsequence converging to an optimal solution of , i.e. to an element of .
More precisely the problem is called Tykhonov well-posed in the generalized sense if and every sequence such that has some subsequence with .
From the definition it follows, obviously, that, if the problem is Tykhonov well-posed in the generalized sense, then it has a non-empty compact set of solutions, i.e. is nonempty and compact. Moreover, when is well-posed in the generalized sense and is a singleton (i.e. its solution is unique), then is Tykhonov well-posed.
When is a singleton, the previous definition reduces to the classical notion of Tykhonov well-posedness or rather the problem is Tykhonov well-posedness if it is Tykhonov well-posed in the generalized Tykhonov sense and is a singleton; thus generalized well-posedness is really a generalization of Tykhonov well-posedness.
In order to weaken the requirement of uniqueness of the solution, other more general notions of well-posedness have been introduced, depending on the hypotheses made on f (and K). Here, the author recall the concept of well-setness introduced in  .
Definition 4.2: Problem is said to be well-set when, for every minimizing sequence
, , as ,
where denotes the set of solutions of problem while is the distance of the point x from the set K.
The idea of the behaviour of the minimizing sequences was used by different authors also to extend this concept to strengthened notions. These notions are not suitable for numerical methods, where the function f is approximated by a family or a sequence of functions. For this reason new notions of well-posedness have been introduced and studied.
Before, however, we consider two generalizations of the notion of minimizing sequence.
The first was introduced and studied by  ; they introduced a new notion of well-posedness that strengthened the Tykhonov’s concept as it required the convergence to the optimal solution of each sequence belonging to a larger set of minimizing sequences. The Levitin-Polyak well-posedness has been investigated intensively in the literature, such as     .
Konsulova and Revalski  studied Levitin-Polyak well-posedness for convex scalar optimization problems with functional constraints. While, recently,  generalized the results of Konsulova and Revalski  for non convex optimization problems with abstract and functional constraints.
The well-posedness of the minimization problem in the sense of Tykhonov concerns the behaviour of the function f in the set K but it does not take into account the behaviour of f outside K  . Of course, often, one can come across with minimizing sequences that do not lie necessarily in K and one wants to control the behaviour of these minimizing sequences, as well. Levitin and Polyak in  considered such kind of sequences.
Definition 4.3: Let K be a nonempty subset of . The sequences is a Levitin-Polyak minimizing sequences for the minimization problem if
where is the distance from the point to the set K while is the Euclidean norm.
In other words, a sequences is a Levitin-Polyak minimizing sequences for if not only approaches the greatest lower bound of f over K but also the sequence tends to K.
Then, the well-posedness concept can be strengthened as follows:
Definition 4.4: The minimization problem is called Levitin-Polyak well-posed if it has unique solution and, moreover, every Levitin-Polyak minimizing sequence for converges to .
Of course, this definition is stronger than that of Tykhonov since requires that each sequence, belonging to a larger set of minimizing sequences, convergs to the unique solution, namely Levitin-Polyak well-posedness implies Tykhonov well-posedness.
The converse is true provided that f is uniformly continuous but not necessarily true if f is only continuous. It is enough to consider
and the generalized minimizing sequence .
As Tykhonov well-posedness can be characterized by the behaviour of , as Levitin Polyak well-posedness can be characterized by the behaviour of the set:
defined for and for f bounded from below on K.
In analogy with Theorem 3.1, the following result gives  :
Theorem 4.2: If K is closed and f is lower semicontinuous and bounded from below on K, then as implies Levitin-Polyak well-posedness of
A second generalization of the usual notion of minimizing sequences is the following:
Definition 4.5: A sequence is said to be a generalized minimizing sequence for the minimization problem if are fulfilled both:
Consequently another strengthened version of the well-posedness is the following:
Definition 4.6: The minimization problem is said strongly well-posed if it has unique solution and, moreover, every generalized minimizing sequences for converges to .
Obviously, in general strong well-posedness of the problem implies that of Levitin-Polyak, which in its turn implies the Tykhonov well-posedness. It is important underline that, each of the previous definitions, widely studied in many papers    , is based on the behaviour of a certain set of minimizing sequences.
The corresponding generalization of Levitin-Polyak well-posedness in the case of non-uniqueness of the solution, or when the uniqueness of the solution is dropped, is:
Definition 4.7: The minimization problem is called generalized Levitin-Polyak well-posed if every Levitin-Polyak minimizing sequence for has a subsequence converging to a solution of .
Of course, any of the notions of generalized well-posedness, at which is added the uniqueness of the solution, is equivalent, obviously, to corresponding non generalized notion.
5. Well-Posedness of Vector Optimization Problems
In scalar optimization, the different notions of well-posedness are based either on the behaviour of “appropriate” minimizing sequences or on the dependence of optimal solution with respect to the data of optimization problems. In vector optimization, instead, there is not a commonly accepted definition of well-posedness but there are different notions of well-posedness of vector optimization problems. For a detailed survey on these problems it is possible to refer to      .
In this section, we propose some of these definitions of well-posedness for a vector optimization problem; in particular, among the various vector well-posedness notions known in the literature, the attention is focused on the concept of pointwise well-posedness, introduced in  .
We consider the vector optimization problem:
where K is a nonempty, closed, convex subset of , is a continuous function and is a closed, convex, pointed cone and with nonempty interior. Denoted by the interior of C.
A point is said to be an efficient solution or minimal solution of problem when:
If, in the above definition, instead of the cone C is used the cone , is said weak minimal solution. Then, a point is said to be a weakly efficient solution or weak minimal solution of problem when:
The set of all efficient solutions (minimal solutions) of problem is denoted by while denotes the set of weakly efficient solutions (weak minimal solutions) of . Moreover, every minimal is also a weak minimal solution but the converse is not generally true.
In this section the authors recall a notion of well-posedness that considers a single point (a fixed efficient solution) and not the whole solution set: a particular type of pointwise well-posedness and strong pointwise well-posedness for vector optimization problems. This definition can be introduced considering, as in the scalar case, the diameter of the level sets of the function f.
Generalizing Tykhonov’s definition of well-posedness for a scalar optimization problem, in  are introduced the notions of well-posedness and of strong well-posedness of vector optimization problem at a point and are provided, also, some conditions to guarantee well-posedness according to these definitions.
Definition 5.1: The vector optimization problem is said to be pointwise well-posed at the efficient solution or Tykhonov well-posed at , if:
Definition 5.2: The vector optimization problem is said to be strongly pointwise well-posed at the efficient solution , or Tykhonov strongly well-posed at , if:
For the sake of completeness, we recall that it is also possible to introduce another type of well-posedness of the vector optimization problem at a point  .
Definition 5.3: The vector optimization problem is said to be H-well-posed at a point if for any sequence , such that .
Definition 5.4: The vector optimization problem is said to be strongly H-well-posed at a point if for any sequence such that with .
If , then well-posedness at a point of the vector optimization problem , according to definition 5.1 [resp. to def. 5.2], implies well-posedness according to definition 5.3 [resp. to def. 5.4]. It is easy realize that the pointwise well-posedness of type 5.1 is weaker than pointwise well-posedness of type 5.3  .
An useful tool in the study of vector optimization problems is provided by the vector variational inequalities, that, introduced first by Giannessi in 1980, have been studied intensively because they can be efficient tools for investigating vector optimization problems and also because they provide a mathematical model for equilibrium problems; they provide, namely, an unified and efficient framework for a wide spectrum of applied problems.
Before, however, it is important to underline that the theory of variational inequalities provides a convenient mathematical apparatus for obtain result relating to a large number of problems with a wide range of applications in economics, finance, social, pure and applied sciences. In fact, it is well known that many equilibrium problems, arising in finance, economics, transportation science and contact problems in elasticity, can be formulated in terms of the variational inequalities  . In other words, the ideas and the techniques of variational inequalities are being applied in a variety of diverse areas of sciences and prove to be productive and innovative.
There is a very close connection between the optimization problems and the variational inequalities. In fact, the well-posedness of a scalar minimization problem is linked to that of a scalar variational inequality and, in particular, to a variational inequality of differential type (i.e. in which the operator involved is the gradient of a given function). The links between variational inequalities of differential type and optimization problems have been deeply studied in     . Furthermore, by means of Ekeland’s variational principle  , that, as it is well known, is an important tool to prove some results in well-posedness for optimization, a notion of well-posed scalar variational inequality has been introduced and its links with the concept of well-posed optimization problem have been investigated  .
In this section, are treated the vector variational inequalities of differential type.
Let be a function differentiable on an open set containing the closed convex set . The vector variational inequality problem of differential type consists in finding a point such that:
where denotes the Jacobian of f and is the vector whose components are the l inner products .
It is well known that provides a necessary condition for to be an efficient solution of . It is, instead, a sufficient condition for to be an efficient solution of if f is -convex while, if f is C-convex, is a sufficient condition for to be an weakly efficient solution of . These remarks underline the links between optimization problems and variational inequalities also for vector case. This is a further reason for a suitable definition of well-posedness for a vector variational inequality which could be compared and related to the given definition for vector optimization. Then, a notion of well-posedness is introduced for the vector variational inequality problem , obtained by generalizing the definition of the scalar case and it is defined the following set:
where and . is a directional generalization of the set of the scalar case.
Definition 5.5: The variational inequality is well-posed if, for every , where .
The following result states the relationship between well-posed optimization problem and a well-posed variational inequality, in the vector case  .
Theorem 5.1: If the variational inequality is well-posed, then problem is well-posed at .
For C-convex functions, in particular, well-posedness of and substantially coincide. To show that, it is necessary to assume that f is differentiable on an open set containing K and observe that:
Definition 5.6: The function is said to be C-convex when:
Lemma 1: If is C-convex, then:
Theorem 5.2: Let f be a C-convex function. Assume that , and that is bounded for some . Then is well-posed.
Therefore, if f is a C-convex function, the well-posedness of is ensured and, namely, by theorem 3.2, substantially coincide with well-posedness of .
6. Main Results
In this section, the authors, after a review of well-posedness, focus their attention on a scalarization procedure that preserve well-posedness of the notions listed above and among various scalarization procedures known in the literature, they consider the one based on the so called “oriented distance” function from a point to a set. This special scalarizing function, introduced by Hiriart-Urruty in  , has been applied to scalarization of vector optimization problem  . The scalarization method is, namely, a powerful tool for studying vector optimization problems.
This function allows to establish a parallelism between the well-posedness of the original vector problem and the well-posedness of the associate scalar problem. Indeed, the authors show that one of the weakest notions of well-posedness in vector optimization is linked to the well-setness of the scalarized problem, while some stronger notion of well-posedness in the vector case is related to Tykhonov well-posedness of the associated scalarization.
These results constitute a simple tool to show that, under some additional compactness assumptions, quasiconvex vector optimization problems are well-posed. Thus, a known result about scalar problems can be extended to vector optimization and improves a previous result concerning convex vector problems.
Throughout this section we assume that is differentiable on an open set containing the closed convex set .
Definition 6.1: For a set , let be defined as:
where is the distance from the point y to the set A.
Function is called the oriented distance function from the point y to the set A and it has been introduced in the framework of nonsmooth scalar optimization.
for (the interior of A), for (the boundary of A) and positive elsewhere.
The main properties of function are gathered in the following theorem  :
1) if and then is real valued;
2) is 1-Lipschitzian;
3) , , , and ,
where the notation denotes the frontier of the set A and the complementary of set A.
4) if A is closed, then it holds ;
5) if A is convex, then is convex;
6) if A is a cone, then is positively homogeneous;
7) if A is a closed convex cone, then is non increasing with respect to the ordering relation induced by A on , i.e. the following is true:
if A has nonempty interior, then
The oriented distance function , used also to obtain a scalarization of a vector optimization problem   allows to establish a relationship between the well-posedness of the original vector problem and the well posedness of the associate scalar problem. More precisely, in  it is known that one of notions of well-posedness in vector optimization can be rephrased as a suitable well-posedness of a corresponding scalar optimization problem, i.e. is linked to well-posedness of a suitable scalar variational inequality of differential type. The construction of this scalar variational inequality represents on interesting application of the “oriented distance function”.
It has been proved in  that when A is closed, convex, pointed cone, then we have:
where is the positive polar of the cone of A and S the unit sphere in .
The function is used in order to give scalar characterizations of some notions of efficiency for problem . Furthermore, some results characterize pointwise well-posedness of problem through function  . Given a point , it is considered the function:
where denotes the positive polar of C and S the unit sphere in . Clearly
The function is directionally differentiable  and hence it is can consider the directional derivative
and the associated scalar problem: find , such that:
The solutions of problem coincide with the solutions of .
Proposition 6.1: Let K be a convex set. If solves problem for some , then is a solution of . Conversely, if solves , then solves problem .
The scalar problem associated with the vector problem is:
The relations among the solutions of problem and those of problem are refers investigated in  . Here it refers only to the characterization of weak efficient solution.
Proposition 6.2: The point is a weak efficient solution of if and only if is a solution of .
The proof is omitted and for it refer to   , for details.
Also well-posedness of can be linked to that of   .
Proposition 6.3: Let f be a continuous function and let be an efficient solution of . Problem is pointwise well-posed at if and only if problem is Tykhonov well-posed.
The next proposition links the well-posedness of to pointwise well-posedness of . It is need to recall Ekeland’s variational principle  : it say that there is a “nearby point” which actually minimizes a slightly perturbed given functional. More precisely it asserts that a particular optimization problem can be replaced by other which is near the original and has a unique solution  . In fact, often the mathematical model of a phenomenon is so complicated that is necessary to replace it by other model which has a solution “near” the original one.
Proposition 6.4: If is pointwise well-posed at , then problem is pointwise well-posed at .
Proof: By proposition 6.3, it is enough to prove that if is pointwise well-posed at , then problem is Tykhonov well-posed.
In fact, for every and , by Ekeland’s variational principle, there exists such that:
If it is introduced the set
then, it follows that
It get, then, that , there exist x such that and
Since , it follows that and so:
Since as , then is Tykhonov well-posed.
Now, the authors prove that the converse of the previous proposition holds under convexity assumptions, namely it is true if f is C-convex. Before, they need the following Lemma:
Lemma 6.1: If is C-convex function, then the function , is convex .
Proposition 6.5: Let f be C-convex and assume is pointwise well posed at . ( is an efficient solution). Then is pointwise well-posed at .
Assuming, ab absurdo, that is not pointwise well-posed at , it follows that exist and , with and one can find some , with .
Without loss of generality, it is possible to put . Since is convex, it follows that:
where . The boundedness of implies that it is can assume (here it is need K closed). Further, since ,
from the continuity of , it is possible to obtain
The last inequality follows from the convexity of  .
Sending n to we obtain which contradicts Tykhonov well-posedness by Proposition 6.3. So, the thesis is true.
7. Concluding Remarks and Future Perspectives for Research
In this paper, the authors have reviewed and studied some properties of well-posedness, a field that has attracted attentions of many researchers for various types of problems and that requests intellectual endeavours. In reality, almost all the literature deals with directly specific notions of well-posedness but there is no general research to the relations between them for different problems and therefore is much needed the research, mostly in this area, to develop and to foster new and innovative applications in various branches of pure and applied sciences. The authors have given only a brief review of this fast growing field and hope that the general theories and results surveyed in this paper can be used to formulate and to outline some connections with other mathematical fields.