1 Introduction

The well-known Banach–Picard–Caccioppoli contraction principle states that:

Theorem 1.1

Let T be a self-mapping of a complete metric space (Xd) such that \( d(Tx, Ty) \le ad(x, y), 0\le a<1,\) for each \(x, y \in X\). Then, T has a unique fixed point. Moreover, the Picard iteration \(\{x_n\}\) defined by \( x_{n+1} = Tx_n, (n = 0, 1, 2, \ldots )\) converges to the unique fixed point \( x_* \in X\) for any initial value \(x_0 \in X\).

The mapping T of Banach–Picard–Caccioppoli contraction is continuous in the entire domain of X.

In 1981, Istr\(\check{a}\)tescu Istrǎtescu (1981b) extended the well-known Banach–Picard–Caccioppoli contraction principle by introducing a convexity condition, namely convex contraction mapping of order m, where \(m\ge 2\). Meanwhile, a more complete study (data dependence, well-posedness, Ulam–Hyers stability, limit shadowing property and Ostrowski property) for convex contraction mapping of order 2 was recently proposed in Mureşan and Mureşan (2015) (see also Rus 2016). In Georgescu (2017), Georgescu initiated study of iterated function system consisting of generalized convex contractions.

Definition 1.1

A continuous function \(T:X\rightarrow X\), where (Xd) is a complete metric space is called convex contraction of order m (also called generalized convex contraction Georgescu 2017) if there exist \(m\in \mathbb {N}\) and \( a_{0},a_{1},\ldots ,a_{m-1}\ge 0\) such that \(\sum _{j=0}^{m-1}a_{j}<1\) and

$$\begin{aligned} d\left( T^{m}x,T^{m}y\right) \le \overset{m-1}{\underset{j=0}{\sum }} a_{j}d\left( T^{j}x,T^{j}y\right) \text {,} \end{aligned}$$
(1.1)

for each \(x,y\in X\), where by \(T^{j}\), we mean the composition of T by itself j times.

Theorem 1.2

Let T be a continuous self-mapping of a complete metric space (Xd) satisfying (1.1). Then, T has a unique fixed point.

It is easy to observe that convex contraction of order \(m-1\) is convex contraction of order m but converse need not be true.

Example 1.1

Let \(X=\mathbb {N}\cup \{0\}\) and d be the usual metric on X. Define \( T:X\rightarrow X\) by

$$\begin{aligned} Tx=\left\{ \begin{array}{ll} 0 &{}\quad \text {if} \; x=0,2\;\text { and }\;x=3k,k\;\in \; \mathbb {N} \\ 1 &{}\quad \text {if} \; x=3k+1 \\ 2 &{}\quad \text {if} \; x=1,3k+2 \end{array} \right. . \end{aligned}$$

Then, T satisfies the condition \(d(T^{3}x,T^{3}y)\le \sum _{j=0}^{2}a_{j}d(T^{j}x,T^{j}y),\) for all \(x,y\in X\), but \( d(T^{2}x,T^{2}y)>\sum _{j=0}^{1}a_{j}d(T^{j}x,T^{j}y),\)(take \(x=4,y=5\) Sastry et al. 2012).

Recall that the set \(O(x;T) = \{T^{n}x : n = 0, 1, 2,\ldots \}\) is called the orbit of the self-mapping T at the point \(x \in X\).

Definition 1.2

A self-mapping T of a metric space (Xd) is called orbitally continuous at a point \(z\in X\) if for any sequence \(\{x_{n}\}\subset O(x;T)\) (for some \(x\in X)\) \(x_{n}\rightarrow z\) implies \(Tx_{n}\rightarrow Tz\) as \( n\rightarrow \infty \).

Every continuous self-mapping of a metric space is orbitally continuous, but the converse need not be true. The following example illustrates this fact.

Example 1.2

Let \(X=[0,1]\) and d be the usual metric on X. Define \(T:X\rightarrow X\) by

$$\begin{aligned} Tx=\left\{ \begin{array}{ll} 1/3 &{}\quad \text {if} \; x\in \left[ 0,1/3\right] \\ 0 &{}\quad \text {if} \; x\in \left( 1/3,1\right] \end{array} \right. . \end{aligned}$$

Then, T is orbitally continuous. However, T is not continuous at \(x=1/3\).

Definition 1.3

A self-mapping T of a metric space (Xd) is called k-continuous Pant and Pant (2017), \(k = 1, 2, 3,\ldots ,\) if \(T^{k}x_n \rightarrow Tz\), whenever \(\{x_n\}\) is a sequence in X such that \(T^{k-1}x_n \rightarrow z\).

Remark 1.1

It is important to note that for a self-mapping T of a metric space (Xd), the notion of 1-continuity coincides with continuity. However,

1-continuity \(\Rightarrow \) 2-continuity \(\Rightarrow \) 3-continuity \( \Rightarrow \cdots ,\)

but not conversely. In Example 1.2, one may observe that \(Tx_n \rightarrow t \Rightarrow T^{2}x_n \rightarrow Tt \) since \(Tx_n \rightarrow t\) implies \( t=0 \) or \(t=1/3\) and \(T^{2}x_{n} \rightarrow 1/3 =T1/3\) for all n. Hence, T is 2-continuous. However, T is not 1-continuous at \(x=1/3\).

The study of contractive conditions which does not force the mappings to be continuous at the fixed point is presently an active area of research. The question whether there exists a contractive definition which is strong enough to generate a fixed point but which does not force the mapping to be continuous at the fixed point was reiterated by Rhoades (1988) as an existing open problem. Few answers to the open question have been given in Bisht and Pant (2017), Bisht (2017), Bisht and Rakočević (2018), Pant (1999), Pant et al. (2019, 2020), and Pant and Pant (2017).

In this paper, we present new fixed point theorems using the notions of orbital continuity or k-continuity of self-mappings on a complete metric space. Our results provide new solutions to Rhoades open question and have significance in both theory and application. It is well known that some fixed point results such as Brouwer’s fixed point theorem, Banach fixed point theorem, Kakutani’s fixed point theorem and Krasnoselskii fixed point theorem, have been used in the theoretical studies of neural networks. Activation functions are basically used in a neural network to transform an input signal into an output signal. A suitable and more generalized activation function can greatly improve the performance of neural networks (Zhang et al. 2014). For example, structures of the activation functions have an important influence on the multiple stability of the neural networks as the different structure of activation function will affect the number of equilibria and the number of stable equilibria of neural networks (Huang et al. 2019b). Continuity or discontinuity of an activation function is also an important aspect for the stability issue of neural networks (for more details see Huang et al. 2012, 2019a; Nie and Zheng 2014; Nie et al. 2019a, b; Zhang et al. 2014 and the references therein). Especially, the network model with discontinuous activation function has a wide range of applications in real-life problems ranging from image compression to character recognition, stock market prediction, etc., (for some examples see Huang et al. 2019b, 2012; Nie and Cao 2012; Nie and Zheng 2014). Forti and Nistri (2003) initiated the study of discontinuous-type neural networks to solve non-negative sparse approximation problems to model the mammalian olfactory system. Therefore, designing a new neural network with a more generalized activation function is very significant. We think that new answers of the Rhoades open question and theoretical fixed point results can contribute to design of several kind neural networks under suitable conditions. In a recent work, possible applications of the notion of k-continuity of self-mappings to the theory of neural networks was stated as an open problem in Bisht and Özgür (2020).

2 Main results

Our first main result is the following:

Theorem 2.1

Let T be a self-mapping of a complete metric space (Xd) such that for each \(x,y\in X\);

$$\begin{aligned} d\left( T^{m}x,T^{m}y\right) \le \overset{m-1}{\underset{j=0}{\sum }} a_{j}d\left( T^{j}x,T^{j}y\right) \text {,} \end{aligned}$$
(2.1)

where \(a_{0},a_{1},\ldots ,a_{m-1}\ge 0\) such that \(\sum _{j=0}^{m-1}a_{j}<1\). Suppose T is orbitally continuous or T is k-continuous. Then, T has a unique fixed point.

Proof

Let \(x_{0}\) be any point in X. If \(Tx_{0}=x_{0}\) then we are done, hence assume \(Tx_{0}\ne x_{0}\). Define a sequence \(\{x_{n}\}\) in X given by the rule \(x_{n+1}=T^{n}x_{0}=Tx_{n}\) for all \(n\in \mathbb {N}\cup \{0\}\). Set

$$\begin{aligned} M=d\left( x,y\right) +d\left( Tx,Ty\right) +\cdots +d\left( T^{m-1}x,T^{m-1}y\right) \quad \text { and }\quad \alpha =\sum _{j=0}^{m-1}a_{j}<1. \end{aligned}$$
(2.2)

Now we claim that for all \(n\ge m\),

$$\begin{aligned} d\left( T^{n}x_{0},T^{n+1}x_{0}\right) \le \alpha ^{[\frac{n}{m}]}M \text {,} \end{aligned}$$
(2.3)

where \([\frac{n}{m}]\) is the integral part of \(\frac{n}{m}\).

Obviously, (2.3) holds for \(n=m\). Assuming (2.3) is true for some n, that is,

$$\begin{aligned} d\left( T^{n}x_{0},T^{n+1}x_{0}\right) \le \alpha ^{[\frac{n}{m}]}M\quad \text { for all }n\ge m\ge 2\text {.} \end{aligned}$$
(2.4)

Consider

$$\begin{aligned} d\left( T^{n+1}x_{0},T^{n+2}x_{0}\right)\le & {} a_{0}d\left( T^{n-m+1}x_{0},T^{n-m+2}x_{0}\right) +a_{1}d\left( T^{n-m+2}x_{0},T^{n-m+3}x_{0}\right) +\cdots \\&+a_{n-1}d\left( T^{n}x_{0},T^{n+1}x_{0}\right) \\\le & {} a_{0}\alpha ^{[\frac{n-m+1}{m}]}M+a_{1}\alpha ^{[\frac{ n-m+2}{m}]}M+\cdots +a_{n-1}\alpha ^{[\frac{n}{m}]}M \\\le & {} (a_{0}+a_{1}+\cdots +a_{n-1})\alpha ^{[\frac{n-m+1}{m}]}M \\\le & {} \alpha ^{[\frac{n-m+1}{m}]+1}M=\alpha ^{[\frac{n+1}{m} ]}M. \end{aligned}$$

Thus, induction argument shows that

$$\begin{aligned} d\left( T^{n}x_{0},T^{n+1}x_{0}\right) \le \alpha ^{[\frac{n}{m}]}M\quad \text { for all }n\ge m\ge 2\text {.} \end{aligned}$$
(2.5)

We shall now show that \(\{x_{n}\}\) is a Cauchy sequence. For \(p<n\) and \( n=p+k,n>p\ge m\), we have

$$\begin{aligned} d(T^{p}x_{0},T^{p+k}x_{0})\le & {} d(T^{p}x_{0},T^{p+1}x_{0})+d(T^{p+1}x_{0},T^{p+2}x_{0})+\cdots +d(T^{p+k-1}x_{0},T^{p+k}x_{0})\\\le & {} \alpha ^{[\frac{p}{m}]}M{+}\alpha ^{[\frac{p+1}{m} ]}M{+}\cdots {+}\alpha ^{[\frac{p+m-1}{m}]}M{+}\alpha ^{[\frac{p+m}{m} ]}M{+}\cdots {+}\alpha ^{[\frac{p+k-1}{m}]}M \\\le & {} m\alpha ^{[\frac{p}{m}]}M{+}m\alpha ^{[\frac{p}{m} ]+1}M{+}\cdots {+}m\alpha ^{[\frac{p}{m}]{+}[\frac{k-1}{m}]}M \end{aligned}$$

(since the first m terms in the sum are equal and the next m terms are equal and so on)

$$\begin{aligned}\le & {} \left( 1+\alpha +\alpha ^{2}+\cdots +\alpha ^{[\frac{k-1}{m}]}\right) m\alpha ^{[\frac{p}{m}]}M \\ \frac{m\alpha ^{[\frac{p}{m}]}}{1-\alpha }M\rightarrow & {} 0\text { as } p\rightarrow \infty \text {.} \end{aligned}$$

From these estimates, we conclude that \(\{x_{n}\}\) is a Cauchy sequence. Since X is complete, there exists a point \(z\in X\) such that \( x_{n}\rightarrow z\) as \(n\rightarrow \infty \). Orbital continuity of T implies that \(lim_{n\rightarrow \infty }Tx_{n}=Tz.\) This yields \(Tz=z,\) that is, z is a fixed point of T. Uniqueness of the fixed point follows easily.

Now suppose that T is k-continuous. Since \(T^{k-1}x_{n}\rightarrow z\), k-continuity of T implies that \(\mathrm{lim}_{n\rightarrow \infty }T^{k}x_{n}=Tz\) . This yields \(z=Tz,\) that is, z is a fixed point of T. \(\square \)

We now give an example to show that convex contraction mapping of order 3 is strong enough to generate a fixed point but may to be continuous at the fixed point.

Example 2.1

Let \(X=[-2,4]\) and d be the usual metric on X. Define \(T:X\rightarrow X\) by

$$\begin{aligned} Tx=\left\{ \begin{array}{ll} 1/2 &{} \quad \text {if} \; -2\le x<0 \\ 1 &{}\quad \text {if} \; 0\le x\le 1 \\ 0 &{}\quad \text {if} \; 1<x\le 2 \\ -3/2 &{}\quad \text {if} \; 2<x\le 4 \end{array} \right. . \end{aligned}$$

Then, T satisfies all the conditions of Theorem 2.1\((m\ge 3)\) and has a unique fixed point \(x=1\) at which T is discontinuous. It may be noted that T is orbitally continuous.

The following example depicts that Theorem 2.1 is not valid if we replace the convex coefficients \(a_{0},a_{1},\ldots ,a_{m-1}\) by their sum \( \alpha =\sum _{j=0}^{m-1}a_{j}\).

Example 2.2

Let \(X=[0,\infty )\) and d be the usual metric on X. Define \( T:X\rightarrow X\) by

$$\begin{aligned} Tx=x+k,\quad k\in \mathbb {N}. \end{aligned}$$

Then, T satisfies

$$\begin{aligned} d(T^{m}x,T^{m}y)\le \alpha \left[ d(x,y)+d(Tx,Ty)+\cdots +d(T^{m-1}x,T^{m-1}y) \right] \text { for }\alpha =1/m, \end{aligned}$$

but T is fixed point free.

Putting \(m=2\) in Theorem 2.1, we get the following corollary which is an extended version of Istr\(\check{a}\)tescu’s result for convex contraction mapping of order 2.

Corollary 2.1

Let T be a self-mapping of a complete metric space (Xd) such that for each \(x, y \in X\);

$$\begin{aligned} d(T^{2}x,T^{2}y)\le a_{0}d(x,y)+a_{1}d(Tx,Ty)\text {,} \end{aligned}$$
(2.6)

where \(a_{0}, a_{1} \ge 0\) such that \(a_{0}+ a_{1} <1\). Suppose T is orbitally continuous or T is k-continuous. Then, T has a unique fixed point.

In the next theorem, we replace condition (2.1) of Theorem 2.1 by a more general condition, which does not imply continuity at the fixed point.

Definition 2.1

A function \(\varphi : [0,\infty ) \rightarrow [0,\infty )\) is called a comparison function if:

(i):

\(\varphi \) is increasing;

(i):

\(lim_{n\rightarrow \infty }\varphi ^{n}(t)=0\) for every \( t\in [0,\infty )\).

It is well known that a comparison function \(\varphi \) has the property that \( \varphi (0) = 0\) and \(\varphi (t) < t\) for every \(t \in (0,\infty )\).

Theorem 2.2

Let T be a self-mapping of a complete metric space (Xd) such that for each \(x,y\in X\);

$$\begin{aligned} d(T^{m}x,T^{m}y)\le \varphi \left\{ \max _{\left( j=0,1,\ldots ,m-1\right) }d(T^{j}x,T^{j}y)\right\} , \end{aligned}$$
(2.7)

where \(\varphi \) is a comparison function and \(m\in \mathbb {N}\). Suppose that T is k-continuous. Then, T has a unique fixed point.

Proof

Let \(x_{0}\) be any point in X. If \(Tx_{0}=x_{0}\) then we are done, hence assume \(Tx_{0}\ne x_{0}\). Define a sequence \(\{x_{n}\}\) in X given by the rule \(x_{n+1}=T^{n}x_{0}=Tx_{n}\) for all \(n\in \mathbb {N}\cup \{0\}\). If \( x_{n}=x_{n+1}\) for some n then \(x_{n}=x_{n+1}=x_{n+2}=\cdots \), i.e., \( \{x_{n}\} = \{T^{n}x_{0}\}\) is a Cauchy sequence and \(x_{n}\) is a fixed point of T. We can, therefore, assume that \(x_{n} \ne x_{n+1}\) for each n. Then, using the argument employed by Miculescu and Mihail (2017) (for constant \(s=1\)), it follows that \(\{x_{n}\}\) is a Cauchy sequence. Rest of the proof follows from the proof of Theorem 2.1. \(\square \)

The following example illustrates Theorem 2.2.

Example 2.3

Let \(X=[0,2]\) and d be the usual metric on X. Define \(T:X\rightarrow X\) by

$$\begin{aligned} Tx=\left\{ \begin{array}{ll} 1 &{}\quad \text {if} \; 0\le x\le 1 \\ 0 &{}\quad \text {if} \; 1<x\le 2 \end{array} \right. . \end{aligned}$$

Then, T satisfies all the conditions of Theorem 2.2 and has a unique fixed point \(x=1\). The mapping T satisfies the condition (2.7) for \(m\ge 2\) with

$$\begin{aligned} \varphi \left( t\right) =\left\{ \begin{array}{ll} 1 &{}\quad \text {if} \; t>1 \\ \frac{t}{2} &{}\quad \text {if} \; t\le 1 \end{array} \right. \end{aligned}$$

and \(lim_{n\rightarrow \infty }\varphi ^{n}(t)=0\) for every \(t\in [0,\infty )\). It can also be easily seen that T is discontinuous at the fixed point \(x=1\).

Remark 2.1

Theorem 2.2 is an effective generalization of Theorem 2.1 if we put \(\varphi (t)=(\sum _{j=0}^{m-1}a_{j})t\).

In view of a result of Pant and Pant (2017), we now state a new fixed point theorem by considering the restriction \(\lim _{n\rightarrow \infty }\varphi ^{n}(t)=0\) in the open interval (0, d(T(X))) instead of in whole of \([0,\infty )\), where d(T(X)), the diameter of the range of T.

Theorem 2.3

Let T be a self-mapping of a complete metric space (Xd) such that for each \(x,y\in X\) satisfying (2.7), where \(\varphi \) is a comparison function with \(\lim _{n\rightarrow \infty }\varphi ^{n}(t)=0\) for each t in the open interval (0, d(T(X))) and \(m\in \mathbb {N}\). Suppose that T is k-continuous. Then, T has a unique fixed point.

Remark 2.2

Using the notion of k-continuity, a host of theorems proved in Alghamdi et al. (2011), Ghorbanian et al. (2012), Hussain and Salimi (2015), Latif et al. (2015), Latif et al. (2016), Miandaragh et al. (2013), Miculescu and Mihail (2017) and Sastry et al. (2012) can also be extended to a wider class of mappings which need not be continuous at the fixed point.

Remark 2.3

The above theorems generalize and subsume various results due to Istrǎtescu (1981b, 1982), Boyd and Wong (1969), Matkowski (1975), Miculescu and Mihail (2017), Rakotch (1962), Rus (2016, 2001) and Sastry et al. (2012). In addition to it, we provide more answers to the open problem regarding to the existence of contractive mappings which have fixed points but are discontinuous at the fixed point (Rhoades 1988).

3 Examples of discontinuous activation functions

In the previous section, we have proved new fixed point results that require the uniqueness of the fixed point, but do not require the self-mapping to be continuous at the fixed point. Such mappings are frequently appeared in the study of neural networks as activation functions. In Sharma et al. (2020), a brief description of various activation functions used in the study of artificial neural networks was given and explained the need of activation functions. Mainly, activation functions are very important because of their role in learning and making sense of non-linear and complicated mappings between the inputs and corresponding outputs in a neural network (for more details and examples of activation functions see Calin 2020; Sharma et al. 2020 and the references therein). One of the main types of activation functions is the unit step function (or binary step function or Heaviside function). Heaviside function H is defined by

$$\begin{aligned} Hx=\left\{ \begin{array}{ll} 0, &{}\quad x<0 \\ 1, &{}\quad x\ge 0 \end{array} \right. \text {.} \end{aligned}$$

Notice that \(Hx_{n}\rightarrow t\) implies \(H^{2}x_{n}\rightarrow Ht\) since \( Hx_{n}\rightarrow t\) implies \(t=0\) or \(t=1\) and \(H^{2}x_{n}\rightarrow 1=H1\) for all n. Then, H is 2-continuous. It is easy to verify that the unit step function H verifies the conditions of Theorem 2.1 for \(m=2\) and any numbers \(a_{0},a_{1}\ge 0\) with \(a_{0}+a_{1}<1\). H has a unique fixed point \(x=1\) and continuous at its fixed point (see Fig. 1 which is drawn using Mathematica Wolfram Research 2019). This example shows the effectiveness of our results.

Fig. 1
figure 1

The graph of the Heaviside function H

Another example of a discontinuous activation function is the function G defined by

$$\begin{aligned} Gx=\left\{ \begin{array}{ll} x-1, &{}\quad x<0 \\ 0, &{}\quad x\ge 0 \end{array} \right. \text {,} \end{aligned}$$

(see Huang et al. 2013 for more details). G has a unique fixed point \(x=0\) and discontinuous at this fixed point (see Fig. 2). In Huang et al. (2013), the existence of positive periodic solution for a class of delayed neural networks was presented based on a new method involving the application of set-valued version of Krasnoselskii’s fixed point theorem in a cone and the framework of Filippov differential inclusions.

Fig. 2
figure 2

The graph of the discontinuous activation function G

In recent years, the dynamical behavior of neural networks with discontinuous activation functions has been extensively analyzed [(for example see (Du and Xu 2015; Du et al. 2013; Forti and Nistri 2003; Huang et al. 2019a, b, 2012; Kong et al. 2019; Liu et al. 2020, 2014; Nie and Cao 2012; Nie and Zheng 2014; Wang et al. 2010)]. In Huang et al. (2019a), it was stated that one of the main advantages of the incorporation of discontinuous activation functions is the higher storage capacity of neural networks than continuous ones. Now, we consider two more examples of discontinuous activation functions. We see that one of them is continuous at its unique fixed point while the other is discontinuous.

In Liu et al. (2014), finite-time stabilization of neural networks with discontinuous activations via discontinuous controllers was discussed. In numerical examples, the discontinuous activation function S defined by

$$\begin{aligned} Sx=x+\left( 0.2\right) signx=\left\{ \begin{array}{ll} x-0.2, &{}\quad x<0 \\ 0, &{}\quad x=0 \\ x+0.2, &{}\quad x>0 \end{array} \right. \end{aligned}$$

was used. Clearly, the point \(x=0\) is the unique fixed point of S and S is discontinuous at its fixed point (see Fig. 3).

Fig. 3
figure 3

The graph of the discontinuous activation function S

In Huang et al. (2012), the multistability and multiperiodicity issues of 2n -dimensional delayed bidirectional associative memory (BAM) neural networks with r-level discontinuous activation functions were discussed. Sufficient conditions were established to ensure the existence of \(r^{n}\) locally exponentially stable equilibria for 2n-dimensional BAM neural networks with discontinuous activation functions due to Brouwer’s fixed point theorem and stability analysis. In numerical simulations, the following discontinuous activation function T was used:

$$\begin{aligned} Tx=\left\{ \begin{array}{ll} -2, &{}\quad x\in \left( -\infty ,-1\right) \\ -1, &{}\quad x\in \left( -1,1\right) \\ 1, &{}\quad x\in \left( 1,+\infty \right) \end{array} .\right. \end{aligned}$$

Notice that this discontinuous activation function T has the unique fixed point \(x=-2\) and continuous at this fixed point (see Fig. 4).

Fig. 4
figure 4

The graph of the discontinuous activation function T

For more examples, one can see Bisht and Özgür (2020), Cai et al. (2018), Calin (2020), Du and Xu (2015), Du et al. (2013), Forti and Nistri (2003), Huang et al. (2019a, 2019b, 2012), Kong et al. (2019), Liu et al. (2020), Nie and Cao (2012), Nie and Zheng (2014), Nie et al. (2019a, 2019b), Pant et al. (2019, 2020), Wang et al. (2010), Zhang et al. (2014) and the references therein.