Finite-time passivity of neutral-type complex-valued neural networks with time-varying delays
View abstract on PubMed
Summary
This summary is machine-generated.This study addresses finite-time passivity in complex-valued neural networks with delays. New criteria ensure finite-time boundedness and passivity, validated by numerical examples.
Area Of Science
- Complex-valued neural networks
- Control theory
- Nonlinear systems
Background
- Neural networks are crucial in AI, but their stability with delays is complex.
- Passivity is key for system stability and energy dissipation analysis.
- Neutral-type systems with time-varying delays present unique analytical challenges.
Purpose Of The Study
- Investigate finite-time passivity for neutral-type complex-valued neural networks.
- Develop sufficient conditions for finite-time boundedness (FTB) and finite-time passivity (FTP).
- Analyze the impact of time-varying delays on network stability and passivity.
Main Methods
- Lyapunov functional approach for stability analysis.
- Wirtinger-type inequality techniques to handle delays.
- Linear matrix inequalities (LMIs) for deriving sufficient conditions.
Main Results
- New sufficient conditions for finite-time boundedness (FTB) were established.
- Novel criteria for finite-time passivity (FTP) were derived.
- The proposed methods effectively analyze the stability of the specified neural network model.
Conclusions
- The derived conditions guarantee finite-time passivity and boundedness for the neural network model.
- The Lyapunov functional and LMI approach provide a robust framework for analyzing such systems.
- Numerical simulations confirm the theoretical findings and the validity of the proposed criteria.
Related Concept Videos
Continuous-time systems have continuous input and output signals, with time measured continuously. These systems are generally defined by differential or algebraic equations. For instance, in an RC circuit, the relationship between input and output voltage is expressed through a differential equation derived from Ohm's law and the capacitor relation,
Discrete-time systems have input and output signals at specific intervals, defined at distinct instants by difference equations. An example...
A system is linear if it displays the characteristics of homogeneity and additivity, together termed the superposition property. This principle is fundamental in all linear systems. Linear time-invariant (LTI) systems include systems with linear elements and constant parameters.
The input-output behavior of an LTI system can be fully defined by its response to an impulsive excitation at its input. Once this impulse response is known, the system's reaction to any other input can be...
Linear systems are characterized by two main properties: superposition and homogeneity. Superposition allows the response to multiple inputs to be the sum of the responses to each individual input. Homogeneity ensures that scaling an input by a scalar results in the response being scaled by the same scalar.
In contrast, nonlinear systems do not inherently possess these properties. However, for small deviations around an operating point, a nonlinear system can often be approximated as linear....
The frequency-domain technique, commonly used in analyzing and designing feedback control systems, is effective for linear, time-invariant systems. However, it falls short when dealing with nonlinear, time-varying, and multiple-input multiple-output systems. The time-domain or state-space approach addresses these limitations by utilizing state variables to construct simultaneous, first-order differential equations, known as state equations, for an nth-order system.
Consider an RLC circuit, a...
First-order systems, such as RC circuits, are foundational in understanding dynamic systems due to their straightforward input-output relationship. Analyzing their responses to different input functions under zero initial conditions reveals significant insights into system behavior.
When a first-order system is subjected to a unit-step input, its response is characterized by its transfer function. By applying the Laplace transform of the unit-step input to the transfer function, expanding the...
Long-term potentiation, or LTP, is one of the ways by which synaptic plasticity—changes in the strength of chemical synapses—can occur in the brain. LTP is the process of synaptic strengthening that occurs over time between pre- and postsynaptic neuronal connections. The synaptic strengthening of LTP works in opposition to the synaptic weakening of long-term depression (LTD) and together are the main mechanisms that underlie learning and memory.
Hebbian LTP
LTP can occur when...

