vault backup: 2025-05-12 11:19:21

This commit is contained in:
Dane Sabo 2025-05-12 11:19:21 -04:00
parent e6d95a8972
commit a6e088240a
9 changed files with 303 additions and 3 deletions

View File

@ -232,7 +232,29 @@
"distributed-systems": 220,
"PRET-machines": 221,
"real-time-systems": 222,
"time-synchronization": 223
"time-synchronization": 223,
"Computer-networks": 224,
"Concurrent-computing": 225,
"Convergence-of-numerical-methods": 226,
"Electrical-equipment-industry": 227,
"Industrial-control": 228,
"Proposals": 229,
"Taxonomy": 230,
"Neural-networks-Computer-science": 231,
"Nonlinear-control-theory": 232,
"General": 233,
"Adaptive-control": 234,
"Control": 235,
"artificial-intelligence": 236,
"complexity": 237,
"Control-Engineering": 238,
"development": 239,
"genetic-algorithms": 240,
"Identification": 241,
"model": 242,
"Modelling": 243,
"Nonlinear-control": 244,
"Wavelets": 245
},
"_version": 3
}

View File

@ -0,0 +1,62 @@
---
authors:
- "Agarwal, M."
citekey: "agarwalSystematicClassificationNeuralnetworkbased1997"
publish_date: 1997-04-01
journal: "IEEE Control Systems Magazine"
volume: 17
issue: 2
pages: 75-93
last_import: 2025-05-12
---
# Indexing Information
Published: 1997-04
**DOI**
[10.1109/37.581297](https://doi.org/10.1109/37.581297)
#Control-systems, #Stability-analysis, #Computer-networks, #Concurrent-computing, #Convergence-of-numerical-methods, #Electrical-equipment-industry, #Industrial-control, #Neural-networks, #Proposals, #Taxonomy
#ToRead
>[!Abstract]
>Successful industrial applications and favorable comparisons with conventional alternatives have motivated the development of a large number of schemes for neural-network-based control. Each scheme is usually composed of several independent functional features, which makes it difficult to identify precisely what is new in the scheme. Help from available overviews is therefore often inadequate, since they usually discuss only the most important overall schemes. This work breaks the available schemes down to their essential functional features and organizes the latter into a multi-level classification. The classification reveals that similar schemes often get placed in different categories, fundamentally different features often get lumped into a single category, and proposed new schemes are often merely permutations and combinations of the well-established fundamental features. The classification has two main sections: neural network only as an aid; and neural network as controller.>[!seealso] Related Papers
>
# Annotations
## Notes
![[Paper Notes/A systematic classification of neural-network-based control.md]]
## Highlights From Zotero
>[!highlight] Highlight
> rigorous comparisons neural-network controllers have fared better than well- established conventional options when the plant characteristics are poorly known [2-61.
> 2025-04-15 9:18 am
>[!done] Important
> In order to illus- tr te the unavoidable basic terminology for the unfamiliar re 9 der, a neural network can be regarded simply as a generic
> 2025-04-15 9:22 am
>[!highlight] Highlight
> mapping,
> 2025-04-17 4:06 pm
>[!highlight] Highlight
> d also for classifi- cation and optimization tasks. An overview of the proposed classification is shown in Fig. 1. The relatively limited option of using neural networks to merely aid a non-neural controller is further classified in the following section. Of the schemes in which the con
> 2025-04-17 4:06 pm
>[!highlight] Highlight
> Of the schemes in which the controller itself is a neural network, the section “Train Based on U” classifies the alternative where control-input signals U are available for training the neural controller and the section “Train Based on Goal”classifies the option where the network devises the needed control strategy on its own, based on the ultimate control objective. C
> 2025-04-17 1:02 pm
## Follow-Ups

View File

@ -0,0 +1,39 @@
---
authors:
- "Poznyak, Alexander S."
- "Yu, Wen"
- "Sanchez, Edgar N."
citekey: "poznyakDifferentialNeuralNetworks2001"
publish_date: 2001-01-01
publisher: "World Scientific"
location: "Singapore ;"
last_import: 2025-05-12
---
# Indexing Information
Published: 2001-01
**ISBN**
[9786611956738](https://www.isbnsearch.org/isbn/9786611956738)
#Neural-networks-Computer-science, #Nonlinear-control-theory
#ToRead
>[!Abstract]
>This book deals with continuous time dynamic neural networks theory applied to the solution of basic problems in robust control theory, including identification, state space estimation (based on neuro-observers) and trajectory tracking. The plants to be identified and controlled are assumed to be a priori unknown but belonging to a given class containing internal unmodelled dynamics and external perturbations as well. The error stability analysis and the corresponding error bounds for different problems are presented. The effectiveness of the suggested approach is illustrated by its ap>[!seealso] Related Papers
>
# Annotations
## Notes
![[Paper Notes/Differential neural networks for robust nonlinear control- identification, state estimation and trajectory tracking.md]]
## Highlights From Zotero
## Follow-Ups

View File

@ -0,0 +1,44 @@
---
authors:
- "Emami, Seyyed Ali"
- "Castaldi, Paolo"
- "Banazadeh, Afshin"
citekey: "emamiNeuralNetworkbasedFlight2022"
publish_date: 2022-01-01
journal: "Annual Reviews in Control"
volume: 53
pages: 97-137
last_import: 2025-05-12
---
# Indexing Information
Published: 2022-01
**DOI**
[10.1016/j.arcontrol.2022.04.006](https://doi.org/10.1016/j.arcontrol.2022.04.006)
#Neural-networks, #Flight-control, #Intelligent-control, #Reinforcement-learning
#ToRead
>[!Abstract]
>As the first review in this field, this paper presents an in-depth mathematical view of Intelligent Flight Control Systems (IFCSs), particularly those based on artificial neural networks. The rapid evolution of IFCSs in the last two decades in both the methodological and technical aspects necessitates a comprehensive view of them to better demonstrate the current stage and the crucial remaining steps towards developing a truly intelligent flight management unit. To this end, in this paper, we will provide a detailed mathematical view of Neural Network (NN)-based flight control systems and the challenging problems that still remain. The paper will cover both the model-based and model-free IFCSs. The model-based methods consist of the basic feedback error learning scheme, the pseudocontrol strategy, and the neural backstepping method. Besides, different approaches to analyze the closed-loop stability in IFCSs, their requirements, and their limitations will be discussed in detail. Various supplementary features, which can be integrated with a basic IFCS such as the fault-tolerance capability, the consideration of system constraints, and the combination of NNs with other robust and adaptive elements like disturbance observers, would be covered, as well. On the other hand, concerning model-free flight controllers, both the indirect and direct adaptive control systems including indirect adaptive control using NN-based system identification, the approximate dynamic programming using NN, and the reinforcement learning-based adaptive optimal control will be carefully addressed. Finally, by demonstrating a well-organized view of the current stage in the development of IFCSs, the challenging issues, which are critical to be addressed in the future, are thoroughly identified. As a result, this paper can be considered as a comprehensive road map for all researchers interested in the design and development of intelligent control systems, particularly in the field of aerospace applications.>[!seealso] Related Papers
>
# Annotations
## Notes
![[Paper Notes/Neural network-based flight control systems- Present and future.md]]
## Highlights From Zotero
>[!highlight] Highlight
> Although the words Intelligence and Autonomy have been widely employed interchangeably, there is an essential conceptual difference between them [1]. Different definitions have been given for both concepts in the literature [2, 3]. However, in a general view, the intelligence may be defined as a very general mental capability that involves the ability to reason, plan, solve problems, think abstractly, comprehend complex ideas, learn quickly and learn from experience [3]. On the other hand, the ability to generate ones own purposes without any instruction from outside can be interpreted as the autonomy of a system [1]
> 2025-04-07 12:54 pm
## Follow-Ups

View File

@ -0,0 +1,38 @@
---
authors:
- "Liu, G. P."
- "Grimble, Michael J."
- "Johnson, Michael A."
citekey: "liuNonlinearIdentificationControl2001"
publish_date: 2001-01-01
publisher: "Springer"
location: "London"
last_import: 2025-05-12
---
# Indexing Information
Published: 2001-01
**ISBN**
[978-1-4471-1076-7 978-1-4471-0345-5](https://www.isbnsearch.org/isbn/978-1-4471-1076-7 978-1-4471-0345-5)
#Adaptive-control, #Control, #artificial-intelligence, #complexity, #Control-Engineering, #development, #genetic-algorithms, #Identification, #learning, #model, #Modelling, #Neural-Networks, #Nonlinear-control, #Wavelets
#ToRead
>[!seealso] Related Papers
>
# Annotations
## Notes
![[Paper Notes/Nonlinear Identification and Control.md]]
## Highlights From Zotero
## Follow-Ups

View File

@ -0,0 +1,43 @@
---
authors:
- "Carr, Steven"
- "Jansen, Nils"
- "Junges, Sebastian"
- "Topcu, Ufuk"
citekey: "carrSafeReinforcementLearning2023"
publish_date: 2023-06-26
journal: "Proceedings of the AAAI Conference on Artificial Intelligence"
volume: 37
issue: 12
pages: 14748-14756
last_import: 2025-05-12
---
# Indexing Information
Published: 2023-06
**DOI**
[10.1609/aaai.v37i12.26723](https://doi.org/10.1609/aaai.v37i12.26723)
#General
#ToRead
>[!Abstract]
>Safe exploration is a common problem in reinforcement learning (RL) that aims to prevent agents from making disastrous decisions while exploring their environment. A family of approaches to this problem assume domain knowledge in the form of a (partial) model of this environment to decide upon the safety of an action. A so-called shield forces the RL agent to select only safe actions. However, for adoption in various applications, one must look beyond enforcing safety and also ensure the applicability of RL with good performance. We extend the applicability of shields via tight integration with state-of-the-art deep RL, and provide an extensive, empirical study in challenging, sparse-reward environments under partial observability. We show that a carefully integrated shield ensures safety and can improve the convergence rate and final performance of RL agents. We furthermore show that a shield can be used to bootstrap state-of-the-art RL agents: they remain safe after initial learning in a shielded setting, allowing us to disable a potentially too conservative shield eventually.>[!seealso] Related Papers
>
# Annotations
## Notes
![[Paper Notes/Safe Reinforcement Learning via Shielding under Partial Observability.md]]
## Highlights From Zotero
## Follow-Ups

View File

@ -0,0 +1,46 @@
---
authors:
- "Alshiekh, Mohammed"
- "Bloem, Roderick"
- "Ehlers, Rüdiger"
- "Könighofer, Bettina"
- "Niekum, Scott"
- "Topcu, Ufuk"
citekey: "alshiekhSafeReinforcementLearning2018"
publish_date: 2018-04-29
journal: "Proceedings of the AAAI Conference on Artificial Intelligence"
volume: 32
issue: 1
last_import: 2025-05-12
---
# Indexing Information
Published: 2018-04
**DOI**
[10.1609/aaai.v32i1.11797](https://doi.org/10.1609/aaai.v32i1.11797)
#Formal-Methods
#ToRead
>[!Abstract]
>Reinforcement learning algorithms discover policies that maximize reward, but do not necessarily guarantee safety during learning or execution phases. We introduce a new approach to learn optimal policies while enforcing properties expressed in temporal logic. To this end, given the temporal logic specification that is to be obeyed by the learning system, we propose to synthesize a reactive system called a shield. The shield monitors the actions from the learner and corrects them only if the chosen action causes a violation of the specification. We discuss which requirements a shield must meet to preserve the convergence guarantees of the learner. Finally, we demonstrate the versatility of our approach on several challenging reinforcement learning scenarios.>[!seealso] Related Papers
>
# Annotations
## Notes
![[Paper Notes/Safe Reinforcement Learning via Shielding.md]]
## Highlights From Zotero
## Follow-Ups

Binary file not shown.

After

Width:  |  Height:  |  Size: 178 KiB

View File

@ -21,8 +21,14 @@
{"id":"bbe4a58a5af748eb","type":"text","text":"OLDER","x":-1000,"y":1600,"width":250,"height":60},
{"id":"b081d5b64b515d56","type":"text","text":"2017","x":-1000,"y":890,"width":250,"height":60},
{"id":"2e49adbdc97ce1c2","type":"text","text":"2015","x":-1000,"y":1260,"width":250,"height":60},
{"id":"6b62b3a19b06fc0a","type":"file","file":"Literature Notes/A Review of Formal Methods applied to Machine Learning.md","x":-100,"y":210,"width":420,"height":400,"color":"3"},
{"id":"bac68ab8fa2c1c9b","type":"file","file":"Literature Notes/Formal verification of neural network controlled autonomous systems.md","x":-720,"y":530,"width":485,"height":360,"color":"3"}
{"id":"6b62b3a19b06fc0a","type":"file","file":"Literature Notes/A Review of Formal Methods applied to Machine Learning.md","x":-200,"y":240,"width":520,"height":370,"color":"3"},
{"id":"bac68ab8fa2c1c9b","type":"file","file":"Literature Notes/Formal verification of neural network controlled autonomous systems.md","x":-720,"y":530,"width":485,"height":360,"color":"3"},
{"id":"7ad1df279923e7f4","x":-720,"y":1660,"width":480,"height":400,"color":"1","type":"file","file":"Literature Notes/A systematic classification of neural-network-based control.md"},
{"id":"b16d77e25162b164","x":-200,"y":1660,"width":400,"height":400,"color":"1","type":"file","file":"Literature Notes/Nonlinear Identification and Control.md"},
{"id":"dae189dd0beeb46d","x":-720,"y":100,"width":480,"height":360,"color":"1","type":"file","file":"Literature Notes/Neural network-based flight control systems - Present and future.md"},
{"id":"8af915dccc8e1757","x":240,"y":1660,"width":400,"height":400,"color":"1","type":"file","file":"Literature Notes/Differential neural networks for robust nonlinear control - identification, state estimation and trajectory tracking.md"},
{"id":"034773fb85df43e9","x":-200,"y":-183,"width":520,"height":363,"color":"1","type":"file","file":"Literature Notes/Safe Reinforcement Learning via Shielding under Partial Observability.md"},
{"id":"1b0680305c662125","x":-200,"y":670,"width":520,"height":330,"color":"1","type":"file","file":"Literature Notes/Safe Reinforcement Learning via Shielding.md"}
],
"edges":[
{"id":"766a32fa431c6398","fromNode":"bcc9c2b5dc701338","fromSide":"bottom","toNode":"bac68ab8fa2c1c9b","toSide":"top","color":"2"},