Obsidian/Zettelkasten/Permanent Notes/Literature Notes/to_process/Runtime Safety Assurance Using Reinforcement Learning-Note.md
Dane Sabo 8824324149 Auto sync: 2025-08-26 11:21:28 (90 files changed)
R  "Zettelkasten/Literature Notes/.archive/CH4System_Representation_S2020pdf2254.md" -> ".archive/Literature Notes/.archive/CH4System_Representation_S2020pdf2254.md"

R  "Zettelkasten/Literature Notes/.archive/IntroductionDiffusionModels2022.md" -> ".archive/Literature Notes/.archive/IntroductionDiffusionModels2022.md"

R  "Zettelkasten/Literature Notes/.archive/Kry10TechnicalOverview.md" -> ".archive/Literature Notes/.archive/Kry10TechnicalOverview.md"

R  "Zettelkasten/Literature Notes/.archive/ME2046_Sampled_Data_Analysis_Reading_Chapter_2pdf2254ME.md" -> ".archive/Literature Notes/.archive/ME2046_Sampled_Data_Analysis_Reading_Chapter_2pdf2254ME.md"

R  "Zettelkasten/Literature Notes/.archive/ME2046_The_z_transform_Chapter_3pdf2254ME.md" -> ".archive/Literature Notes/.archive/ME2046_The_z_transform_Chapter_3pdf2254ME.md"

R  "Zettelkasten/Literature Notes/.archive/My Library.bib" -> ".archive/Literature Notes/.archive/My Library.bib"

R  "Zettelkasten/Literature Notes/.archive/aModeladoNucleoAnalisis2023.md" -> ".archive/Literature Notes/.archive/aModeladoNucleoAnalisis2023.md"

R  "Zettelkasten/Literature Notes/.archive/atsumiModifiedBodePlots2012.md" -> ".archive/Literature Notes/.archive/atsumiModifiedBodePlots2012.md"
2025-08-26 11:21:28 -04:00

36 lines
1.0 KiB
Markdown

# First Pass
**Category:**
Simulation paper that talks about using a recovery controller in combination
with a nominal controller that is learned. The boundary to swtich between
the two is a learned boundary.
**Context:**
Drones
**Correctness:**
Not very. They do really well in the intro and methodology, but shit hits the fan
when it comes to the results. They don't explain things well, and also don't
really establish why their learned boundary between nominal and recovery
controllers should be successful.
**Contributions:**
The biggest contributions these guys make is demonstrating feasibility of their
reinforcement learned switching from the recovery controller and the nominal
controller.
**Clarity:**
Well written until the whole thing came apart at the end.
# Second Pass
I read this a second time but I don't think it's worth it.
Their main contribution is trying to use RL to learn when to switch to a
recovery controller.
# Third Pass
**Recreation Notes:**
**Hidden Findings:**
**Weak Points? Strong Points?**