Demonstration of the Dyna Reinforcement Learning Framework for Reactive Close Proximity Operations
dc.contributor.author | Majumdar, Ritwik | |
dc.contributor.author | Sternberg, David | |
dc.contributor.author | Albee, Keenan | |
dc.contributor.author | Jia-Richards, Oliver | |
dc.date.accessioned | 2025-01-06T14:54:42Z | |
dc.date.available | 2025-01-06T14:54:42Z | |
dc.date.issued | 2025-01 | |
dc.identifier.citation | AIAA SciTech Forum, AIAA 2025-1002, Orlando, FL, USA, 2025 | en_US |
dc.identifier.uri | https://hdl.handle.net/2027.42/195994 | en |
dc.description.abstract | Lessons from the International Space Station (ISS) emphasize the necessity of exterior inspection for anomaly detection and maintenance, but current methods rely on costly and limited human extravehicular activities and robotic arms. Deployable free-flying small spacecraft offer a flexible, autonomous solution, capable of comprehensive exterior inspections without human involvement. However, the safety of these spacecraft during close proximity operations remains a concern, particularly given uncertain variability in thruster performance. This paper presents SmallSat Steward, a reactive and integrated architecture for online model learning and trajectory planning based on the Dyna reinforcement learning architecture. By combining model-based planning and direct reinforcement learning, Dyna offers a potentially flexible and computationally efficient solution capable of adapting to changes in thruster performance and other system uncertainties. Preliminary results in both simulation and hardware environments demonstrate the potential of this architecture to successfully regulate position under single and double thruster failures. In simulation, the Dyna-based controller outperformed a PD-LQR controller in $\sim$70\% of all cases. On hardware, Dyna was able to eliminate the steady state error caused by thruster failures. | en_US |
dc.description.sponsorship | NASA University SmallSat Technology Partnerships (80NSSC23M0237) | en_US |
dc.language.iso | en_US | en_US |
dc.publisher | American Institute of Aeronautics and Astronautics | en_US |
dc.title | Demonstration of the Dyna Reinforcement Learning Framework for Reactive Close Proximity Operations | en_US |
dc.type | Conference Paper | en_US |
dc.subject.hlbsecondlevel | Aerospace Engineering | |
dc.subject.hlbtoplevel | Engineering | |
dc.contributor.affiliationum | Aerospace Engineering, Department of | en_US |
dc.contributor.affiliationother | Jet Propulsion Laboratory, California Institute of Technology | en_US |
dc.contributor.affiliationumcampus | Ann Arbor | en_US |
dc.description.bitstreamurl | http://deepblue.lib.umich.edu/bitstream/2027.42/195994/1/10.2514:6.2025-1002.pdf | |
dc.identifier.doi | 10.2514/6.2025-1002 | |
dc.identifier.doi | https://dx.doi.org/10.7302/24930 | |
dc.identifier.source | AIAA SciTech Forum | en_US |
dc.description.filedescription | Description of 10.2514:6.2025-1002.pdf : Main article | |
dc.description.depositor | SELF | en_US |
dc.working.doi | 10.7302/24930 | en_US |
dc.owningcollname | Aerospace Engineering, Department of |
Files in this item
Remediation of Harmful Language
The University of Michigan Library aims to describe its collections in a way that respects the people and communities who create, use, and are represented in them. We encourage you to Contact Us anonymously if you encounter harmful or problematic language in catalog records or finding aids. More information about our policies and practices is available at Remediation of Harmful Language.
Accessibility
If you are unable to use this file in its current format, please select the Contact Us link and we can modify it to make it more accessible to you.