{"id":406,"date":"2024-10-18T15:06:28","date_gmt":"2024-10-18T19:06:28","guid":{"rendered":"https:\/\/sites.bu.edu\/learningautonomy\/?page_id=406"},"modified":"2024-11-20T12:06:49","modified_gmt":"2024-11-20T17:06:49","slug":"poster-session","status":"publish","type":"page","link":"https:\/\/sites.bu.edu\/learningautonomy\/poster-session\/","title":{"rendered":"Poster Session"},"content":{"rendered":"<p><strong><span style=\"color: #003366;\"><img loading=\"lazy\" src=\"\/learningautonomy\/files\/2024\/11\/Screenshot-2024-11-15-at-3.56.07\u202fPM-150x150.png\" alt=\"\" width=\"212\" height=\"212\" class=\" wp-image-485 alignleft\" srcset=\"https:\/\/sites.bu.edu\/learningautonomy\/files\/2024\/11\/Screenshot-2024-11-15-at-3.56.07\u202fPM-150x150.png 150w, https:\/\/sites.bu.edu\/learningautonomy\/files\/2024\/11\/Screenshot-2024-11-15-at-3.56.07\u202fPM-550x550.png 550w, https:\/\/sites.bu.edu\/learningautonomy\/files\/2024\/11\/Screenshot-2024-11-15-at-3.56.07\u202fPM-710x710.png 710w, https:\/\/sites.bu.edu\/learningautonomy\/files\/2024\/11\/Screenshot-2024-11-15-at-3.56.07\u202fPM-300x300.png 300w, https:\/\/sites.bu.edu\/learningautonomy\/files\/2024\/11\/Screenshot-2024-11-15-at-3.56.07\u202fPM-600x600.png 600w\" sizes=\"(max-width: 212px) 100vw, 212px\" \/><img loading=\"lazy\" src=\"\/learningautonomy\/files\/2024\/10\/EhsanSabouni_2-150x150.jpg\" alt=\"\" width=\"212\" height=\"212\" class=\" wp-image-422 alignleft\" srcset=\"https:\/\/sites.bu.edu\/learningautonomy\/files\/2024\/10\/EhsanSabouni_2-150x150.jpg 150w, https:\/\/sites.bu.edu\/learningautonomy\/files\/2024\/10\/EhsanSabouni_2.jpg 200w\" sizes=\"(max-width: 212px) 100vw, 212px\" \/>H M Sabbir Ahmad and Ehsan Sabouni<\/span><\/strong><br \/>\n<span style=\"color: #003366;\">Advisor: Wenchao Li and Christos Cassandras<\/span><br \/>\n<span style=\"color: #003366;\">Boston University<\/span><\/p>\n<p><strong>Talk Title: <\/strong>Bridging Learning and Safety: A Reinforcement Learning Based Approach for Joint Receding Horizon and Control Barrier Function Framework<\/p>\n<p><span><strong>Abstract: <\/strong><\/span>Optimal control methods provide solutions to safety-critical problems but easily become intractable. Control Barrier Functions (CBFs) have emerged as a popular technique that facilitates their solution by provably guaranteeing safety, through their forward invariance property. This approach involves defining a performance objective alongside CBF-based safety constraints that must always be enforced. However, achieving a balance between performance and solution feasibility can be heavily influenced by two main factors: the choice of cost function and its parameters, as well as the calibration of parameters within the CBF-based constraints. To address these challenges, we propose a Reinforcement Learning (RL)-based Receding Horizon Control (RHC) approach leveraging Model Predictive Control (MPC) with CBFs (MPC-CBF). In particular, we parameterize our controller and use bilevel optimization, where RL is used to learn the optimal parameters while MPC computes the optimal control input. We validate our method by applying it to the challenging automated merging control problem for Connected and Automated Vehicles (CAVs) at conflicting roadways. This work contributes to enhancing the safety of learning-based autonomous systems in real-world applications, offering a promising direction for addressing the performance-safety trade-off in learning-based control systems.<\/p>\n<p>&nbsp;<\/p>\n<p><span style=\"color: #003366;\"><img loading=\"lazy\" src=\"\/learningautonomy\/files\/2024\/10\/Unknown-2-150x150.jpeg\" alt=\"\" width=\"214\" height=\"214\" class=\" wp-image-415 alignleft\" span=\"\" style=\"color: #003366;\" \/><strong>Abin Binoy George<\/strong><\/span><br \/>\n<span style=\"color: #003366;\">Advisor: Sabrina Neuman<\/span><br \/>\n<span style=\"color: #003366;\">Boston University<\/span><\/p>\n<p><strong>Talk Title: <span>Co-Design of Mechanical &amp; Computer Hardware for Robotics<\/span><\/strong><\/p>\n<p><strong>Abstract: <\/strong>Efficiently executing specific tasks is the primary goal of robotics, with key paradigms including mechanical design, control systems, and computational hardware. These three facets are interdependent; altering one will affect the others. For example, using research frameworks like RoboGrammar, we demonstrate that variations in hardware mass significantly affect optimal robot designs. Our focus is on co-designing the robot&#8217;s mechanical structure and control systems to enhance efficiency, using an origami quadruped as our experimental platform. Initial explorations involve tuning design and gait parameters to improve power consumption and performance, employing Bayesian Optimization for optimal configurations. Future work will integrate advanced controllers and a computer hardware simulator to create a comprehensive framework for optimizing robotic systems across various tasks and terrains.<\/p>\n<p>&nbsp;<\/p>\n<p><span style=\"color: #003366;\"><img loading=\"lazy\" src=\"\/learningautonomy\/files\/2024\/10\/Zijian-Guo-150x150.jpg\" alt=\"\" width=\"214\" height=\"214\" class=\" wp-image-390 alignleft\" srcset=\"https:\/\/sites.bu.edu\/learningautonomy\/files\/2024\/10\/Zijian-Guo-150x150.jpg 150w, https:\/\/sites.bu.edu\/learningautonomy\/files\/2024\/10\/Zijian-Guo-550x550.jpg 550w, https:\/\/sites.bu.edu\/learningautonomy\/files\/2024\/10\/Zijian-Guo-710x710.jpg 710w, https:\/\/sites.bu.edu\/learningautonomy\/files\/2024\/10\/Zijian-Guo-300x300.jpg 300w, https:\/\/sites.bu.edu\/learningautonomy\/files\/2024\/10\/Zijian-Guo-600x600.jpg 600w\" sizes=\"(max-width: 214px) 100vw, 214px\" \/><strong>Zijian Guo<\/strong><\/span><br \/>\n<span style=\"color: #003366;\">Advisor: Wenchao Li<\/span><br \/>\n<span style=\"color: #003366;\">Boston University<\/span><\/p>\n<p><strong>Talk Title: <span>Temporal Logic Specification-Conditioned Decision Transformer for Offline Safe Reinforcement Learning<\/span><\/strong><\/p>\n<p><strong>Abstract: <\/strong>Offline safe reinforcement learning (RL) aims to train a constraint satisfaction policy from a fixed dataset. Current state-of-the-art approaches are based on supervised learning with a conditioned policy. However, these approaches fall short in real-world applications that involve complex tasks with rich temporal and logical structures. In this paper, we propose temporal logic Specification- conditioned Decision Transformer (SDT), a novel framework that harnesses the expressive power of signal temporal logic (STL) to specify complex temporal rules that an agent should follow and the sequential modeling capability of Decision Transformer (DT). Empirical evaluations on the DSRL benchmarks demonstrate the better capacity of SDT in learning safe and high-reward policies compared with existing approaches. In addition, SDT shows good alignment with respect to different desired degrees of satisfaction of the STL specification that it is conditioned on.<\/p>\n<p>&nbsp;<\/p>\n<p><span style=\"color: #003366;\"><img loading=\"lazy\" src=\"\/learningautonomy\/files\/2024\/10\/RanJing_2022-150x150.png\" alt=\"\" width=\"214\" height=\"214\" class=\" wp-image-419 alignleft\" srcset=\"https:\/\/sites.bu.edu\/learningautonomy\/files\/2024\/10\/RanJing_2022-150x150.png 150w, https:\/\/sites.bu.edu\/learningautonomy\/files\/2024\/10\/RanJing_2022-550x550.png 550w, https:\/\/sites.bu.edu\/learningautonomy\/files\/2024\/10\/RanJing_2022-300x300.png 300w, https:\/\/sites.bu.edu\/learningautonomy\/files\/2024\/10\/RanJing_2022.png 600w\" sizes=\"(max-width: 214px) 100vw, 214px\" \/><strong>Ran Jing<\/strong><\/span><br \/>\n<span style=\"color: #003366;\">Advisor: Andrew Sabelhaus<\/span><br \/>\n<span style=\"color: #003366;\">Boston University<\/span><\/p>\n<p><strong>Talk Title: <span>Self-Sensing for Proprioception and Contact Detection in Soft Robots Using Shape Memory Alloy Artificial Muscles<\/span><\/strong><\/p>\n<p><strong>Abstract: <\/strong>Estimating a soft robot&#8217;s pose and applied forces, also called proprioception, is crucial for safe interaction of the robot with its environment. However, most solutions for soft robot proprioception use dedicated sensors, particularly for external forces, which introduce design trade-offs, rigidity, and risk of failure. This work presents an approach for pose estimation and contact detection for soft robots actuated by shape memory alloy (SMA) artificial muscles, using no dedicated force sensors. Our framework uses the unique material properties of SMAs to self-sense their internal stress, via offboard measurements of their electrical resistance and in-situ temperature readings, in an existing fully-soft limb design. We demonstrate that a simple polynomial regression model on these measurements is sufficient to predict the robot&#8217;s pose, under no-contact conditions. Then, we show that if an additional measurement of the true pose is available (e.g. from an already-in-place bending sensor), it is possible to predict a binary contact\/no-contact using multiple combinations of self-sensing signals. Our hardware tests verify our hypothesis via a contact detection test with a human operator. This proof-of-concept validates that self-sensing signals in soft SMA-actuated soft robots can be used for proprioception and contact detection, and suggests a direction for integrating proprioception into soft robots without design compromises. Future work could employ machine learning for enhanced accuracy.<\/p>\n<p>&nbsp;<\/p>\n<p><span style=\"color: #003366;\"><img loading=\"lazy\" src=\"\/learningautonomy\/files\/2024\/10\/Anni-Li-150x150.jpg\" alt=\"\" width=\"214\" height=\"214\" class=\" wp-image-420 alignleft\" srcset=\"https:\/\/sites.bu.edu\/learningautonomy\/files\/2024\/10\/Anni-Li-150x150.jpg 150w, https:\/\/sites.bu.edu\/learningautonomy\/files\/2024\/10\/Anni-Li-550x550.jpg 550w, https:\/\/sites.bu.edu\/learningautonomy\/files\/2024\/10\/Anni-Li-710x710.jpg 710w, https:\/\/sites.bu.edu\/learningautonomy\/files\/2024\/10\/Anni-Li-300x300.jpg 300w, https:\/\/sites.bu.edu\/learningautonomy\/files\/2024\/10\/Anni-Li-600x600.jpg 600w\" sizes=\"(max-width: 214px) 100vw, 214px\" \/><strong>Anni Li<\/strong><\/span><br \/>\n<span style=\"color: #003366;\">Advisor: Christos Cassandras<\/span><br \/>\n<span style=\"color: #003366;\">Boston University<\/span><\/p>\n<p><strong>Talk Title: <span>Human-Autonomous Vehicle Safe Interactions in Multi-agent Systems<\/span><\/strong><\/p>\n<p><strong>Abstract: <\/strong>We study safe driving interactions between Human-Driven Vehicles (HDVs) and Connected and Autonomous Vehicles (CAVs) in mixed traffic to derive time and energy-optimal policies for CAVs to complete lane change maneuvers. The interaction between CAVs and HDVs can be formulated using a bilevel optimization setting with an appropriate behavioral model for HDV, requiring the best possible response from a CAV to actions by its neighboring HDVs. An iterated best response (IBR) method is then used to determine a Nash equilibrium. Besides, we also show that CAV cooperation can eliminate or greatly reduce the interaction between CAVs and HDVs. We derive a simple threshold-based criterion to select an optimal policy for the lane-changing CAV to merge ahead of a cooperating CAV in the target lane. In this case, the trajectory of the lane-changing CAV is independent of HDV behavior. Moreover, in the case where the dynamics and control policies of HDVs are unknown and hard to predict, we employ event-triggered Control Barrier Functions (CBFs) to estimate the HDV model online, construct data-driven and state-feedback safety controllers, and transform constrained optimal control problems for CAVs into a sequence of event-triggered quadratic programs. We show that we can ensure collision-free interactions between HDVs and CAVs and demonstrate the robustness and flexibility of our framework on different types of human drivers in lane-changing scenarios while guaranteeing the satisfaction of safety constraints.<\/p>\n<p>&nbsp;<\/p>\n<p><span style=\"color: #003366;\"><img loading=\"lazy\" src=\"\/learningautonomy\/files\/2024\/10\/EhsanSabouni_2-150x150.jpg\" alt=\"\" width=\"215\" height=\"215\" class=\" wp-image-422 alignleft\" srcset=\"https:\/\/sites.bu.edu\/learningautonomy\/files\/2024\/10\/EhsanSabouni_2-150x150.jpg 150w, https:\/\/sites.bu.edu\/learningautonomy\/files\/2024\/10\/EhsanSabouni_2.jpg 200w\" sizes=\"(max-width: 215px) 100vw, 215px\" \/><strong>Ehsan Sabouni<\/strong><\/span><br \/>\n<span style=\"color: #003366;\">Advisor: Christos Cassandras<\/span><br \/>\n<span style=\"color: #003366;\">Boston University<\/span><\/p>\n<p><strong>Talk Title: <span>Bridging Learning and Safety: A Reinforcement Learning Based Approach for Joint Receding Horizon and Control Barrier Function Framework<\/span><\/strong><\/p>\n<p><strong>Abstract: <\/strong>Optimal control methods provide solutions to safety-critical problems but easily become intractable. Control Barrier Functions (CBFs) have emerged as a popular technique that facilitates their solution by provably guaranteeing safety, through their forward invariance property. This approach involves defining a performance objective alongside CBF-based safety constraints that must always be enforced. However, achieving a balance between performance and solution feasibility can be heavily influenced by two main factors: the choice of cost function and its parameters, as well as the calibration of parameters within the CBF-based constraints. To address these challenges, we propose a Reinforcement Learning (RL)-based Receding Horizon Control (RHC) approach leveraging Model Predictive Control (MPC) with CBFs (MPC-CBF). In particular, we parameterize our controller and use bilevel optimization, where RL is used to learn the optimal parameters while MPC computes the optimal control input. We validate our method by applying it to the challenging automated merging control problem for Connected and Automated Vehicles (CAVs) at conflicting roadways. This work contributes to enhancing the safety of learning-based autonomous systems in real-world applications, offering a promising direction for addressing the performance-safety trade-off in learning-based control systems.<\/p>\n<p>&nbsp;<\/p>\n<p><span style=\"color: #003366;\"><img loading=\"lazy\" src=\"\/learningautonomy\/files\/2024\/10\/Zili-Wang-9-150x150.jpg\" alt=\"\" width=\"214\" height=\"214\" class=\" wp-image-423 alignleft\" srcset=\"https:\/\/sites.bu.edu\/learningautonomy\/files\/2024\/10\/Zili-Wang-9-150x150.jpg 150w, https:\/\/sites.bu.edu\/learningautonomy\/files\/2024\/10\/Zili-Wang-9-300x300.jpg 300w, https:\/\/sites.bu.edu\/learningautonomy\/files\/2024\/10\/Zili-Wang-9.jpg 512w\" sizes=\"(max-width: 214px) 100vw, 214px\" \/><strong>Zili Wang<\/strong><\/span><br \/>\n<span style=\"color: #003366;\">Advisor: Sean Andersson<\/span><br \/>\n<span style=\"color: #003366;\">Boston University<\/span><\/p>\n<p><strong>Talk Title: <span>Learning-enabled Navigation and Nonlinear Control for Resource Constrained Robots<\/span><\/strong><\/p>\n<p><strong>Abstract: <\/strong>Navigating complex and unknown environments is a remarkable ability shared by both humans and animals, allowing them to predict important environmental features and devise effective strategies for success based on past experiences. Equipping robots with a similar capability is a valuable challenge for real-world applications. While recent advances in deep learning have improved robotic task execution through experience-based learning, many existing methods demand extensive sensing and computational resources. This research addresses the challenge of resource-efficient semantic navigation and reliable nonlinear control in unknown structured environments. We propose a novel framework that divides decision-making into two key components: (1) high-level planning under resource constraints, where the selection of intermediate goals is informed by the environment\u2019s predictable structure. This stage includes a scene network module that extrapolates the environment from partial data and a planning module that charts the trajectory toward the target. (2) Low-level control, where a nonlinear controller navigates toward the goals set in the planning stage. This component focuses on a control network designed for stability and safety in polygonal environments. By integrating them, we provide a robust decision-making framework applicable to a wide range of robotic applications.<\/p>\n<p>&nbsp;<\/p>\n<p><span style=\"color: #003366;\"><img loading=\"lazy\" src=\"\/learningautonomy\/files\/2024\/10\/1597599549884-150x150.jpeg\" alt=\"\" width=\"214\" height=\"214\" class=\" wp-image-424 alignleft\" srcset=\"https:\/\/sites.bu.edu\/learningautonomy\/files\/2024\/10\/1597599549884-150x150.jpeg 150w, https:\/\/sites.bu.edu\/learningautonomy\/files\/2024\/10\/1597599549884-300x300.jpeg 300w, https:\/\/sites.bu.edu\/learningautonomy\/files\/2024\/10\/1597599549884.jpeg 400w\" sizes=\"(max-width: 214px) 100vw, 214px\" \/><strong>Alp Eren Yilmaz<\/strong><\/span><br \/>\n<span style=\"color: #003366;\">Advisor: Sabrina Neuman<\/span><br \/>\n<span style=\"color: #003366;\">Boston University<\/span><\/p>\n<p><strong>Talk Title:\u00a0Accelerating Rigid Body Dynamics through Variable Fixed-Precision with Provable Error-Bounds<\/strong><\/p>\n<p><strong>Abstract:\u00a0<\/strong>Robotic applications increasingly demand mobility, imposing strict constraints on battery capacity, weight, and power consumption. The integration of advanced AI and optimal control at the edge has heightened computational demands, worsening these constraints, especially for tiny robots. This drives the need for compressing these algorithms to enable deployment on embedded edge hardware. A promising approach is fixed-point arithmetic, which requires less memory and is faster than standard floating-point arithmetic, and is already demonstrated in optimal control applications. Previous works used a single fixed-point representation through trial and error, missing key optimization opportunities. Tools like Daisy now enable rigorous stability proofs for computational graphs after fixed-point conversion, but their generalized approach often leads to suboptimal results or extended runtimes. Our approach leverages domain-specific knowledge to address these issues. Additionally, we are developing an end-to-end pipeline to provide provable error bounds on variable-representation fixed precision for robotics applications, focusing initially on rigid body dynamics\u2014a known bottleneck in optimal control. Early experiments show that a 20x speedup is achievable on a microcontroller without an FPU, such as Raspberry Pi Pico.<\/p>\n<p>&nbsp;<\/p>\n<p><img loading=\"lazy\" src=\"\/learningautonomy\/files\/2024\/10\/headshotsquare-1-150x150.jpg\" alt=\"\" width=\"249\" height=\"249\" class=\"wp-image-462 alignleft\" srcset=\"https:\/\/sites.bu.edu\/learningautonomy\/files\/2024\/10\/headshotsquare-1-150x150.jpg 150w, https:\/\/sites.bu.edu\/learningautonomy\/files\/2024\/10\/headshotsquare-1-636x636.jpg 636w, https:\/\/sites.bu.edu\/learningautonomy\/files\/2024\/10\/headshotsquare-1-768x768.jpg 768w, https:\/\/sites.bu.edu\/learningautonomy\/files\/2024\/10\/headshotsquare-1-550x550.jpg 550w, https:\/\/sites.bu.edu\/learningautonomy\/files\/2024\/10\/headshotsquare-1-710x710.jpg 710w, https:\/\/sites.bu.edu\/learningautonomy\/files\/2024\/10\/headshotsquare-1-300x300.jpg 300w, https:\/\/sites.bu.edu\/learningautonomy\/files\/2024\/10\/headshotsquare-1-600x600.jpg 600w, https:\/\/sites.bu.edu\/learningautonomy\/files\/2024\/10\/headshotsquare-1.jpg 851w\" sizes=\"(max-width: 249px) 100vw, 249px\" \/> <img loading=\"lazy\" src=\"\/learningautonomy\/files\/2024\/10\/IMG_1232-e1566874260402-636x636-2-150x150.jpg\" alt=\"\" width=\"249\" height=\"249\" class=\"wp-image-463 alignleft\" srcset=\"https:\/\/sites.bu.edu\/learningautonomy\/files\/2024\/10\/IMG_1232-e1566874260402-636x636-2-150x150.jpg 150w, https:\/\/sites.bu.edu\/learningautonomy\/files\/2024\/10\/IMG_1232-e1566874260402-636x636-2.jpg 636w, https:\/\/sites.bu.edu\/learningautonomy\/files\/2024\/10\/IMG_1232-e1566874260402-636x636-2-550x550.jpg 550w, https:\/\/sites.bu.edu\/learningautonomy\/files\/2024\/10\/IMG_1232-e1566874260402-636x636-2-300x300.jpg 300w, https:\/\/sites.bu.edu\/learningautonomy\/files\/2024\/10\/IMG_1232-e1566874260402-636x636-2-600x600.jpg 600w\" sizes=\"(max-width: 249px) 100vw, 249px\" \/><span style=\"color: #003366;\"><strong>Mela Coffey &amp; Dawei Zhang<\/strong><\/span><br \/>\n<span style=\"color: #003366;\">Advisor: Alyssa Pierson and Roberto Tron<\/span><br \/>\n<span style=\"color: #003366;\">Boston University<\/span><\/p>\n<p><strong>Talk Title: <span>Reactive and Safe Co-Navigation with Haptic Guidance<\/span><\/strong><\/p>\n<p data-olk-copy-source=\"MessageBody\"><strong>Abstract: <\/strong>We propose a co-navigation algorithm that facilitates collaborative navigation between a human and a robot toward a shared goal. In this framework, the human pilot makes high-level directional decisions, while the robot provides haptic feedback for collision avoidance and path guidance, adapting to dynamic environmental changes. Our algorithm uses optimized Rapidly-exploring Random Trees (RRT*) to generate paths to lead the user to the goal, via an attractive force feedback computed using a Control Lyapunov Function (CLF). Simultaneously, we ensure collision avoidance through a Control Barrier Function (CBF) where necessary. Our approach is validated through simulations with a virtual pilot and hardware experiments with a human pilot. The results show that integrating RRT*\u00a0with CBFs is a promising tool for effective human-robot co-navigation.<\/p>\n<p>&nbsp;<\/p>\n<p><img loading=\"lazy\" src=\"\/learningautonomy\/files\/2024\/11\/1661210515997-150x150.jpeg\" alt=\"\" width=\"249\" height=\"249\" class=\"wp-image-492 alignleft\" srcset=\"https:\/\/sites.bu.edu\/learningautonomy\/files\/2024\/11\/1661210515997-150x150.jpeg 150w, https:\/\/sites.bu.edu\/learningautonomy\/files\/2024\/11\/1661210515997-300x300.jpeg 300w, https:\/\/sites.bu.edu\/learningautonomy\/files\/2024\/11\/1661210515997.jpeg 400w\" sizes=\"(max-width: 249px) 100vw, 249px\" \/> <img loading=\"lazy\" src=\"\/learningautonomy\/files\/2024\/11\/1698183959296-150x150.jpeg\" alt=\"\" width=\"249\" height=\"249\" class=\"wp-image-493 alignleft\" srcset=\"https:\/\/sites.bu.edu\/learningautonomy\/files\/2024\/11\/1698183959296-150x150.jpeg 150w, https:\/\/sites.bu.edu\/learningautonomy\/files\/2024\/11\/1698183959296-300x300.jpeg 300w, https:\/\/sites.bu.edu\/learningautonomy\/files\/2024\/11\/1698183959296.jpeg 400w\" sizes=\"(max-width: 249px) 100vw, 249px\" \/><\/p>\n<p><span style=\"color: #003366;\"><strong>Zirui Zang &amp; Ahmad Anime<\/strong><\/span><br \/>\n<span style=\"color: #003366;\">Advisor: Rahul Mangharam<\/span><br \/>\n<span style=\"color: #003366;\">University of Pennsylvania<\/span><\/p>\n<p><strong>Talk Title: <span data-olk-copy-source=\"MessageBody\">IT-LMPC: Information Theoretic Learning Model Predictive Control for Safe Iterative Learning Control<\/span><\/strong><\/p>\n<p data-olk-copy-source=\"MessageBody\"><strong>Abstract: <\/strong>We present an approach to solve the Learning Model Predictive Control (LMPC) problem using Model Predictive Path Integral (MPPI) as a means of extending LMPC to stochastic non-linear systems. LMPC solves the optimal control policy that minimizes a value function by iteratively improving performance and learning from previous trajectories. Unlike reinforcement learning, the LMPC formulation provides desirable theoretical properties for safety and optimality, which is valuable for safety-critical applications. Previous methods utilizing optimization-based techniques or Cross-Entropy-Method sampling to solve the LMPC problem fall short when dealing with high-dimensional, nonlinear dynamics. Our MPPI-based framework optimizes the control policy from an information-theoretic perspective and overcomes these limitations by providing a systematic way of handling constraints without sacrificing sample spread. We validate our approach through simulations and real-world experiments, demonstrating significant improvements in constraint satisfaction and final trajectory performance.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>H M Sabbir Ahmad and Ehsan Sabouni Advisor: Wenchao Li and Christos Cassandras Boston University Talk Title: Bridging Learning and Safety: A Reinforcement Learning Based Approach for Joint Receding Horizon and Control Barrier Function Framework Abstract: Optimal control methods provide solutions to safety-critical problems but easily become intractable. Control Barrier Functions (CBFs) have emerged as [&hellip;]<\/p>\n","protected":false},"author":18553,"featured_media":0,"parent":0,"menu_order":0,"comment_status":"closed","ping_status":"closed","template":"page-templates\/no-sidebars.php","meta":[],"_links":{"self":[{"href":"https:\/\/sites.bu.edu\/learningautonomy\/wp-json\/wp\/v2\/pages\/406"}],"collection":[{"href":"https:\/\/sites.bu.edu\/learningautonomy\/wp-json\/wp\/v2\/pages"}],"about":[{"href":"https:\/\/sites.bu.edu\/learningautonomy\/wp-json\/wp\/v2\/types\/page"}],"author":[{"embeddable":true,"href":"https:\/\/sites.bu.edu\/learningautonomy\/wp-json\/wp\/v2\/users\/18553"}],"replies":[{"embeddable":true,"href":"https:\/\/sites.bu.edu\/learningautonomy\/wp-json\/wp\/v2\/comments?post=406"}],"version-history":[{"count":26,"href":"https:\/\/sites.bu.edu\/learningautonomy\/wp-json\/wp\/v2\/pages\/406\/revisions"}],"predecessor-version":[{"id":501,"href":"https:\/\/sites.bu.edu\/learningautonomy\/wp-json\/wp\/v2\/pages\/406\/revisions\/501"}],"wp:attachment":[{"href":"https:\/\/sites.bu.edu\/learningautonomy\/wp-json\/wp\/v2\/media?parent=406"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}