Skip to content

Commit b769c86

Browse files
committed
Hanlan's paper
1 parent fd90f40 commit b769c86

File tree

2 files changed

+27
-1
lines changed

2 files changed

+27
-1
lines changed

_bibliography/papers.bib

Lines changed: 14 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -6,10 +6,23 @@ @string{iros
66
@string{ral = {IEEE Robotics and Automation Letters}}
77
@string{tro = {IEEE Transactions on Robotics}}
88
@string{ijrr = {The International Journal of Robotics Research}}
9+
@misc{yang2025pachs,
10+
title = {Parallel Heuristic Search as Inference for Actor-Critic Reinforcement Learning Models},
11+
author = {Hanlan Yang and Itamar Mishani and Luca Pivetti and Zachary Kingston and Maxim Likhachev},
12+
abstract = {Actor-Critic models are a class of model-free deep reinforcement learning (RL) algorithms that have demonstrated effectiveness across various robot learning tasks. While considerable research has focused on improving training stability and data sampling efficiency, most deployment strategies have remained relatively simplistic, typically relying on direct actor policy rollouts. In contrast, we propose \pachs{} (\textit{P}arallel \textit{A}ctor-\textit{C}ritic \textit{H}euristic \textit{S}earch), an efficient parallel best-first search algorithm for inference that leverages both components of the actor-critic architecture: the actor network generates actions, while the critic network provides cost-to-go estimates to guide the search. Two levels of parallelism are employed within the search -- actions and cost-to-go estimates are generated in batches by the actor and critic networks respectively, and graph expansion is distributed across multiple threads. We demonstrate the effectiveness of our approach in robotic manipulation tasks, including collision-free motion planning and contact-rich interactions such as non-prehensile pushing.},
13+
eprint = {2509.25402},
14+
archiveprefix = {arXiv},
15+
primaryclass = {cs.RO},
16+
year = 2025,
17+
pdf = {https://arxiv.org/abs/2509.25402},
18+
projects = {implicit},
19+
note = {Under Review},
20+
abbr = {ARXIV}
21+
}
922
@misc{meng2025look,
1023
title = {Look as You Leap: Planning Simultaneous Motion and Perception for High-{DoF} Robots},
1124
author = {Qingxi Meng and Emiliano Flores and Carlos Quintero-Peña and Peizhu Qian and Zachary Kingston and Shannan K. Hamlin and Vaibhav Unhelkar and Lydia E. Kavraki},
12-
abstract = { In this work, we address the problem of planning robot motions for a high-degree-of-freedom (DoF) robot that effectively achieves a given perception task while the robot and the perception target move in a dynamic environment. Achieving navigation and perception tasks simultaneously is challenging, as these objectives often impose conflicting requirements. Existing methods that compute motion under perception constraints fail to account for obstacles, are designed for low-DoF robots, or rely on simplified models of perception. Furthermore, in dynamic real-world environments, robots must replan and react quickly to changes and directly evaluating the quality of perception (e.g., object detection confidence) is often expensive or infeasible at runtime. This problem is especially important in human-centered environments such as homes and hospitals, where effective perception is essential for safe and reliable operation. To address these challenges, we propose a GPU-parallelized perception-score-guided probabilistic roadmap planner with a neural surrogate model (PS-PRM). The planner explicitly incorporates the estimated quality of a perception task into motion planning for high-DoF robots. Our method uses a learned model to approximate perception scores and leverages GPU parallelism to enable efficient online replanning in dynamic settings. We demonstrate that our planner, evaluated on high-DoF robots, outperforms baseline methods in both static and dynamic environments in both simulation and real-robot experiments.},
25+
abstract = {In this work, we address the problem of planning robot motions for a high-degree-of-freedom (DoF) robot that effectively achieves a given perception task while the robot and the perception target move in a dynamic environment. Achieving navigation and perception tasks simultaneously is challenging, as these objectives often impose conflicting requirements. Existing methods that compute motion under perception constraints fail to account for obstacles, are designed for low-DoF robots, or rely on simplified models of perception. Furthermore, in dynamic real-world environments, robots must replan and react quickly to changes and directly evaluating the quality of perception (e.g., object detection confidence) is often expensive or infeasible at runtime. This problem is especially important in human-centered environments such as homes and hospitals, where effective perception is essential for safe and reliable operation. To address these challenges, we propose a GPU-parallelized perception-score-guided probabilistic roadmap planner with a neural surrogate model (PS-PRM). The planner explicitly incorporates the estimated quality of a perception task into motion planning for high-DoF robots. Our method uses a learned model to approximate perception scores and leverages GPU parallelism to enable efficient online replanning in dynamic settings. We demonstrate that our planner, evaluated on high-DoF robots, outperforms baseline methods in both static and dynamic environments in both simulation and real-robot experiments.},
1326
eprint = {2509.19610},
1427
archiveprefix = {arXiv},
1528
primaryclass = {cs.RO},

_data/coauthors.yml

Lines changed: 13 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -51,6 +51,19 @@
5151
member: true
5252
url: talha
5353

54+
"yang":
55+
- firstname: ["Hanlan"]
56+
member: true
57+
url: hanlan
58+
59+
"likhachev":
60+
- firstname: ["Maxim"]
61+
url: https://www.cs.cmu.edu/~maxim/
62+
63+
"mishani":
64+
- firstname: ["Itamar"]
65+
url: https://imishani.github.io/
66+
5467
"blumenschein":
5568
- firstname: ["Laura H."]
5669
url: "https://engineering.purdue.edu/ME/People/ptProfile?resource_id=241064"

0 commit comments

Comments
 (0)