Documenti full-text disponibili:
![[thumbnail of Thesis]](https://amslaurea.unibo.it/style/images/fileicons/application_pdf.png) |
Documento PDF (Thesis)
Full-text accessibile solo agli utenti istituzionali dell'Ateneo
Disponibile con Licenza: Salvo eventuali più ampie autorizzazioni dell'autore, la tesi può essere liberamente consultata e può essere effettuato il salvataggio e la stampa di una copia per fini strettamente personali di studio, di ricerca e di insegnamento, con espresso divieto di qualunque utilizzo direttamente o indirettamente commerciale. Ogni altro diritto sul materiale è riservato
Download (6MB)
| Contatta l'autore
|
Abstract
This thesis explores the use of foundation models in the context of robotic grasp planning, addressing the challenge of achieving both geometric robustness and task-oriented adaptability in
automated manipulation.
The research focuses on the integration of two state-of-the-art frameworks: GraspGen, a diffusion-based 6-DoF grasp generation model capable of producing diverse and physically consistent grasp configurations, and FoundationGrasp which leverages large-scale vision-language representations for task-aware grasp evaluation.
The proposed work aims to combine the semantic understanding of FoundationGrasp with the generative capabilities of GraspGen, developing a hybrid planning architecture implemented within the ROS 2 ecosystem.
The system is designed to generate grasp candidates through probabilistic sampling and subsequently refine them using task compatibility scores derived from multimodal foundation models. YOLO was employed for perception (detection and segmentation), behavior trees for control logic, and Docker for containerization ensuring reproducible deployments across development and testing phases.
The experimental evaluation demonstrates the benefits of this integration, showing improved task consistency and generalization across unseen objects and manipulation scenarios.
This work contributes to the development of intelligent grasp planning systems that bridge the gap between low-level geometric reasoning and high-level semantic understanding, paving the way toward more generalizable and adaptive robotic manipulation.
Abstract
This thesis explores the use of foundation models in the context of robotic grasp planning, addressing the challenge of achieving both geometric robustness and task-oriented adaptability in
automated manipulation.
The research focuses on the integration of two state-of-the-art frameworks: GraspGen, a diffusion-based 6-DoF grasp generation model capable of producing diverse and physically consistent grasp configurations, and FoundationGrasp which leverages large-scale vision-language representations for task-aware grasp evaluation.
The proposed work aims to combine the semantic understanding of FoundationGrasp with the generative capabilities of GraspGen, developing a hybrid planning architecture implemented within the ROS 2 ecosystem.
The system is designed to generate grasp candidates through probabilistic sampling and subsequently refine them using task compatibility scores derived from multimodal foundation models. YOLO was employed for perception (detection and segmentation), behavior trees for control logic, and Docker for containerization ensuring reproducible deployments across development and testing phases.
The experimental evaluation demonstrates the benefits of this integration, showing improved task consistency and generalization across unseen objects and manipulation scenarios.
This work contributes to the development of intelligent grasp planning systems that bridge the gap between low-level geometric reasoning and high-level semantic understanding, paving the way toward more generalizable and adaptive robotic manipulation.
Tipologia del documento
Tesi di laurea
(Laurea magistrale)
Autore della tesi
Tomasinelli, Lorenzo
Relatore della tesi
Correlatore della tesi
Scuola
Corso di studio
Indirizzo
AUTOMATION ENGINEERING
Ordinamento Cds
DM270
Parole chiave
UR5e, robotiq 2F-140, RealSense D435, YOLO, Docker, Behavior Trees, ROS2, MoveIt2, Rviz2, Grasping, Foundational Model, Foundation Model, FoundationGrasp, GraspGen, Task-Agnostic Grasping Generator, Task-Oriented Grasping Evaluator
Data di discussione della Tesi
25 Marzo 2026
URI
Altri metadati
Tipologia del documento
Tesi di laurea
(NON SPECIFICATO)
Autore della tesi
Tomasinelli, Lorenzo
Relatore della tesi
Correlatore della tesi
Scuola
Corso di studio
Indirizzo
AUTOMATION ENGINEERING
Ordinamento Cds
DM270
Parole chiave
UR5e, robotiq 2F-140, RealSense D435, YOLO, Docker, Behavior Trees, ROS2, MoveIt2, Rviz2, Grasping, Foundational Model, Foundation Model, FoundationGrasp, GraspGen, Task-Agnostic Grasping Generator, Task-Oriented Grasping Evaluator
Data di discussione della Tesi
25 Marzo 2026
URI
Statistica sui download
Gestione del documento: