Il full-text non è disponibile per scelta dell'autore.
      
        (
Contatta l'autore)
      
    
  
    
  
  
    
      Abstract
      This thesis is focused on deep reinforcement learning for mobile robot navigation in unstructured environments. Autonomous navigation has been mainly tackled through map-based algorithms, which are ineffective in many situations like exploration or rescue missions. For mapless scenarios, the simultaneous localization and planning (SLAM) represents a cornerstone on which a wide variety of algorithms are built. However the difficulty of maintaining a map from sensory inputs, typical of these methods, is leading the research community to look for alternatives. Deep reinforcement learning aims at solving the autonomous navigation problem end-to-end, by directly mapping high-dimensional inputs to actions, without the need for a model of the environment.
In this thesis, a model-free reinforcement learning approach is adopted: a variant of the tabular Q-learning algorithm, called deep Q-learning, uses a deep neural network to approximate the action-value function and to map states into velocity commands, without the need of an expert or supervisor. The learning model is trained in simulation on TurtleBot3 and Curiosity mobile robots in two different environments. After that, the neural network trained on TurtleBot3 is transferred on Curiosity and then fine-tuned on new navigation environments. The results are then compared to those obtained by training the model from scratch, with random initialization of the parameters: this comparison shows how, thanks to the pre-training, the rover manages to reach on average a higher number of targets per episode throughout the entire simulation.
     
    
      Abstract
      This thesis is focused on deep reinforcement learning for mobile robot navigation in unstructured environments. Autonomous navigation has been mainly tackled through map-based algorithms, which are ineffective in many situations like exploration or rescue missions. For mapless scenarios, the simultaneous localization and planning (SLAM) represents a cornerstone on which a wide variety of algorithms are built. However the difficulty of maintaining a map from sensory inputs, typical of these methods, is leading the research community to look for alternatives. Deep reinforcement learning aims at solving the autonomous navigation problem end-to-end, by directly mapping high-dimensional inputs to actions, without the need for a model of the environment.
In this thesis, a model-free reinforcement learning approach is adopted: a variant of the tabular Q-learning algorithm, called deep Q-learning, uses a deep neural network to approximate the action-value function and to map states into velocity commands, without the need of an expert or supervisor. The learning model is trained in simulation on TurtleBot3 and Curiosity mobile robots in two different environments. After that, the neural network trained on TurtleBot3 is transferred on Curiosity and then fine-tuned on new navigation environments. The results are then compared to those obtained by training the model from scratch, with random initialization of the parameters: this comparison shows how, thanks to the pre-training, the rover manages to reach on average a higher number of targets per episode throughout the entire simulation.
     
  
  
    
    
      Tipologia del documento
      Tesi di laurea
(Laurea magistrale)
      
      
      
      
        
      
        
          Autore della tesi
          Sarti, Lorenzo
          
        
      
        
          Relatore della tesi
          
          
        
      
        
          Correlatore della tesi
          
          
        
      
        
          Scuola
          
          
        
      
        
          Corso di studio
          
          
        
      
        
      
        
      
        
          Ordinamento Cds
          DM270
          
        
      
        
          Parole chiave
          deep reinforcement learning,autonomous navigation,mobile robot,unstructured environment,pre-training,robot navigation
          
        
      
        
          Data di discussione della Tesi
          20 Luglio 2021
          
        
      
      URI
      
      
     
   
  
    Altri metadati
    
      Tipologia del documento
      Tesi di laurea
(NON SPECIFICATO)
      
      
      
      
        
      
        
          Autore della tesi
          Sarti, Lorenzo
          
        
      
        
          Relatore della tesi
          
          
        
      
        
          Correlatore della tesi
          
          
        
      
        
          Scuola
          
          
        
      
        
          Corso di studio
          
          
        
      
        
      
        
      
        
          Ordinamento Cds
          DM270
          
        
      
        
          Parole chiave
          deep reinforcement learning,autonomous navigation,mobile robot,unstructured environment,pre-training,robot navigation
          
        
      
        
          Data di discussione della Tesi
          20 Luglio 2021
          
        
      
      URI
      
      
     
   
  
  
  
  
  
  
    
      Gestione del documento: 
      
        