Theses in Robotics: Difference between revisions

From Intelligent Materials and Systems Lab

Line 151: Line 151:
*Madis K Nigol, [http://hdl.handle.net/10062/64339 Õppematerjalid robotplatvormile Robotont (Study materials for robot olatform Robotont)], MS thesis, 2019
*Madis K Nigol, [http://hdl.handle.net/10062/64339 Õppematerjalid robotplatvormile Robotont (Study materials for robot olatform Robotont)], MS thesis, 2019
*Renno Raudmäe, [http://hdl.handle.net/10062/64341 Avatud robotplatvorm Robotont (Open source robotics platform Robotont)], MS thesis, 2019
*Renno Raudmäe, [http://hdl.handle.net/10062/64341 Avatud robotplatvorm Robotont (Open source robotics platform Robotont)], MS thesis, 2019
*Asif Sattar, tba
*Asif Sattar, [http://hdl.handle.net/10062/64352 Human detection and distance estimation with monocular camera using YOLOv3 neural network (Inimeste tuvastamine ning kauguse hindamine kasutades kaamerat ning YOLOv3 tehisnärvivõrku)], MS thesis, 2019
*Ragnar Margus, tba
*Ragnar Margus, tba
*Pavel Šumejko, tba
*Pavel Šumejko, tba

Revision as of 21:57, 15 June 2019

Projects in Advanced Robotics

The main objective of the follwing projects is to give students experience in working with advanced robotics tehcnology. Our group is active in several R&D projects involving human-robot collaboration, intuitive teleoperation of robots, and autonomous navigation of unmanned mobile platforms. Our main software platforms are Robot Operating System (ROS) for developing software for advanced robot systems and Gazebo for running realistic robotic simulations.

For further information, contact Karl Kruusamäe

The following is not an exhaustive list of all available thesis/research topics.


KUKA youBot

Development of demonstrative and promotional applications for KUKA youBot

The goal of this project is to develop promotional use cases for KUKA youBot, that demonstrate the capabilities of modern robotics and inspire people to get involved with it. The list of possible robotic demos include:

  • demonstration of motion planning algorithms for mobile manipulation,
  • using 3D vision for human and/or environment detection,
  • interactive navigation,
  • autonomous path planning,
  • different pick-and-place applications,
  • and human-robot collaboration.
Universal Robots UR5

Development of demonstrative and promotional applications for Universal Robots UR5

Sample demonstrations include:

  • autonomous pick-and-place,
  • load-assistance for human-robot collaboration,
  • packaging,
  • physical compliance during human-robot interaction,
  • tracing objects surface during scanning,
  • robotic kitting,
  • grinding of non-flat surfaces.

Development of demonstrative and promotional applications for Clearpath Jackal

Sample demonstrations include:

  • human-robot interaction,
  • multi-robot mapping,
  • autonomous driving,

ROS support and educational materials for open-source mobile robot

ROS

The goal of the project is to develop Robot Operating System (ROS) wrapper functions for drivers of an open-source mobile robot platform, that has been successfully used in Robotex competitions. Additionally, educational materials must be developed to integrate the robot platform into ROS ecosystem. The outcome of this work will give a lot of educational uses and rapid prototyping opportunities for the platform.

What is ROS?

Detecting features of urban and off-road surroundings

Accurate navigation of self-driving unmanned robotic platforms requires identification of traversable terrain. A combined analysis of point-cloud data with RGB information of the robot's environment can help autonomous systems make correct decisions. The goal of this work is to develop algorithms for terrain classification.
Mapping

Robotic simulations (in Gazebo)

1) Developing large area simulation worlds for mobile robotics. In order to develop robot navigation algorithms, it is more time- and cost-efficient to test robot behavior in a wide range of realistically simulated worlds. These simulated worlds include both indoor and outdoor environments. This work focuses on designing robot simulation environments using an open-source platform Gazebo.
2) Humans in Gazebo. Integrating walking and gesturing humans to Gazebo.
3) Tracked Robots in Gazebo. Creating and testing tracked robotis in Gaxebo and testing different track configurations.
4) Robot basketball simulation for game strategy and shot accuracy
5) Robotont at the Institute of Technology
Gazebo NASA Robonaut simulation in Gazebo

Follow-the-leader robotic demo

Detecting 2 hands with Leap Motion Controller

The idea is to create a robotic demonstration where a mobile robot is using Kinect or similar depth-camera for identifying a person and then starts following that person. The project will be implemented using Robot Operating System (ROS) on either KUKA youbot or similar mobile robot platform.

Detecting hand signals for intuitive human-robot interface

This project involves creating ROS libraries for using either a Leap Motion Controller or an RGB-D camera to detect most common human hand signals (e.g., thumbs up, thumbs down, all clear, pointing into distance, inviting).

Virtual reality user interface (VRUI) for intuitive teleoperation system

Adding virtual reality capability to a gesture- and natural-language-based robot teleoperation system.
Gesture-based teleoperation
Kruusamäe et al. (2016) High-precision telerobot with human-centered variable perspective and scalable gestural interface

Health monitor for intuitive telerobot

Intelligent status and error handling for an intuitive telerobotic system.

Dynamic stitching for achieveing 360° FOV

Automated image stitching of images from multiple camera sources for achieveing 360° field-of-view during mobile telerobotic inspection of remote areas.

3D scanning of industrial objects

Using laser sensors and cameras to create accurate models of inustrial producst for quality control or further processing.

Modeling humans for human-robot interaction

True human-robot collaboration means that the robot must understand the actions, intention, and state of its human partner. This work invovlves using cameras and other human sensors for digitally representing and modelling humans. There are multiple stages for modeling: a) physical models of human kinematics and dynamics; b) higher level-models for recognizing human intent.
ROS & Kinect & Skeleton-Markers Package

ROS wrapper for Estonian Speech Synthesizer

Creating a ROS package that enables robots to speak in Estonian. The basis of the work is the existing Estonian language speech synthesizer that needs to be integrated with ROS sound_play package or a stand-alone ROS wrapper package.

Robotic avatar for telepresence

Integrating hand gestures and head movements to control a robot avatar in virtual reality user interface.

ROS driver for Artificial Muscle actuators

Desigining a controller box and writing software for interfacing artificial muscle actuators [1, 2] ROS.

Upgrading KUKA youBot

KUKA youBot is omnidirectionl mobile manipulator platform. The project involves replacing the current onboard computer with a more capable one, installing newer version of Ubuntu Linux and ROS, testing core capabilities, and creating documentation for the robot.

Development of strategies for inter-robot knowledge representation

An inseparable part for making robots work together is to enable them to share knowledge about surrounding environment and robot's intention. This project focuses on developing and testing the methodologies for inter-robot knowledge representation. This project is developed as a subsystem of TeMoto.

TeMoto based smart home control

The project involves designing a open-source ROS+TeMoto based scalable smart home controller.

Detection of hardware and software resources for smart integration of robots

Vast majority of today’s robotic applications rely on hard-coded device and algorithm usage. This project focuses on developing a Resource Snooper software, that can detect addition or removal of resources for the benefit of dynamic reconfiguration of robotic systems. This project is developed as a subsystem of TeMoto.

Sonification of feedback during teleoperation of robots

Humans are used to receiving auditory feedback in their everyday lives. It helps us make decision and be aware of potential dangers. Telerobotic interfaces can deploy the same idea to improve the Situational Awareness and robotic task efficiency. The thesis project involves a study about different sonification solutions and implementation of it in a telerobotic application using ROS.

Human-Robot and Robot-Robot collaboration applications

Creating a demo or analysis of software capabilities related to human-robot or tobot-robot teams

  • human-robot collaborative assembly
  • distributed mapping; analysis and demo of existing ROS (e.g., segmap https://youtu.be/JJhEkIA1xSE) packages for multi-robot mapping
  • Inaccessible region teamwork
    • youbot+drone - a drone maps the environment (for example a maze) and ground vehicle uses this information to traverse the maze
    • youbot+clearbot - youbot cannot go up ledges but it can lift smaller robot, such as clearbot, up a ledge.

Developing ROS driver for a robotic gripper

The goal for this project is to develop ROS drivers for LEHF32K2-64 gripper. The work is concluded by demonstrating the functionalities of the gripper via pick-and-place task.
SMC LEHF32K2-64 gripper.

Mirroring human hand movements on industrial robots

The goal of this project is to integrate continuous control of industrial robot manipulator with a gestural telerobotics interface. The recommended tools for this thesis project are Leap Motion Controller or a standard web camera, Universal Robot UR5 manipulator, and ROS.

ROS2-based robotics demo

Converting ROS demos and tutorials to ROS2.

ROS2 for robotont

Creating ROS2 support for robotont mobile platform

TeMoto for robotont

Swarm-management for robotont using TeMoto framework.

3D lidar for mobile robotics

Analysing the technical characteristics of 3D lidar.. Desinging and constructing the mount for Ouster OS-1 lidar and validating its applicability for indoor and outdoor scenarios.

Making KUKA youBot user friendly again

This thesis focuses on integrating the low-level software capabilities of KUKA youBot in order to achieve high-level commonly used functionalities such as

  • teach mode - robot can replicate user demonstrated trajectories
  • end-effector jogging
  • gripper control
  • gamepad integration - user can control the robot via gamepad
  • web integration - user can control the robot via internet browser

The thesis is suitable for both, master and bachelor levels, as the associated code can be scaled up to generic "user-friendly control" package.

Flexible peer-to-peer network infrastructure for environments with restricted signal coverage

A very common issue with robotics in real world environments is that the network coverage is highly dependent on the environment. This makes the communication between the robot-to-base-station or robot-to-robot unreliable, potentially compromising the whole mission. This thesis focuses on implementing a peer-to-peer based network system on mobile robot platforms, where the platforms extend the network coverage between, e.g., an operator and a Worker robot. The work will be demonstrated in a real world setting, where common networking strategies for teleoperation (tethered or single router based) do not work.

Completed projects

Masters's theses

Bachelor's theses