My Work Time line: Difference between revisions

From Intelligent Materials and Systems Lab

Line 50: Line 50:
# '''laser''' was selected in fixed frame option menu in Display<br />
# '''laser''' was selected in fixed frame option menu in Display<br />
# '''laser scan''' was added from '''add''' option<br />
# '''laser scan''' was added from '''add''' option<br />
# Recording was started by using this command rosbag record LaserScanner /scan
# Recording was started by using this command   rosbag record LaserScanner /scan
==
=== Playing bag file ===

Revision as of 15:29, 31 July 2017

Welcome to my Activity Log!!! My Research topic is " Co-operative mappingg by using a swarm of sensors". The first target to achieve is to build up knowledge about using Ubuntu and ROS kinetics.

Installation

Ubuntu seems new to me as i always worked on window versions of operating systems. So for me task 1 was to install Ubuntu and make myself familiar to its working. When i started Ubuntu installation, i had to face problem because my system was not ready to accept different operating system. It started screaming out loud to warn me but i keep going. After 3 failures i became successful in Ubuntu installation over the cost of data in my hard drive.

[Ubuntu Official Website]

Installing ROS

Next task was to install ROS kinetics. Installing ROS Kinetics was not difficult after watching few tutorials on you tube.

Learning ROS

Use ROS Tutorials

How to add a new Package in workspace (27-06-2017)

Insert non-formatted text here

$ cd catkin_ws // go in workspace
$ cd src //go in source directory
$ git clone https://github.com/ros-drivers/usb_cam.git  // download package
/catkin_ws$ catkin_make //go in workspace and run catkin_make
$ source devel/setup.bash   // run it
$roslaunch usb_cam usb_cam-test.launch // to launch package

(29-06-2017)

Interfacing 3D Camera

http://wiki.ros.org/realsense_camera
https://01.org/developerjourney/recipe/intel-realsense-robotic-development-kit

Working with Lidar

Read this tutorial

http://wiki.ros.org/hokuyo_node/Tutorials/UsingTheHokuyoNode
==For Lidar==

first run roscore
then hokuyo_node
after it tf function
for tf function
rosrun tf static_transform_publisher 0.0 0.0 0.0 0.0 0.0 0.0 map my_frame 100

http://answers.ros.org/question/195962/rviz-fixed-frame-world-does-not-exist/

Then run rviz
rosrun rviz rviz

For navigation stack

https://github.com/ut-ims-robotics/tutorials/wiki/Navigation-stack

Recording and playing bag file

For this purpose Lidar scan was used and following steps were performed.

  1. roscore
  2. hokuyo_node was run rosrun but lidar was connected rosrun hokuyo_node hokuyo_node
  3. rviz run by using this command rosrun rviz rviz
  4. laser was selected in fixed frame option menu in Display
  5. laser scan was added from add option
  6. Recording was started by using this command rosbag record LaserScanner /scan

Playing bag file