All the news that's fit to assimilate[ Home | Blogs | Events | Robots | Humans | Projects | Podcasts | About | Account ]
For those of you getting started with ROS, I have written up a do-it-yourself tutorial for tracking a colored object using a web cam and AX-12 pan and tilt servos. You can find it here:
In this short article on using SLAM with ROS, I have posted a couple of videos showing Pi Robot mapping out part of an apartment using a Hokuyo laser scanner and the gmapping package. See
For the past several months, I have been learning the basics of ROS from Willow Garage. At the same time, I have been testing Mike Ferguson's "Poor Man's Lidar" or PML as an alternative to a more expensive laser range finder. The results are encouraging--at least for obstacle avoidance and simple navigation tasks. You can see the report at:
I recently converted most of the C# code for my Pi Robot project to Python. At the same time, I am changing the programming architecture to use message passing among nodes. To get started, I wrote up a little introduction to the topic at:
I finally had a chance to write up the math behind the Pi Robot arm tracking video. Keep in mind that I am only using the two shoulder joints in each arm--the elbow and wrist servos are fixed--so the inverse kinematics is fairly straightforward. Later on I'll have to deal with the other joints...
Here is the link to the write-up:
Here is a followup video to my previous blog entry. In this video, a number of independent behavioral threads are running to enable the robot to track and grasp the green balloon. Whenever the balloon is grasped, the robot turns its attention to the red balloon. When the green balloon is released, tracking turns again to it and the red balloon is ignored. I use RoboRealm to do the green/red tracking. There is a sonar sensor on the inside of the left hand that tells the robot when something is ready to be grasped. It can also do this using vision alone along with some trigonometry, but the result is more reliable when using the sensor.
I just finished up some work on using RoboRealm to guide my robot as it reaches toward a target object. The ultimate goal is for the robot to be able to pick up the object from a random location or take it from someone's hands. For now, I simply wanted to work out the coordinate transformations from visual space to arm space to get the two hands to point in the right direction as the target is moved about. The following video shows the results so far:
I don't have a full write-up yet on how I did this but it basically just uses 3-d coordinate transformations from the head angles and distance to the target (as measured by sonar and IR sensors mounted near the camera lens) to a frame of reference attached to each shoulder joint. The Dynamixel AX-12 servos are nice for this application since they can be queried for their current position info. The distance to the balloon as measured by the sonar and IR sensors is a little hit and miss and I think I'd get better performance using stereo vision instead.
I put together a new robot using Dynamixel AX-12+ servos and I wanted to test an algorithm for tracking a moving object. The camera being used is a DLink 920 wireless operating over 802.11g and the visual tracking is done using RoboRealm. All processing is done on my desktop PC. The full writeup can be found here:
2012 Top 10 Robot Christmas Gift Ideas
DARPA Robotics Challenge Kick Off
2012 ASABE Robot Contest Photos
Interview with David L. Heiserman
David Anderson on Subsumption Robots
Review: Apocalyptic AI by Robert M. Geraci
Raspberry Pi Interview with Eben Upton
2012 VEX Robotics World Championship
Giant Dallas Robot Cited as Best Public Art
There's More Than One Way to Skin a Robot
Day of the Androids at Hanson Robotics