Add Vision to Robots — See the Difference

Robots Paired With Machine Vision Create Automated and Efficient Motion, Leaving the Old Generation Behind

By Dan Hebert, PE, Senior Technical Editor

1 of 2 < 1 | 2 View on one page

CD1402 webcoverRobots have been used for decades in industrial manufacturing applications to perform tasks such as welding, part picking and metalworking. But without vision systems to guide robot movements and actions, these applications were limited to performing the same or very similar operations over and over on precisely located parts.

Consider an application where five similar, but not identical, sets of parts need to be welded. Without a vision system, five robot stations are required. Each part must be mounted in a costly high-precision fixture manually and then welded by a tool attached to a robot arm that moves with high accuracy and absolute repeatability.

SEE ALSO: Machine Vision and Industrial Robots

When a vision system is paired with a robot, one welding station can weld all five part types, as well as other parts. Each part doesn't need to be manually and precisely mounted, but instead just needs to be securely attached to a servo-driven housing, with the vision system aligning the parts as required.

Vision systems free robots from the narrow restrictions of time and space, providing a host of advantages in a wide variety of manufacturing applications. These advantages enable more widespread use of robots by guiding their actions to closely fit ever-changing and more-demanding production requirements.

See It, Do It
When a robot can't see, the part on which it will perform an action must be precisely placed, requiring a very accurate fixture or positioning system and a high-precision robot. But with vision, close is good enough, as the robot can find its own way. "Vision systems make it possible to use a less precise and expensive robot for precise placement applications such as wafer alignment, part orientation or part presence," says Bob Fung, vice-president of engineering at Owens Design, Fremont, Calif., an automation services company that creates, designs and builds equipment for the semiconductor, disk drive, solar, flat-panel, LED, fuel cell and mobile electronics markets.

Chris White, project manager for automation at CMD, Appleton, Wis., agrees. CMD built an automated packaging system based on a Yamaha robot and vision system that eliminates most of the problems with fixtures, part-positioning, orientation, part defects and product changeovers. "The vision system provides the robot with the position and orientation of the part, allowing the robot to pick and place the product as desired," he explains.

Matt Wicks, vice president of product development at Intelligrated, Mason, Ohio,, a materials handling specialist, says his company uses vision in applications when the material handling operations require more understanding about the environment, material or products. "We recently demonstrated a robotic de-palletizing application (Figure 1) that used 2-D and 3-D vision to determine the location and orientation of cases stacked on a pallet," Wicks says. "Then the robot was commanded to de-palletize the products onto a takeaway conveyor. This type of operation wouldn't be possible without the use of vision-guided robotics."

In some instances, seeing the parts and performing actions could be performed by sensors, but cameras are often a better solution. "As the price of vision systems drops, they're replacing sensors as they provide more functionality and flexibility," Fung adds. "For example, in the past we had to use multiple sensors or move the product past the sensors to determine the quantity and orientation of parts in a carrier. A single camera now can do the same without moving the part, and it can support product changes without changing hardware."

A bakery used a 2-D vision system to guide cake-decorating robots until about a year ago. It upgraded to 3-D to improve flexibility and increase throughput. (See "Robotic Cake Decorator" sidebar)

How Hard Is It?
Adding vision to a robot isn't always easy. "There are challenges with the seamless integration of sophisticated vision, robotics and the accompanying automation systems," Wicks says. "This advanced level of integration requires a collaborative effort on the part of the vision systems providers, robotic arm manufacturers and robotic systems integrators because it requires in-depth knowledge of not only the vision piece, but also how it integrates with the robot and other automation equipment."

Fung says that most vision systems must be designed specifically to meet the requirements of the tool into which they are to be incorporated. "This increases the overall tool cost as well as the design time," he explains. "There's a lack of generic, industrial-quality vision systems that can be programmed easily to meet a variety of system requirements. Most vision systems are difficult and complex to operate. Much of the system complexity is due to the high image quality required for most applications, as well as challenges involved in accurately measuring parts that don't conform to regular shapes."

SEE ALSO: Vision Takes Control

Software and lighting pose problems, too. "We've seen advances in both software and lighting for vision systems, but there are many opportunities for further improvement," CMD's White says. "While the software has become increasingly user-friendly, end users still need some proficiency with programming and setting up the cameras when they introduce a new product. Depending on the type of features the customer is looking for, this can be a tedious process of adjusting the camera and lighting to show the intended features. Products have different patterns, colors, sheens and materials that can complicate setup."

Fung agrees and suggests improvements. "Lighting systems must be able to fit into the tool without adding excessively to tool size," he says. "Designers usually have to use very thin, industrial-grade backlighting, which is expensive and further increases overall system cost. Filtering to prevent stray light reflection is also a concern."

Most robot and vision vendors recognize the integration challenges and provide solutions. For example, White explains how Yamaha software helps, saying, "The software program in the Yamaha RCX controller is a modified form of BASIC. The commands are actually quite simple. For instance, a MOVEP command (move to position) tells the robot to move to a given X,Y,Z coordinate. The Yamaha firmware and hardware determine the most efficient path and control all the motors to arrive at the destination as smoothly as possible and within the defined limits."

Steve Zhu, director of business development for Teledyne Dalsa, describes a system the company built at a Honda plant that also uses vendor software to advantage. "Our GEVA1000 vision system and Sherlock software is being used for China's Dongfeng Honda Assembly line, where an ABB robotic gripper is used to grip the car body accurately and reliably (Figure 2). Our GEVA1000 with Sherlock guides the giant ABB robot to locate the car body properly."

1 of 2 < 1 | 2 View on one page

Join the discussion

We welcome your thoughtful comments. Please comply with our Community rules.
All comments will display your user name.

Want to participate in the discussion?

Register for free

Log in for complete access.


No one has commented on this page yet.

RSS feed for comments on this page | RSS feed for all comments