Etr

  • November 2019
  • PDF

This document was uploaded by user and they confirmed that they have the permission to share it. If you are author or own the copyright of this book, please report to us by using this DMCA report form. Report DMCA


Overview

Download & View Etr as PDF for free.

More details

  • Words: 3,837
  • Pages: 12
ROBOTICS 1. Introduction One of the main goals of the study of robotics today is to understand not only the control systems used to stabilize the position, motion, and other forces of the robot, but also to allow robots to effectively work via sensors and actuators in dynamic environments. The field of robotics has its origin in science fiction, the term ROBOT was derived from the English translation of a fantasy play written in Czecholosolvakia around 1920. It took another 40 years before the modern technology of industrial ROBOTS began. Today robots are highly automated mechanical manipulators controlled by computers. First, let's decide what a 'robot' is. To be a robot, it should have the ability to think - make decisions. This may sound hard IF FRONT LEFT WHISKER SENSOR IS ON THEN STOP, GO BACKWARDS 2 FEET, TURN RIGHT, CONTINUE. This is a very common 'IF-THEN' statement. A machine that can perform this instruction is truly a robot. So, the conclusion is that to be called a robot; you really need an on-board brain and a way to program it. Classification of ROBOTS: •

Robot like devices: 1. Prostheses 2. Exoskeletons 3. Telecherics 4. Locomotive mechanism



Classification by coordinate system: 1. Cylindrical coordinate robots 2. Spherical coordinate robots 3. Jointed arm robots 4. Cartesian coordinate robots



Classification by control method: 1. Non-servo controlled robots 2. Servo controlled robots 3. Point-to-point servo controlled robots 4. Continuous-path servo-controlled robots

2. SENSORS IN ROBOTICS The sensors used in robotics include wide range of devices that can be divided into the following general categories: 1. Tactile sensors:  Touch Sensors  Force sensors 

Force sensing wrist



Joint sensing



Tactile array sensors

2. Proximity and Range sensors. 3. Miscellaneous sensors and sensor based system

2. 1. Tactile Sensors: Tactile sensors are devices that indicate contact between themselves and some other solid object.

2.1.1. Touch Sensors: Touch sensors are used to indicate that contact has been made between two objects without regard to the magnitude of the contacting force. Another use for touch sensing device would be as a part of inspection probe that is manipulated by the robot to measure the dimensions on the work part.

2.1.2. Force Sensors: Capacity to measure forces permits the robots to perform a number of tasks. Force sensing in robotics can be accomplished in several ways. A commonly used technique is “force-sensing wrist”. This consists of a special load-cell mounted between the gripper and the wrist. Another technique is to measure the torque being exerted by each joint; this is usually accomplished by sensing motor current for each of the joint motors.

2.1.3 Force-sensing wrist: The purpose of force sensing wrist is to provide information about three components of force and three moments being applied at the end of one arm. Since the forces are usually applied to the wrist in the combination it is necessary to first resolve the forces and moments into their six components. The kind of computation can be carried out by the robot controller or by a specialized amplifier designed for this purpose. Based on this calculations the, the robot controller can obtain the required information of the forces and the moments being applied at the wrist.

2.1.4. Joint sensing: If the robot uses dc servomotors then the torque being exerted by the motors is proportional to the current flowing through the armature. A simple way to measure this current is to measure the voltage drop across a small precision resistor in series with the motor and power amplifier. This simplicity makes this technique attractive. However measuring the joint torque has several disadvantages. First measurements are made in joint space, while the forces of interest are applied by the tool. The measurements therefore not only reflect the forces being applied at the tool, but also the forces and torques required to accelerate the links of the arms to overcome the friction and transmission losses of the joints.

2.1.5. Tactile array sensing: A tactile array sensor is a special type of force sensor composed of a matrix of force sensing elements. The force data provided by this type of device may be combined with pattern recognition to describe a number of characteristics about the impression contacting the array sensor surface. Among these characteristics are (1.) The presence of an object (2.) The objects contact area, shape, location and orientation, (3.) The pressure and pressure distribution and (4.) Force magnitude and location. Tactile array sensors can be mounted on the fingers of the robot gripper or attached to a worktable as a flat touch surface. The device is typically composed of an array of conductive elastomer pads. As each pad is squeezed its electrical resistance changes in response to the amount of deflection in the pad, which is proportional to the applied force. By measuring the resistance of each pad, information about the shape of the object against the array of sensing elements can be determined; in the background is the CRT monitor display of the tactile impression made by the object placed on the surface of the sensor device. Research into potential materials for tactile sensors has lead to the development of a force sensing skin of polyvinylidene di-Fluoride. This is a piezoelectric material which means that it generates an output voltage when it is squeezed.

2 .2. Proximity Sensors: Proximity sensors are used in robotics for the near-field work in connection with object grasping or avoidance. The proximity sensors are further classified as:

2.2.1. Inductive sensors: Sensors based on the change of inductance due to the presence of metallic object are among the most widely used industrial proximity sensors. The effect of bringing the sensor in close proximity to a ferromagnetic material causes a change in position

of the permanent magnet. Under static conditions there is no movement of the flux lines and therefore, no current is induced in the coil. However as a ferromagnetic object enters or leaves the field of the magnet, the resulting change in the flux lines induces the current pulse whose amplitude and shape are proportional to the rate of change in the flux. The voltage waveform observed at the output of the coil provides an effective means for proximity sensing.

2.2.2. Hall Effect Sensors: Hall effect relates the voltage between two points in a conducting or semi conducting material to magnetic field across the material. When used by themselves, Hall-effect sensors can only detect magnetized objects. However when used in conjunction with a permanent magnet in a configuration they are capable of detecting all ferromagnetic materials. Hall effect sensors are based on the principle of Lorentz force which acts on a charged particle through a magnetic field. This force acts on a charged particle traveling through a magnetic field whose axis is perpendicular to the plane established by the direction of motion of the charged particle and the direction of the field. Bringing ferromagnetic material close to the semiconductor-magnet device would decrease the strength of the magnetic field, thus reducing the Lorentz force and ultimately, the voltage across the semiconductor. This drop in voltage is the key for sensing proximity with Hall-effect sensors.

2.2.3. Capacitive sensors: Capacitive sensors are potentially capable of detecting all solid and liquid materials. These sensors are based on detecting a change in capacitance induced by a surface that is brought near the sensing element. The sensing element is a capacitor composed of sensitive electrode and a reference electrode. A cavity of dry air is usually placed behind the capacitive element to provide isolation. The rest of the sensor consists of electronic circuitry which can be included as an integral part of the unit, in which case it is normally embedded in a resin to provide sealing and mechanical support. Typically, these sensors are operated in a binary mode so that a change in the capacitance greater than a preset threshold ‘T’ indicates the presence of the object, while changes below the threshold indicate the absence of an object with respect to detection limits established by the value of ‘T’.

2.2.4. Ultrasonic sensors: The basic element is an electro acoustic transducer, often of the piezoelectric ceramic type. The resin layer protects the transducer from humidity, dust, and other environmental factors; it also acts as an acoustical impedance matcher, since the

same transducer is generally used for both transmitting and receiving fast damping of acoustic energy is necessary to detect objects at close range.

2.2.5. Optical proximity sensors: This sensor consists of a solid-state light emitting diode (LED), which acts as a transmitter of infrared light, and a solid-state photodiode which acts as the receiver. The cones of light formed by focusing the source and the detector on the same plane intersect in a long, pencil like volume. This volume defines the field of operation of the sensor since a reflective surface which intersects the volume is illuminated by the source and simultaneously “seen” by the receiver.

2. 3. Miscellaneous Sensors and Sensor-based Systems: This category includes devices with the capability to sense variables such as temperature, pressure, fluid flow, and electrical properties. Voice programming systems can be used in robotics for oral communication of instructions to the robot. Voice sensing relies on the techniques of speech recognition to analyze spoken words uttered by a human being and compare those words with a set of stored word patterns. When the spoken word matches the stored word pattern, this indicates that the robot should perform some particular actions which correspond to the word or series of words.

3. Evolutionary Algorithms Charles Darwin first identified the process of natural selection in his monumental work The Origin of Species. Certain characteristics that govern an individual's chances of survival are passed to offspring during reproduction. Individuals with poor characteristics die off, making the species stronger in general. Inspired by this natural process of "survival of the fittest," evolutionary algorithms (EAs) attempt to find a solution to a problem using simulated evolution in a computer. There are two related, yet distinct, types of EAs. The first type, genetic algorithms (GAs), involves manipulating a fixed-length bit string. The bit string represents a solution to the problem being solved; it is up to the programmer to determine the meaning of the string. The second, genetic programming, involves generating expression trees as used by languages such as LISP and SCHEME. With genetic programming, actual programs can be created and then executed. Generally speaking, the process used in evolutionary learning begins by randomly generating a population of individuals where each individual is a potential solution to the problem. The population is the set of individuals generated. Each individual contains a genome, the content produced during the execution of EAs. In

the case of GAs, the fixed-length bit string is the genome. Next, each individual in the population is evaluated using a fitness function. Fitness functions use a method of testing solution quality for the respective problem domain. How the fitness of an individual is used depends on the implementation of the evolutionary algorithm. Usually, a higher fitness value corresponds to a greater chance of the individual being selected for reproduction. Different methods of producing the next generation exist, but most commonly, two operators are employed: mutation and breeding. Mutation involves randomly altering one or more genes in an individual's genome. Breeding uses a crossover operation to combine components of two parents' genomes to produce one or more children. Once a new generation is created, the old one is discarded. A cycle of evaluation and reproduction is repeated through several generations, as specified by the programmer. The evolution process ceases after a termination criterion has been met, and the result of the evolution run is the individual found to be best so far. A summary of the components needed for EAs follows as: •

A genetic representation for potential solutions to the problem (i.e. fixed (or variable)-length bit string, expression tree)



A method to create an initial population of individuals



A fitness function that plays the role of the environment, rating solutions in terms of their "fitness"



Operators that determine the composition of children



Values for various parameters, which the evolutionary algorithm uses (solution set size, probabilities of applying genetic operators, etc.)

3. 1 Robot Navigation: As humans, we enjoy the luxury of having an amazing computer, the brain, and thousands of sensors to help navigate and interact with the real world. The product of aeons of evolution has enabled our minds to model the world around us based on the information gathered by our senses. In order to navigate successfully, we can make high-level navigation decisions, such as how to get from point A to point B, as well as low-level navigation decisions, such as how to pass through a doorway. The brain's capacity to adapt has also made it possible for people without certain sensory capabilities to navigate throughout their environments. For example, blind people can maneuver through unfamiliar areas with the aid of seeing-eye dogs or canes. Even without all of our sensors, we are able to cope with familiar and unfamiliar environments.

3. 2. Applying EAs to Robotic Navigation: Evolutionary algorithms have been implemented to solve problems in robot navigation. In particular, EAs have been used to get a robot to learn how to adapt to its limited capabilities. Using GP in this way is termed evolutionary learning. Robot navigation is a difficult problem to solve, and it becomes increasingly more difficult when a robot encounters a failure in either its sensors, devices that tell a robot about its environment, or its actuators, devices that allow a robot to physically interact with its environment. In most other EA applications, two distinct steps occur: an initial training period is conducted by running the EA on a training set, followed by the execution of the best-fit solution. With Continuous and Embedded Learning, the two steps are linked and operate concurrently while the robot is performing its task. Thus, the learning process is continuous. Figure 1 shows an outline of this approach.

Figure 1. The Continuous and Embedded Learning Model.

Key components of the model are: •

Learning continues indefinitely, allowing adaptation to sensory failure.



Learning is done on a simulation model.



The simulation model is updated to reflect changes in the real robot or environment.

3. 3. Performance: In an experiment, a robot was given the task of navigating to the opposite side of a room through a passage in a wall starting from one wall and heading in a random direction from -90 degrees to 90 degrees, with 0 pointing directly to the opposite wall. An example course is shown in Figure 2. The measure of success of CEL was the

percentage of times the robot successfully reached the goal without bumping into walls along the way or running out of time.

Figure 2 A Nomadic Technologies Nomad 200 Mobile Robot was used for the real-world tests. The robot is a three-wheeled synchronized-steering unit that features seven front-mounted sonar units. The monitor in the execution system combines the output of sensors with information about the robot's actuator execution to determine sensor failures. If a frontal sonar unit continuously outputs zero as the distance from the robot to an obstacle but the robot keeps moving forward, then that sensor is marked as having failed. In the experiment, sonar "failure" was simulated by covering the sensor with a hard material. The initial rule-set used in the experiment gave the robot a basic notion of how to get to the other side of the room, but did not take into account obstacles or walls. The simulation model initially assumed all sensors were working. With no evolution and all sensors functioning, the robot was able to navigate successfully about 25% of the time. After 50 generations of evolution, the robot's success rate increased to about 61%. With three sensors on the right side of the robot disabled after 50 generations, the success rate dropped to about 42%, but increased over 50 more generations to about 63%. Figure 3 shows the learning curve.

Figure 3

The results of the experiment show that a robot using CEL can not only learn how to improve its navigation abilities by itself, but also re-learn how to navigate after suffering the loss of some sensory capability. While the results may not be

terribly impressive, improvement in navigation and adaptation to changes in capability are clearly shown. 4. Living on Their Own Ecosystem-like settings are interesting from an alife perspective. Within ecosystems, the main goal of a robot is self-preservation (staying operational for an extended period of time). Resources, especially energy, are limited in time and space. Consequently, robots must compete for them. Consequently, competition forms the basis of all robot interactions in the system. There is a substantial amount of alife research based on simulated ecosystems. The basic ecosystem located at the Flemish Free University of Brussels (VUB) is a 5 m X 3 m space enclosed by walls (Figure 4). Initially it includes simple mobile robots, the moles (Figure 5). The photo sensors, positioned at the front of the robot, are used to navigate and to find objects in the ecosystem.

Figure 4

Figure 5 The arena also contains a charging station, where the robots can autonomously recharge their batteries. The robots drive into the charging stations and make contact with conductive plates that connect them to the energy source. This electrical energy is food for robots. The robots constantly monitor their energy level and in this way know when they are hungry. The robots' whole world revolves around earning and competing for the electrical food. There is a limited amount of food in the ecosystem. Also, the ecosystem contains competitors: small boxes that house

lamps emitting a modulated light. These lamps are connected to the same global energy-source as the charging station; they therefore feed on the same source as the robots. The robots must knock out the competitors (Figure 6). If a robot knocks a competitor several times, the lamps inside the competitor dim, and the robot has an additional amount of energy at its disposal from the charging station. After a while, the competitor recovers and the lamps inside return to their default intensity. The competitors establish a kind of work task for the robots that is paid in electric energy.

Figure 6 Research on robotic ecosystems usually deals with homogeneous agents, which is also the case for the ecosystem described above. A heterogeneous set-up, in contrast, is more interesting for several reasons. First, heterogeneity substantially adds to the complexity of the environment, which is of key interest to alife-oriented research. Second, artificial ecosystems with just one species are hardly biologically relevant. The extended ecosystem features a new inhabitant in the form of the head, which consists of a camera on a pan-tilt-unit and substantial vision capabilities. However, as it is fixed to its position, the head cannot access the charging station, and it is forced to cooperate with the mobile moles. When a mobile robot approaches a pitfall, which it cannot distinguish from the charging station, the head can warn the mobile robot. The mobile robot in exchange can share the benefit of the saved energy with the head. Recent developments have also shown that communication between robots is very important; robots need to be able to share information and intentions. And they have to negotiate their relationship with each other: "I will show you a food source! Let me warn you of danger! What will you give me if I do this?" Recently scientists have started to study how robots can construct a primitive language to communicate exactly these kinds of concepts.

5. Emerging Trends and Applications of Robotics 5.1 Mars Rovers:

Planetary scientists have a unique problem when trying to collect data: in the absence of an interplanetary manned space program, they must rely upon remote sensing techniques. An invaluable part of their arsenal is spacecraft that can study the region of interest in detail. Landers in particular are able to bring sensors and instruments into close contact with the planetary environment. This allows them to take direct measurements of properties such as soil mechanics and elemental composition, as well as image surface features in greater detail than is possible with orbiters or Earth-bound instruments. Landers, however, have the shortcoming that they are limited to a single site for study. An important addition, then, is a mobile robot, which can roam over a much larger segment of the terrain and can carry imagers and other instruments to a variety of features spread over this larger area. The recent Mars Pathfinder mission demonstrated the utility of such an approach. The Sojourner Rover was carried to the surface of Mars by a lander. Over an area approximately 10 meters in radius, it conducted soil experiments in several different terrains, took detailed images of rocks and soils from centimeters away, and placed its on-board spectrometer on 16 distinct targets (9 rocks and 7 soil sites).

Figure 8 Future missions plan to expand this successful technology by incorporating mobile robots (“rovers”) which are capable of traversing even larger distances, carrying their instruments to a wider variety of features and even caching samples along the way. A key requirement for these new planetary robots is greater navigational autonomy, since longer distances than in prior missions must be covered between communications with Earth. The field of planetary exploration offers a rich environment for mobile robotics research, encompassing many complicated issues which must be tackled in order to produce successful missions. In particular, some of the issues related to autonomous path-planning for planetary rovers. With the tools provided by improvements in these areas, mobile robots will prove to be an even more useful and robust addition to the techniques available for planetary exploration. 5.2 Biological Inspiration: When trying to re-create biological phenomena, it makes sense to look at the biological world for some good initial pointers. For instance, the study of robotic

locomotion is aided by observing and imitating biological systems. A robot is limited when using wheels (very un-biological) in rough terrains, whereas a biologically inspired robot would have legs like that of a spider for maneuvering around and over obstacles. Another example may be how a spider-like robot that loses a leg might adjust its gait to complete a mission-critical task. One can look at an experiment performed as an example. Here two legs from opposite sides of a spider were amputated. Two possibilities existed for the spider: it could either retain a diagonal rhythm of locomotion using four legs, but lose its ability to balance; or it could move three legs at a time, offsetting its diagonal rhythm. The experiment showed that the spider chose the second alternative, which was more mechanically stable. This apparently has to do with the fact that although the walking patterns are determined in the spiders “central oscillator,” they can be modified by feedback from sensory organs in the legs. In effect, the spider adapted by reprogramming itself.

Related Documents

Etr
November 2019 17
Etr
May 2020 19
Etr
May 2020 13
Capacitadores Etr
May 2020 8