EQUINE CLICKER TRAINING.....
                                                               using precision and positive reinforcement to teach horses and people
 
         

 il.com

Clicker Expo 2008:  Lexington, Kentucky

 

 

        I was fortunate to be able to attend Clicker Expo in Lexington, Kentucky in March 2008.  This was my third expo and like the ones before, it was a great experience. I met some very nice people, learned a lot, and had the opportunity to share ideas with other clicker trainers.  I thought I would share some of the highlights of my experience because I think that this expo generated a lot of new ideas for me about how to clicker train horses more efficiently and more creatively.  I need to point out that Clicker Expo runs for 3 days with 3 time blocks of lectures and labs each day. Within each time block, there are five choices for what to attend, so every attendee puts together his or her own program and the sessions I am going to describe are just those that I attended.

 

    Friday morning starts with an opening session where Aaron Clayton presents some information on who is attending (what states we come from, what kinds of jobs) and then the faculty are introduced.  Don't hold me to these numbers but someone told me there were about 400 attendees with 150 dogs.  The sessions I attended were:

 

Ken Ramirez:  "Working for the joy of it: A Systematic Look at Non-Food Reinforcers."

 

    Ken Ramirez is the vice president of animal collections and training at the Shedd Aquarium in Chicago and he is always a good presenter with a nice mix of stories and information. The focus of his talk was how to condition and use non-food reinforcers to provide reinforcement variety in your training.  He started by explaining the difference between primary and secondary reinforcers. Primary reinforcers are those things that inherently satisfy a biological need and include food, water, shelter, air, reproduction, and safety.  Secondary reinforcers are reinforcers that acquire their value through association with a primary reinforcer.  There are lots of different kinds of secondary reinforcers including event markers and keep going signals.  For the purpose of his discussion, he was focusing on what he called "reinforcement substitutes" which are reinforcers that a trainer can use in place of a primary reinforcer.  I want to point out that this session was an "advanced" level one and he was very clear that novice trainers should not be using reinforcement substitutes without supervision or guidance from more experienced trainers, although there are exceptions.

 

    Reinforcement substitutes are useful for trainers for those times when a primary reinforcer cannot be used either because it is not available or because the animal does not want it. He explained how they used reinforcement substitutes with one of the dolphins when it was sick and would not eat. They needed to feed her medicated fish so they actually got her to eat the fish by using reinforcement substitutes in training sessions and including some fish eating as a desired behavior. One of the points he made with this example is that the value of reinforcers can change so it is useful to have several reinforcers available to give the trainer a choice of what to use in any given situation.

 

    He presented detailed information on how to condition reinforcement substitutes. He uses classical conditioning to associate the new reinforcement substitute with a primary reinforcer.  The procedure starts with choosing a reinforcement substitute you want to use. He gave us some guidelines for choosing these. It is easier to start with something you think the animal might find reinforcing, but you can start with something neutral.  He does not recommend starting with something the animal finds aversive.  And he is very specific that even if you choose something that you already think the animal finds reinforcing, you still need to go through the conditioning process because just because an animal likes something, that doesn't mean it will work for it.  Later, he mentioned that one of the cautions about using reinforcement substitutes is that in order for them to be strong reinforcers, you need to control access to them.  Therefore if your animal has a behavior that it finds strongly reinforcing, and you want to condition it as a reinforcement substitute, you may need to limit access to that reinforcer under some situations.

 

    The trainer is going to start associating the reinforcement substitute with the primary reinforcer by presenting the reinforcement substitute and then immediately presenting food. He does not click because he does not want the animal to think it needs to "do anything" to earn the primary reinforcer.  At this point the reinforcement substitute is just indicating that primary reinforcement is coming. He does this for many sessions over weeks, months and maybe years, as long as it takes for the animal to make the association.  This is the first step in the process.

 

    The second step is to start using the reinforcement substitute in training and see if the animal accepts it. He had some general rules to follow.  Each step is repeated multiple times.

 

    1.  Start with a easy, well established behavior. For the first reps, ask for the behavior, click, present new reinforcement substitute, follow with primary

    2.  Mix in some times where you follow the click with the reinforcement substitute and NO primary. Only do this 3x max per session.

    3.  Now start using reinforcement substitute after a harder but well established behavior, by asking for the behavior, clicking, offering the new reinforcement substitute, followed by the primary

    4.  Start following a harder but well established behavior with the reinforcement substitute only. Again, only do this 3x max per session.

    5.  Increase use of reinforcement substitutes but do not use more reinforcement substitutes than primary reinforcers. His general guideline is 20/80 max but this can be over multiple sessions so you can have some sessions where you use more reinforcement substitutes than others.

    6.  You must continue to pair the reinforcement substitute with the primary reinforcer on a regular basis to keep the association strong.

    7.  With beginners, he had more specific rules for introducing reinforcement substitutes gradually. These were never use a RS (reinforcement substitute) after 2 consecutive behaviors, avoid using the same RS twice in a row, always ask for behavior followed by a primary reinforcer more often than a secondary, and keep the association strong by reinforcing reinforcement substitute regularly.

 

    He showed how to use reinforcement substitutes in chains and take advantage of Premack.   He gave examples of chains where there were very few primary reinforcers and how this initially looks like a variable reinforcement schedule. But included in the chain, he had behaviors that had been conditioned as reinforcement substitutes so when you identified those in the chain, it showed that the animal was getting steady reinforcement.

 

    When talking about introducing reinforcement substitutes, he said that it is important to keep in mind the animal's expectations.  This means that if you have used the same reinforcement for every click, the animal expects that type of reinforcement. If you now try adding in reinforcement substitutes, you have to be very careful to avoid frustration because the animal has expectations about what follows a click. I found this particular detail very relevant because 99% of the time, I reinforce my horses with food.  If I want to start adding in substitute reinforcers, I need to do it slowly and read my animal to make sure I don't create frustration or confusion.

 

    He also pointed out that not everyone needs to use reinforcement substitutes and not every animal does well with them. He wanted to share how he conditions them because he finds that people are using them whether they are aware of it or not, and he finds that behavior can break down if the trainer is depending upon reinforcement substitutes that have not been systematically conditioned and introduced. 

 

    On a personal note, this session had some "ah-ha's" for me.  I found the discussion on expectations very interesting because I have run into some of that and not been quite sure how to address it. I also found it added to my understanding of how to create stronger chains.   From Kathy Sdao, I had gotten the piece that cues are reinforcers and they help to keep chains strong. From Ken, I got a better idea of how to mix in reinforcement substitutes as components of a chain to further strengthen it and to avoid having to keep clicking, stopping and treating as I am working my horse.

 

Steve White:  "No Problem!  Solve any Training Issue in Four Steps"

 

    Steve White trains police dogs, but recently he has been working with people and he found that some of the problem solving strategies that people use in their jobs work well for animal trainers too.  He presented several models for how organizations use problem solving (OODA loop, SWOT, drill down, IDEAL, SARA) and his own DIP-IT.

 

    The components of DIP-IT are Define the problem, Isolate the problem, Plan your remediation, Implement your plan and Take another look.   These were the basic steps and could be used for solving problem in current behaviors or training new behaviors. He pointed out that teaching new behaviors is as much about problem solving as is working on modifications to existing behaviors.   Some of the problem solving pitfalls that people encounter are focusing on the problem, asking too much, working too long and being short-sighted.  He reviewed the 10 rules of shaping (Karen Pryor) and pointed out that he believes the ultimate solution to many training issues is to change the motivation.

 

    He believes that many people have problems because they stop training too soon. Good trainers overtrain.  They also train equal and opposite behaviors. The example he used for this was a DVD clip of a police dog learning to be called off the person wearing the sleeve. I saw this movie last year and it made a big impression on me.  In the movie, they have a police dog that is happy to go out and grab the "bad guy" but ignores requests to let go or to abort his run toward the victim if he is called off.   To address this issue, they sent the dog toward one person who offered no reinforcement by remaining passive as the dog approached. They called the dog off and at the same time, they presented a new "bad guy"  and sent the dog to him.  To do this, the dog  had to leave the first "bad guy" and go to the new "bad guy" who gave the dog a good reinforcement by being very reactive to the dog's "attack."  The dog learned to listen to the handler because the handler always knew where the best reinforcement was. When I saw this last year it made a big impression on me about how we can teach our horses that listening to us is the best way to earn reinforcement instead of getting carried away offering their own favorite behaviors.

 

    He showed us a training form, the taproot form, that he uses with his dogs and students. It is a way of keeping data on what behaviors are being trained so that the most important behaviors get the time they need. He has examples on his web site in the libraries section (www.i2ik9.com).  The form has boxes for behaviors with the most important ones in the middle and additional behaviors on each side. The trainer works on a behavior in the middle and then can go to one of the "side" behaviors but then has to return to a middle behavior.

 

    He ended with some general thoughts on training.  Training is a mixture of art and science.  If you want to break the "rules," then you need to know the rules.  It is important to be creative and think outside the box, especially with problem solving where most people are tempted to lose their sense of the where they are in training process. He pointed out that knowing where you are is very important and you have to be realistic about it.  He compared it to using Mapquest to get directions. Mapquest doesn't care where you have been, it just wants to know where you are now.

   

 

Jesus Rosales-Ruiz: "Broken Clicks: How Reinforcer Delivery Impacts Learning"

 

    Jesus Rosales-Ruiz and his graduate students are studying various aspects of operant conditioning and clicker training.  He did a series of experiments to see how a delay in the presentation of the treat after the click affected the animal's behavior.  Before he described the studies, he went over some basic information about the use and selection of the click (or other marker signal) as a conditioned reinforcer. 

 

    He believes that the clicker has two functions. It serves as a marker signal to strengthen and select behavior. It also serves as a cue to tell the animal to go collect its reinforcement.  To illustrate this, he went back to some basic information about how to condition the clicker from BF Skinner. He showed how the click ends up being the first step in a chain which is click, approach (or turn head), eat. The conventional thought has been that you pair the clicker with the food using classical conditioning, but he thinks that it is more useful to think that we are teaching that the clicker is a cue to approach the trainer with the expectation of getting reinforcement. With this in mind, he explains why it is so important to maintain the one to one ratio of clicks to treats (or other reinforcement).

 

    If the trainer is not consistent about reinforcing after each click, then the use of the clicker as a cue that reinforcement is available is no longer reliable.  This matters because if the click is no longer a reliable predictor of food, the animal is going to spend time trying to identify other predictors of food and you can end up with an animal that is easily distracted and spends a lot of time looking at the handler for information he should be getting from the click. This is inefficient and adds a level of confusion to the training process. He had some movie clips of dogs trained where they were not treated after every click and in both cases the dogs got confused, showed frustration behavior and their level of performance decreased significantly.

 

    The use of treatless clicks also poisoned the whole training experience so that their general attitude and enthusiasm for the training sessions decreased so much that they no longer enjoyed the training sessions. When the student was first training them using a one click:one treat ratio, the dogs were excited to get started. When she switched to two clicks: one treat, they would avoid the student when she came and the owner had to go get them for the sessions.

 

    He then did another experiment with adding in extra clicks during a training session.  His student trained a dog to touch two foot targets by using a one-click:one-treat ratio.   The dog was focused on the task and very prompt and clear about what behavior it needed to do.  The student then added in a click for the first target touch, but still clicked and treated after the second target touch. The dog did keep working but she seemed to get confused. She looked to the handler more often than before and there were hesitations as she went from one target to another. 

 

    They did another experiment that was similar, but instead of using a treatless click to mark the first target touch, they used the word "bien." This worked out ok.  In this scenario, the word "bien" was being used as a keep going signal and it did not interfere with the one-click:one-treat ratio. The dog seemed to interpret it as additional information that she was working correctly although I did not see any big difference between the behavior as trained with only a click for completion of both target touches vs. trained with "bien" as a keep going signal.

 

    The remainder of his talk focused on what happens when there is a delay between the click and treat.  According to his earlier statement, the click acts as both a marker signal and the start of a chain  What happens if other behaviors creep into the chain when there is a delay between click and treat?  He showed results of a targeting experiment with sheep. They asked the sheep to target and either reinforced immediately or delayed for 5, 10, and 20 seconds and measured how much other behavior occurred between click and treat.  As the delay increased, more frustration behavior (hoofing, nibbling the target stick, walking away) happened between the click and treat.

 

    He also shared how Virginia Broitman and Sherri Lipman did the same experiment with their very clickerwise dogs and also found that the dogs repeated behaviors that occurred between the click and treat. Not only that, but after the experiment was over, they were unable to completely eliminate the frustration behaviors that those dogs had inadvertently gotten reinforced for doing.  The same types of changes in behavior were shown on movie clips studying the effect of a 5 second delay on targeting behavior in monkeys.   In all cases, the monkey's behavior deteriorated. Not only did they add a lot of superstitious behavior between click and treat, but after receiving the treat, they did not immediately go back to the target. Some of the behaviors they showed were submissive or frustration behaviors. 

 

    From these studies, he has concluded that while it is common to say that the clicker allows us to have a delay between click and treat, this is not as straightforward as it seems.  A delay in delivering the treat can allow the animal to include superstitious behaviors and also lead to frustration and confusion over what was being clicked.  Someone asked the question about what if we have a behavior we are training where we cannot immediately deliver the reinforcement, such as working at a distance or at speed.   In that case, he thinks the delay is not a problem as long as the process of getting and delivering the treat starts as soon as the click is given, and the animal can see that the trainer has started what he calls the "reinforcement cycle."  What he stressed was that it was important to recognize that the click starts a behavior chain that ends with the delivery of the treat.  If the trainer keeps this chain intact and always does food delivery in the same way, it is not an issue.  I think what good trainers really need to do is ensure that the animal is not throwing a lot of behaviors in between the click and treat.  This can be accomplished by keeping the time between click and treat short or having a set routine for presentation of the treat so that the animal patiently waits for the treat.

 

 

Ken Ramirez:  LAB:  "Next!  Finding and Using New Reinforcers"

 

    This lab was a chance for people who attended Ken's lecture to practice with their dogs. I attended as an observer and got to watch several people start conditioning new reinforcement substitutes with their dogs. He started by reviewing how to select a good reinforcement substitute and suggested that people practice it without their dogs if the substitute was a hand signal or some kind of physical gesture. I suppose you might practice a verbal reinforcer too to make sure you were saying it consistently. 

 

    A few different people demonstrated with their dogs.  The important thing to watch for is the dog orienting toward the handler when it hears or sees the reinforcement substitute. We had one woman who wanted to use a thumbs-up, another who wanted to use a verbal and one who used clapping. The dog that belonged to the woman who chose clapping took a while to figure out that clapping was not a new cue for begging, which he offered a lot. Ken had her shorten the duration of clapping and present the food before the dog had a chance to offer to beg.  He stressed that it is important to keep the time between the reinforcement substitute and primary reinforcer short so that the dog does not throw in other behaviors, but there does need to be a delay in order for classical conditioning to take place.

 

    I enjoyed watching the dogs work and it was interesting to see how some of them were fine with food just appearing and others seemed to want to figure out why the food was appearing. It was good to see some conditioning of reinforcement substitutes to see how it worked out and how the dogs handled it.

 

Morten Egtvedt and Cecilie Koeste:  "Reliability: Thy Name is Backchaining"

 

    Morten and Cecilie had a number of presentations and I only attended the one on backchaining.  They are top level obedience competitors who also run a chain of dog training schools and publish dog books and videos in Norway.  They have a different approach to training in that they do not put behaviors on cue until they are completely finished. Rather than put early versions on a temporary cue, they do not cue the dog at all. The dogs figure out which behaviors they might want by context (handler position, presence of equipment) and run through their repertoire until they get clicked.  The covered all this material in the lecture and lab I did not attend, but a friend of mine attended and gave me the basic outline.

 

    The lecture I did attend was on how they combine behaviors together by backchaining.  Most of the behaviors they teach are used in obedience trials and the dogs perform them in a set order. To achieve precision and speed, they backchain finished behaviors to prepare for competition.   They stated that they never forward chain if they can backchain.  They have found that backchaining produces faster and more reliable performance because the dog knows what is coming next and is anticipating the cue.  The success of backchaining is due partly to Premack where a less probable behavior is reinforced by a more probable behavior.  In a backchain, the last behavior has the strongest reinforcement history and is therefore more likely to occur. Therefore you can use that behavior to reinforce the previous behavior and so on. When you create a chain in this way, the animal is working from less probable behaviors toward more probable behaviors and therefore it builds enthusiasm.

   

    They showed some examples of how they backchain.  If the chain is short, they just start by reinforcing the last behavior (4) and then asking for the last two behaviors (3 & 4). They reinforce for  3 & 4 until they are really fluent.  The explained that they do not backchain until each individual behavior is very fluent. A lot of errors in chaining come about when trainers try to chain together behaviors that are not fluent enough. They make sure that the dog really knows each individual piece before they create the chain.  Once 3 &4 are fluent, they add in 2 and practice 2 & 3 & 4 until fluent before adding in 1.

 

    If they are building a longer chain, they might make two mini chains separately and then combine them using overlap, if they can. So they might train a chain with behaviors 1, 2 & 3 and another chain with 3, 4, & 5 and then combine them. The overlap helps to make the connection strong.   There are other ways to keep the chain strong. One way is to reinforce the last behavior with a really premium reward. Another is not break the chain unless absolutely necessary. One of the problems with forward chaining is that you have to occasionally reinforce each component of the chain to keep it strong, which reinforces individual behaviors, but breaks the chain.  By making sure each behavior is perfect before backchaining them together, they can avoid having to break out individual components of the chain, which might strengthen an individual component, but breaks the chain. 

 

    They had a few other guidelines. They test each chain 5 times to see if it is reliable. They emphasized that the strength of chains comes from using positively trained behaviors. Behaviors that have a "do it or else" component will actually weaken a chain.  If they have taught a backchain and the dog makes an error, they abort and try again. If the dog makes an error again, then they go back to looking at individual components of the chain.  If the dog makes an error, it could be that they need to go back and work on one component, but they try to correct the problem without breaking the chain if possible.

 

    The last thing they talked about was how dogs will go through a "testing phase" to see if they can skip steps of the chain and still get their reward. They stressed how important it is that the animal goes through this phase and Cecilie said she would never compete a dog that had not tested her.  In most cases, if the dog skips a behavior, they can just prevent it from being reinforced and ask it to start again.  They had a movie clip of a dog checking blinds for a person. The dog has to run a figure 8 pattern checking every blind until it gets to the last one where the person is hiding with a sleeve.  They backchain this search sequence so the dog always knows that the person is in the last blind. When they start backchaining, if the dog skips and goes directly to the last blind, the person just steps out and does not give the opportunity to bite the sleeve which is the reinforcement for this exercise.  Usually just removing the reinforcement is enough to convince the dog to complete the entire chain.

 

    Their presentation had a nice mix of explanations and movie clips. They showed some very good footage of dogs being backchained. They also had the audience participate in some activities by pairing up and backchaining each other. I found this was interesting as I could feel how the backchaining did build anticipation. This also showed how important it was to be very clear about each individual behavior. Sometimes I was cued to do the next behavior, and I started to do it without stopping the previous behavior. One chain was turn, clap, sit and I found myself sitting while clapping until we got more specific about how many times I was supposed to clap. I could also see how keeping the cues in the chain let the trainer use anticipation to their advantage. Prior to this, I had thought that part of the point of creating a chain was to give the final chain one cue instead of retaining the cues for each individual behavior. But Morten and Cecilie keep the cues in place, which means they can control the timing of when the dog goes from one part of the chain to the next.

 

    I am very intrigued by the idea of using more backchaining and I am trying to think of ways to implement this in horse training. So much of riding is about using one behavior to set the horse up for the next behavior that forward chaining seems more obvious, but I have sometimes taken advantage of anticipation by careful selection of the a behavior that improves the one before it and I think this is part of what backchaining is all about. For example, if I have a horse that is slow at the walk, I will do a lot of walk trot transitions, clicking for the trot.  As the horse starts to anticipate the trot, the walk will improve.  In the past I have been clicking the improved walk, but if I think of this as a backchaining exercise, clicking the walk might not be necessary because I can reinforce the improved walk by asking for and rewarding the trot as well.  I think this is definitely worth pursuing and I am putting together some ideas for things to try with my horses.

 

Kathy Sdao: LAB: "What a Cue can do in Action: cue control"

 

    This was the second of two labs that accompanied her lecture.  I did not attend her lecture or the first lab, but I wanted to see her teaching so I signed up for this lab. I spent the winter watching her DVD's so I was hoping to be prepared.  Kathy is a dog trainer from Washington State with extensive experience with marine mammals and a very dynamic and enthusiastic teacher.  She kept us busy with several exercises on cue discrimination and control. 

 

    She started by reviewing the qualities of a good cue. Cues are distinctive, consistent, salient, simple (make sure you know exactly what the cue is - harder than you think), and precious.  By precious she meant that we should be careful with our cues. Don't present a cue unless you are willing to bet money that the dog will do it. 

 

    Then we started off with a cue discrimination exercise.  She had each dog owner write down 5 behaviors their dog had on cue. The helper shuffled the cards and made a random list of 10 behaviors. Then they tested to see how the dog was doing in the new environment. This lab was held in an outdoor tent with blowers going for heat and a lot of distractions.  I noticed that it took some of the dogs a while to get focused on their owners and respond reliably to the cues they knew.  She gave each team a data collection sheet and they recorded how many times the dog was correct.  The owner was to only ask once and this exercise was not about getting the behaviors, but about seeing how the dogs responded to the cues. Every correct behavior has to be clicked. You cannot use variable reinforcement for this exercise as the dogs need the information that they have responded to each cue correctly.

 

    She recommends that dog trainers do discrimination exercises like these on a regular basis to keep the dogs sharp. I think it would also help to make sure that your cues are not morphing over time or getting sloppy.  When she asked for feedback on how the sessions went, a number of handlers reported that they were more careful about their cues than usual.

 

    For the second cue discrimination exercise, she explained how to test to see if you and your dog agreed on the correct cue for any given behavior. She calls this the "prove it" game.  The idea is to vary the cue slightly and see if the dog still responds. So if you wanted to test the cue "sit," you might ask the dog to "hit," "sat," "upsit," or some other variation on "sit." Is the dog sitting for any one syllable word that starts with s? How about any word that ends in -it? If you are using body language cues, you could test your hand gestures by changing some details. Ideas for this included standing on a chair, using the other hand, holding your hand at a different height, making the motion bigger or smaller, wearing sunglasses, wearing a glove, kneeling and so on.  She gave us a handout with suggestions for ways to test cues and said that some people get very creative about doing this. 

 

    Before she let people try this out, she reminded them that they had to decide ahead of time how they wanted the dog to respond to any given cue. Do you want your dog to sit when you say "sat?" Do you want your dog to respond to hand gestures with both hands or just one?  It is important that the dog has the possibility to earn a click each time a cue is presented, but it could earn a click by doing the behavior, or by not doing the behavior, depending upon what the owner has decided. Again, she had people ask their dogs for behaviors in random order so that the dog was not influenced by recent reinforcement for any one behavior.  People had a lot of fun with this. 

 

    This was a fun lab with lots of good training to watch and people were very creative with coming up with variations on cues for the prove it game.  She also handed out a paper that had a list of reasons people might not go at a green light. I am familiar with this list from her DVD's and I think it is a great way to help people understand about cues. In her DVD she makes the distinction between cues and commands. In her view, commands have a component of "do it or..." whereas cues are just indicators that an opportunity for reinforcement is available for a particular behavior. 

 

    To help people understand this, she compares cues to a driver waiting at a red light. The handout lists reasons why you might not go even if the light turned green.  Some reasons are:  you can't see the signal, signal was brief and you sneezed or blinked, didn't recognize signal because it was different somehow (flashing green), distracted by another sight or sound, another overriding signal prevented you (e.g. siren), another car is in your way (inhibition),  unsafe (someone ran a red light), ran out of gas/broke down, and new criterion (standard car stalled on hill).   Her point is that there are lots of reasons a dog might not respond when presented with a known cue and we need to recognize and troubleshoot why instead of assuming the dog chose to ignore it.

 

 

 

Michelle Pouliot: "Training Guide Dogs for the Blind"

 

    Michelle Pouliot presented information on the training program for Guide Dogs for the Blind which has been converting over to clicker training. She presented some historical information about guide dog training and then showed how they are now adding in clicker training. I didn't know much about guide dog training so I was interested to see how they taught some of the behaviors. They use treadmills to teach the dogs to lead and the movie clips showed that the dogs really love the treadmill training.  She had movie clips that showed them teaching the dogs to respond to the collar, ignore food on the ground and back up in a straight line.

 

    She also showed how they teach them to go around obstacles and about intelligent disobedience. They used to teach intelligent disobedience by asking the dog to proceed and then mimicking a fall or bad event. This was stressful on the dogs and with the use of the clicker, they came up with a better way by asking the dog to go forward and clicking before the dog could respond.  The dog learned to evaluate the safety of the situation before responding to the command "forward." 

 

    I thought it was interesting to see how they had added clicker work into their training and I loved seeing how happy the dogs were when they were working.  So even though I am not doing guide dog work, I was happy to have attended this session.

 

Jesus Rosales-Ruiz: "The Poisoned Cue Anew"

 

        The poisoned cue lecture is one that Jesus has presented at several Clicker Expos and he is constantly adding new information to it. He started the session by distinguishing between cues, commands and poisoned cues.  Cues are discriminative stimuli that indicate the possibility of reinforcement as established through positive reinforcement. Commands are discriminative stimuli established through negative reinforcement.  In training a command, the command is presented and if the animal does not comply, negative reinforcement is applied until the animal performs the target behavior at which point the aversive stimulus is terminated. I think it is important to note his terminology here because he does specify that the animal must find the use of negative reinforcement aversive in his example.  Through this scenario, a command becomes a conditioned aversive stimulus.  The presentation of the command can be used to decrease behavior, or removal of the command can be used to shape and capture behavior.   Because of the way we use negative reinforcement with horses, I think we have to look carefully at how we are using negative reinforcement and how our horses interpret our use of negative reinforcement.  In most of his examples, negative reinforcement was used as more of a correction than as a teaching tool.

 

    He argues that a poisoned cue is a cue that is ambiguous because it has been trained with both positive reinforcement and the use of aversive stimuli. I want to note here that the first time I attended this talk, he said that a poisoned cue was one that was trained with both positive and negative reinforcement. So he has changed his terminology somewhat here. He now refers to aversive stimuli instead of negative reinforcement.  I think this is important for horse trainers who use so many pressure and release cues.  It is the ambiguity of the cue that causes the problem.  Possible effects on behavior from poisoned cues are reluctance in the trainee with signs of stress, behavior breaking down both before and after the cue, longer latencies, freezing and avoidance behavior.

 

    There are three ways to create a poisoned cue. The first is to add aversive stimulation to a positive reinforcement program. The second is to teach with aversive stimulation where the good behaviors were positively reinforced and the incorrect responses were "corrected" with an aversive stimulus. The third way is to elicit behavior with an aversive stimulus and capture it with positive reinforcement.

 

    He showed a movie clip of how they created a poisoned cue with a miniature poodle. The poodle had prior experience with clicker training and no experience with aversives.  There were two parts to the experiment. In the first part, they taught the cues. The cue "ven" was used to teach the poodle to come using only positive reinforcement. The cue "punir" was trained by presenting the cue (saying "punir") and giving the dog 2 seconds to respond. If the dog did not respond, it was pulled to the handler and then clicked and treated in position.  They repeated this over a hundred times.  The leash correction was discontinued somewhere around the 60th trial.  They measured the tail height (high vs. low), presence of whining, and snorting. He showed a movie clip of some of the training sessions and the dog's body language for "ven" was bright and enthusiastic. For "punir," the dog looked depressed, with a low tail and low energy. He also pointed out that in order to present the cue, the dog had to be a certain distance away. In "ven," the dog willingly waited at a distance for the cue. In "punir," the dog stayed so close to the handler that she had a hard time getting the required distance.  It seemed as though the dog hoped to avoid the leash correction by staying right next to the handler.

 

    In the second part of the experiment, they used the cues "ven" and "punir" to capture a new behavior.  They chose several behaviors such as the dog's position in the room, back stepping, and touching an object. For example, when they were working on capturing stepping back, the handler said "ven" when the dog stepped back. The dog returned to the handler and got his reinforcement.  In "ven," the dog figured out the game quickly and offered the new behavior with a lot of enthusiasm and the handler was able to maintain a high rate of reinforcement. In "punir," the dog seemed hesitant and slightly aimless. Even if the handler said "punir" and reinforced the dog for the new behavior, it did not immediately return to the new behavior that caused the handler to say "punir." In some cases, the dog created chains where it consistently performed the behavior that had been marked by "ven" even though that was never reinforced during the "punir" sessions.

 

    Once they were done capturing behaviors with "ven" and "punir," the handler did some additional experiments to rule out the presence of the leash and harness used in "punir" as being the source of the dog's attitude. The dog was fine with the leash and harness in other settings.  In the original experiment, the dog was loose for "ven," but on a leash and harness for "punir," so they switched and used the leash and harness for "ven" and had the dog loose for "punir."  There were some small changes in the dog's attitude, but "ven" was still significantly different than "punir."

 

    From this study, Jesus and his student concluded that combining positive and negative reinforcement during training can have detrimental effects.  Emotional behaviors produced by this procedure do not disappear over time despite the use of positive reinforcement and that while a poisoned cue can be used to mark and select new behavior, the results will be different than using a cue that was trained with positive reinforcement only.   He said that he had gotten a lot of requests to repeat this study but he was reluctant to put another dog through the correction process. Then he realized that there are a lot of cues out there that are already poisoned and they didn't need to poison a new cue, they just needed to find one that was already poisoned.

 

    As an example, he showed some work they did with Caesar, a German Shepherd, who reacted to the presence of the leash as a poisoned cue.  If the leash was in the house, the dog happily engaged with his student and responded correctly to behaviors such as "sit." As soon as the leash came out, he started ignoring and avoiding the student. This was with the leash on the ground, not attached. They put the leash in various locations and the dog did not respond enthusiastically until the leash was back in the house. 

 

    I have gone into this in such detail because I think poisoned cues are a real problem for horse trainers, especially if we start with crossover horses. And no matter how careful we are with our rein handling, there are going to be times when the reins are used in a manner that the horse finds aversive.   I am still trying to work out in my brain if the fact that we educate the horses about negative reinforcement means that we can use negative reinforcement without having poisoned cues.  With my own horses, I have used negative reinforcement to shape a lot of behaviors and they don't seem to view those as poisoned cues. But I do have a few issues that have come about where I think that the cue has gotten poisoned and I would like to avoid that in the future.  I am going to take a closer look at how I use negative reinforcement to both shape and maintain behavior.

 

    One of his early statements was that cues or commands trained only with -R (negative reinforcement) are better than poisoned cues because it is the ambiguity of the poisoned cue that creates the problem.  This makes it seem as if it would be better to just use -R alone, instead of using -R and +R (positive reinforcement) combined, which is something a lot of horse clicker trainers do on a regular basis.  But I don't see that trend in my training and with other people's horses.  On the contrary, horses trained with -R and +R combined are much happier and more willing than if -R alone is used.  I do think that if you use escalating pressure and the -R becomes too aversive, that it negates the effect of the positive reinforcement so my guess is that there is a balance here that works.  In any case, I think that horse trainers need to be very aware of poisoned cues because we are somewhat limited in what we can use for cues under saddle and we need to make sure that we don't poison them.

   

Theresa McKeon: LAB:  "TAGteach in Action"

 

    I attended the TAGteach lecture at Clicker Expo 2006, but had not attended a lab so I signed up for the lab to get more information on what TAGteaching really looks like. Theresa McKeon led the discussion and activities and started us off by making a TAGulator out of beads.   She had us string 10 beads and then gave us ideas for how to use them. I had thought of TAGteach as being used in a formal teaching situation, but she described how she uses it to keep track of things in her daily life. She might TAG herself each time she drank a bottle of water or when she made a decision about whether or not to have a cookie. I have to say that this was a big light bulb moment for me as I suddenly realized that there were lots of places in my everyday life where I could TAG myself for doing things.

 

    She talked us through a few exercises on how to implement TAGteach. The first exercise was just practice tagging other people. We divided up into groups of 3 and one person tagged while the other two people passed a balloon back and forth. The TAG point might be touching the balloon with two hands, or calling the balloon.  I learned the importance of picking a good TAG point. If the people were too close together, it was important to pick a really obvious TAG point as there was not much time to decide if the person should get tagged. It is important to pick a TAG point that is easy for the trainee at the start, and easy for the tagger to identify.

 

    She had us do another exercise where we set up "runways" by laying out two strings and putting objects between them to create an obstacle course. One member of the team was blindfolded and another team member guided her/him through the runway using instructions of 5 words or less. This exercise demonstrated that it is hard to explain something in 5 words or less unless you have previously decided upon some definitions, and part of the ability to set good TAG points was knowing your trainee.  A further demonstration on picking TAG points showed how the trainee could define her own tag points by saying what she was feeling in her body when it was correct.

 

    There were some demonstrations by a TAGteach team from Quebec who use TAGteach in their dog training. They start by training the handlers, not the dogs and they use tagging to teach basic positions such as a hand position for heeling, how to handle treat delivery and clicking the handler for recognizing when the dog is giving them attention.  Theresa also showed with a helper how it was important to be unemotional and specific about TAGpoints. She went over some basics about good tagging including having a short phrase to describe the TAGpoint (5 words or less), keeping emotion out by saying "The TAG point is..,"  not correcting or saying that she is changing anything, but just stating "the new TAGpoint is.." She also stressed the importance of separating out information from praise. Praise should come at the end so that the trainee just pays attention to the TAG which is the most important information.  Some people find praise distracting and she warned of the danger of creating praise junkies.

 

    This was a very entertaining session. She had all the attendees involved and she presented enough information to give me some good ideas about how to use TAGteach. Later in the final wrap-up with Karen Pryor, they played a movie clip showing different applications for TAGteach and I am amazed at the breadth and scope of where TAGteach is being used.  The movie showed TAGteach being used by commercial fisherman, painters, teachers and in various sports applications.

 

    In addition to these sessions, I had a lot of great discussions with other clicker trainers. There are so many interesting people who attend these expos and I learned something from each one of them.   

 

Here are a few other odds and ends that I pulled out specifically to share with my horse friends. 

 

Jackpots:

 

There was talk recently on one of the lists about using jackpots and whether or not they were useful. I do use jackpots, but I have to say that I have never really studied whether they work for me or not.  At clicker expo, there were a variety of opinions about jackpots and from listening to the trainers, the idea I got was that if you are going to use jackpots, make sure they work for you.

 

For example, Kay Laurence does not use jackpots (as defined by multiple goodies, not different goodies). She points out that if your animal has just performed a really good effort, you want to have it repeat the behavior right away before it forgets what it did. You don't want to have a big time lapse (due to chewing) between a successful effort and the next effort. I think this is a good point.

 

I think the key here is to think about all aspects of the treat from choice of treat to delivery and make sure they all work to your advantage. Here are some of my common strategies.

 

Vary the treat. Give better treats for better efforts that are better because the animal added enthusiasm and energy. When I taught Willy Spanish Walk, I found out he would lift his leg higher for an apple than for a carrot. He likes apples better so it made him more enthusiastic and generated a better Spanish Walk.  If I am training a behavior where I am trying to get the horse to relax or calm down, I am more likely to jackpot by offering more of a lesser quality treat. Chewing is calming and I don't mind the time lag between offered behaviors. Sometimes I can even have the horse eat while performing the behavior. This works if you are going for duration on the mat. You can feed many treats while the horse is standing on the mat and this will strengthen the behavior.

 

NO Reward Markers:

 

This was discussed at a panel discussion. Most of the trainers do not use NRM. If they do, it is for very specific circumstances. They do not use it for shaping.   In one example, a NRM marker was used to show a dog the difference between an incorrect behavior and one that was correct but that the trainer was not going to click at that moment. In this case, no information (no click, no NRM) means the dog is correct.

 

Cues:

 

I was introduced to the ideas of cues as reinforcing behaviors. I think that with horses, we do a lot of redirecting. I often ask my horse for something because I want him to do something else, not because he has done one thing correctly and I am ready to do the next thing. If you want to chain together many behaviors, the horse needs to know that being asked for the next cue means he was successful, not that you are redirecting him. One way to do this would be with a keep going signal or verbal praise.  My thoughts now are that while I might not use words in shaping, I will use them when I start to connect together several behaviors.

 

Differences between horses and dogs:

 

It is tempting to try and apply all the same ideas from clicker training for dogs to horses. In many cases, such as husbandry, and simple discrete behaviors, this works. But with riding, I think we have to be a bit more creative.

 

If you want the horse to catch on quickly to the idea that the next cue means success, then you need to start chaining behaviors together early. It is tempting to work on the walk and then the trot and then the canter and reinforce each effort, going for refinement. I think the horse needs the trainer to start chaining behaviors together early, even if it is just asking for a few behaviors in a row. The horse needs to really understand that each new cue is a new opportunity for reinforcement.

 

 

 

 

Katie Bartlett, 2008 - please do not copy or distribute without my permission

   

 

 

 

Home | Articles | Clicker Basics | Community | FAQ | Getting Started | Horse Stories | Links | Photos | Resources

 

Equine Clicker Training