Page 26 - JSOM Fall 2018
P. 26

By comparing endpoints, the contribution of metrics to perfor­  TABLE 1  Five Stages of Skill Levels 13
          mance assessment can be inferred. We thought beforehand that   Novice.  The novice has no background or experience in his or her
          performing faster would challenge users more than performing   area. Rules and objective attributes are applied without an under­
          slower. The challenge posed by time­only metrics (i.e., ≤60s and   standing or knowledge of the context of the situation.
          ≤30s) was as expected: The graphed line of each metric’s per­  Advanced Beginner.  The advanced beginner demonstrates margin­
          formance ended at a different point representing that metric’s   ally acceptable performance based on experience acquired under
          count of failures. However, there was an unexpectedly wide   the mentoring of a more­experienced nurse or a teacher. The larger
                                                              context of the situation is difficult to grasp at this stage. There is a
          gap between the maximum endpoint (31) and the minimum   concern for good management of skills and time, but the need for
          (3). That difference (28) in magnitude was surprisingly large.   guidance and assistance remains.
          Likewise, effectiveness and ≤60s were less challenging than ef­  Competent.  Competent performers are able to differentiate between
          fectiveness and ≤30s, and its magnitude (27, 37 − 10) was sim­  the aspects of the current situation and those of the future and can
          ilar. Parsing results by metric component, the greater impacts   select those aspects that are important. The focus on good manage­
          occurred irrespective of whether the metric consisted only of a   ment of time skills remains, but the sense of responsibility is higher.
                                                              However, they may have an unrealistic concept of what they can
          time component or the metric also incorporated effectiveness.  actually handle.
                                                              Proficient.  Proficient performers are able to see the whole situation
          We also thought beforehand that effectiveness alone would   in context and can apply knowledge to clinical practice, identifying
          contribute little to the failure count, because we had picked   the most salient aspects and differentiating them from those that are
          participants who, on average, were more experienced than   less important. Actions are intuitive and skilled. They have confi­
          the general public. As expected, effectiveness contributed little   dence in their own knowledge and abilities, and focus less on rules
                                                              and time management.
          (7). Because the two magnitudes 28 and 27 contributed by the
          metrics with time components were nearly identical, whereas   Expert.  Expert performers are able to focus intuitively on solutions
                                                              to situations without having to explore alternatives. This ability is
          the magnitude contributed by effectiveness alone was at least   based on a rich experiential background. Focus is on meeting pa­
          threefold less (27/7), we inferred that, in this study, the time   tient needs and concerns to the point of being an advocate for the
          metrics were more impactful on assessment than effectiveness.   patient and care. The focus on self and one’s own performance is
          Although these five plots were of the same physical perfor­  diminished.
          mances, the methods of evaluating the performances through
          the different lens of each metric occasionally changed the as­  FIGURE 3  Cumulative sum (CUSUM) for mean times from 10 users
          sessed outcome by a substantial amount.            among 20 uses.

          The limited capacity of effectiveness to discern interuser differ­
          ences in performance tended to result in an easy pass in almost
          every use (193 of 200). Furthermore, when a metric was least
          challenging (in this case, ≤60s), almost every use was a pass
          even when bleeding was not stopped. When only accounting
          for speed and ignoring effectiveness, the assessments of perfor­
          mances were incomplete. Judging fast ineffectiveness as a pass
          was also misleading because the main point was to control
          bleeding. In addition, misunderstanding a metric created an
          illusion of skill, and a pass without control is a fallacy. To see
          the fallacy is to realize the illusion and vice versa. Among the
          criteria defining a pass, the details determined how the criteria   for all 200 uses. At use 6, a change in trend was detected by
          count, while ignoring such details caused problems.  the software as mean times started to decrease significantly.
                                                             The score was useful in detecting change in the overall trend
          Parsing ineffectiveness by user, seven failures occurred among   of all users as a group, but this finding was absent among
          five users, of whom two had a pair of failures each. We cate­  10 individuals. Further, when we tried to overlay 10 trends
          gorized users by levels of skill (i.e., novice, advanced beginner,   of individual users, the result looked like a horse tail with 10
          competent, proficient, and expert) using an established model   hairs meeting at the bottom right corner at the zero value for
          (Table 1),  and the five users were either competent or profi­  the 20th use. The tail was tangled, obscuring individual curves
                  13
          cient (three and two users, respectively). All five competent or   and differences among curves. Furthermore, the score was
          proficient users failed, whereas 0% (none of five) of the others   not designed specifically for learning curves but was intended
          failed (two novices, an advanced beginner, and two experts).   to detect changes after a learning curve has ended. In other
          Two of the three middle levels of skill had the failures.  words, a learning curve  is to discern  when a learner enters
                                                             a controlled state of proficiency or “is in control,” whereas
          The cumulative number of failures showed meaningful trends,   CUSUM detects when they go “out of control.” Surprisingly,
          and we progressed to a tool specifically designed to detect   we found here that a substantial change in performance oc­
          changes in trend.                                  curred as performance improved at use 6. Beforehand, we had
                                                             not expected such an improvement, because we had selected
          Cumulative Sum Trends                              so many experienced users in the sample. Thus, we decided to
          The cumulative sum (CUSUM) method was used to detect   proceed with an appropriate tool to analyze learning curves.
          shifts in mean times (SAS JMP, version 13; SAS Institute,
          https://www.sas.com). CUSUM scores among 10 users were   Learning Curve Cumulative Sum Trends
          calculated for the mean times by use numbers 1–20 (Figure   for a Less Challenging Metric
          3). Each score represented the difference between the mean   Learning curve cumulative sum (LC­CUSUM) trends were pre­
          time of each use and a benchmark: a mean time of 22 seconds   determined with calculated values to decide proficiency at a


          24  |  JSOM   Volume 18, Edition 3 / Fall 2018
   21   22   23   24   25   26   27   28   29   30   31