Welcome to Health Care POV | sign in | join
Stepwise Success

Embracing Point of Care
July 15, 2016 6:20 AM by Scott Warner

For an industry that is frequently at the forefront of new technology in healthcare, laboratory workers can be among the most resistant to change. Computers coexist with paper; manual diffs are still done when automated counts are far superior; point of care technology is disdained as inferior to central lab testing.

But point of care testing is coming into its own. Indeed, as hinted at in a recent Advance article, microfluidics will bring the core laboratory to the patient. Investors are shelling out big bucks for this idea of “lab on a chip” that will make the current crop of glucose meters look as outdated as 1980’s cordless phones. Our world is changing rapidly.

This doesn’t sound like good news for an industry already on the ropes with centralized testing, testing formularies, labor shortages, and demands for faster, cheaper testing. What will happen? I wonder.

Traditionally laboratory professionals have been leery of point of care testing, because non-laboratorians don’t understand the technology or how values are verified. If I ask a lab tech, “How do you know that glucose is accurate?” I will get a detailed response: the QC is within peer-defined limits, the last calibration is acceptable, the other patients I have run meet a statistical norm, etc. If I ask I person expected to do point of care testing - such as an RN - how he or she knows a glucose is accurate, I’ll probably get a less specific response.

But that’s no reason for a lab not to embrace point of care. As point of care testing becomes simpler and more foolproof, laboratories should step up to the plate and make sure they are there to enhance services and improve patient care. This change is inevitable as technology advances.

“Foolproof” is a misnomer as we all know just like a zero-maintenance analyzer. Errors are often subtle, and an understanding of what affect results is necessary to produce quality results. It’s what separates laboratory training from all others. Who better to shepard in a new era of technology?

The lab is unique in that it provides technology useful in point of care. We likewise have a unique opportunity to educate, encourage, and engage others in what we do. This isn’t easy, but it is inevitable. How easy has this been in your lab? I wonder. There are considerable cultural barriers between the evidence-based focus of laboratories and the support-based focus on other departments.

Has your lab embraced point of care testing? If so, are you an active partner with doctors and nurses or is it a necessary evil brought about by affordable technology?

NEXT: The Joys of Shopping

To Cut Costs, Change
July 1, 2016 6:20 AM by Scott Warner

These days it’s all about change. I have heard a constant drumbeat for the last thirty years that change is the only constant we can count on. The only thing more constant than change is the need to cut costs. Now that laboratories are becoming cost centers in small hospitals and groups are recognizing the economies of scale in centralizing testing and developing formularies to cut costs, those two are intertwined.

In one sense, it’s easy to save money: just search for a cheaper vendor. The problem for smaller hospitals is less bargaining power. A Group Purchasing Organization (GPO) can help but hinder depending on compliance terms. Small labs just don’t have the clout, especially in the Critical Access world, to convince vendors. The market charges what consumers will pay the world around. So cutting costs by going cheap without sacrificing quality only goes so far and as a strategy is quickly exhausted.

An easier scenario is one understood by every bean counter out there: cut staff. As a rule of thumb payroll is half of all expenses. If full time can be cut to part time, hours reduced, positions eliminated, or managers replaced with “working” managers, it all looks good on the bottom line. When the above is exhausted, this is the next logical step. Labs are benchmarked against each other to determine staffing load, positions are lost through attrition, and managers are challenged to come up with new ways to do more work with fewer people. It’s too bad benchmarking can’t capture the hard work that goes into making a good lab a great lab. And in fact benchmarking doesn’t take quality into account at all.

When that is done, what’s next? Many labs are facing the question today. They’re compliant with GPO contracts, they have evaluated pricing and chosen the cheapest without sacrificing quality, they have evaluated “make it or buy it” to cut costs, and they have lost people to never be replaced.

The answer is a hard one: change.

As an industry we have to change what we do, how we do it, and how we work to deliver a service that is cheaper, faster, and that doesn’t sacrifice quality. Doctors (and the insurance companies) do not care about the how and why, but they want completely separate things. The first wants speed, reliability, and quality. The second wants the lowest cost. Maybe that means outsourcing, merging, or centralizing. Maybe, it means changing.

If the world around us has changed, we must change with it. This requires creative thinking. It requires rethinking what we have done for the last ten or twenty years. It means trying completely different approaches to getting work done. For example, you may have a test kit in your lab that requires more QC if nonwaived and less if waived, and the only difference is the sample type. As another example, maybe moving a workstation to another department will streamline workflow and improve turnaround time.

Only we can change. If we wait for outside forces to change us, it’s too late.

NEXT: Embracing Point of Care

Do Doctors Read Comments?
June 18, 2016 6:40 AM by Scott Warner

Laboratories add comments to reports, some of which are informative e.g. CRITICAL VALUE REPEATED and others that are interpretive e.g. explaining the meaning and utility of the MDRD estimated GFR equation. It is the latter that brought me to this current question.

Most doctors have little or no idea how results are generated. I think they assume that professional, trained staff under the supervision of a pathologist give them the best number possible. A GFR estimated from a creatinine, age, and sex alone is a perfect example. What if a doc assumes that this is accurate for dosing? It is not, generally, and will tend to overestimate GFR in the elderly, the infirm, and those otherwise with extremes in muscle mass. It is intended to be a screen for chronic kidney disease in a subset of patients with a normal to slightly elevated creatinine value.

The question is how much of this instruction needs to appear on a laboratory report? And will the physicians read it?

Second question first: no.

In my experience unless a result is completely obtuse or difficult to interpret physicians do not read comments for an explanation. They are conditioned (and we are too, when you really think about it) to respond in a knee-jerk fashion to flags generated by the information system. The nuances of utilizing a test are either a) well known by the physician, or he or she would not have ordered the test initially or b) lost in the comments. “Interpretive comments” exist to cover the laboratory.

The real issues here aren’t covering the laboratory or informing the physician, which are two separate issues. The first relates to producing quality results and is boolean in nature: it is or isn’t, regardless of comments. The second is an education issue that will never be resolved by a lengthy comment. And a lab that thinks so is doing a disservice to the patient. No, the real issue is, “How do I add value to the report?”

I had a conversation with a physician recently that brought this issue into sharp focus.

He said, “We can add a comment to any test, such as potassium. But should we? Unless it explains anything we need to know about the particular test, does it add anything?”

That is an excellent question that most laboratorians and many pathologists have difficulty answering. The only way to know for sure is to ask your medical staff what is helpful to them. And you might be surprised.

Glucose is another example. A fasting glucose <= 100 mg/dL is considered normal. But we report out many glucoses on samples drawn throughout the day. Should all laboratories add a comment that states “Glucose ranges are for fasting patients only?” What about non-fasting patients? What about diabetics? What about patients on steroids? There are far too many contingencies, most of which the attending physician is more than aware of, for any laboratory to address on a report. And it would be near crazy to try.

The litmus test (if there is one) is “Does this comment tell the docs something unique about this number?” And that can be a hard call. The best approach (in my experience) is to ask them. You could be surprised.

NEXT: To Cut Costs, Change

Justifying Staff
May 31, 2016 6:41 AM by Scott Warner

These days it’s all about shortages. Shortages of techs, shortages of patients, and shortages of money. In small hospitals there are fewer of us working with fewer patients for less money. Those who are working are older than the average worker, are wondering who is replacing them, and are tired of hearing about doing more with less when new technology requires people to test, validate, and perform the assays. This is what I’ve been hearing for the last few years.

A few random thoughts on this.

A laboratory test menu is constantly in flux as tests are brought in house or sent out to reduce cost or improve quality, but what I hear now is, “If you send out that test can you reduce hours?” I find myself justifying what little staff I have more and more, and I suspect many managers would say the same thing.

Payroll is a huge portion of ongoing expenses, so that’s understandable. I get that reducing expenses is crucial to managing a dwindling cash flow and can make or break a hospital in a competitive market. But the reality of managing laboratories is different from other departments.

Few benchmarks: there are benchmarking factors, but laboratories are so different in employee mix, services, and outreach that it’s difficult to compare them in a meaningful way. I’ve been in labs with many phlebotomists, for example, and some where techs performed most of the phlebotomy. It all depends on how far away a phlebotomy station is, how versatile the information system is, and other factors. Equipment varies greatly from lab to lab, and not all instruments offer the same mix of quality and speed. “Efficiency” varies from lab to lab, often having little to do with the skill level of techs.

Make or buy issues: bringing in tests to justify staff is fine if it’s cheaper to perform a test in house, but that can be a boondoggle if it requires new instrumentation, more maintenance, more training, and more competency. It’s been my experience in general that people are poor multitaskers. Asking people already doing multiple tasks to do one more creates a drag on overall efficiency and a chance to increase errors that will drag a system down even more.

A manager caught in a feeding frenzy of cost cutting has to recognize benchmarking and other comparisons for what they are: an attempt to manage expenses using verifiable data. That can be a smart idea in a big picture sense. But a manager also has to use dwindling resources with a mindset that these issues aren’t going away, necessitating new ways of thinking about old problems. Our futures in this industry will likely be shaped by more than “make it or buy it,” outsourcing, or cost cutting. We can only do so much of this, and in the meantime the demand for faster, better laboratory results is increasing. What we do has never been more important.

This could mean different workflow models, consolidated platforms, software AI, or something completely different. Whatever it turns out to be for our individual labs, it has to be invented under constant pressure to do more with less. Inventing new ways to produce better care may be the best way to justify staff we have.

NEXT: Do Doctors Read Comments?

Slide Review or Manual Diff?
May 18, 2016 5:59 AM by Scott Warner

The CPT code 85004 (Blood count; automated differential WBC count) has many variations, each of which is charged instead of 85004 and including the work. These codes can be distinguished in the CPT code book, because they are indented. Examples:

  • 85007 (blood smear, microscopic examination with manual differential WBC count)
  • 85008 (blood smear, microscopic examination without manual differential WBC count)

In other words, performed a manual differential if a CBC with automated differential is performed is not billable in addition to the code 85004. All of these are “bundled” into 85004. Neither is a slide review, a spun hematocrit, or an automated reticulocyte count. All are considered iterations of a blood count with an automated diff, recognizing that the term “CBC” encompasses variations.

One of which is: should you reflex to a slide review or a manual differential?

The consensus rules from the ISLH (International Society for Laboratory Hematology) suggest the former. A slide review is a targeted review of a peripheral smear specific to the particular parameter or instrument flag. For example, a white blood cell count of >30 thousand reflexes to a “slide review” if it is the first time or a delta check failure within 3 days. That slide review logically involves checking for abnormal cells or cells that suggest a leukemoid reaction. A manual differential is necessary only to enumerate abnormal cells.

The slide review concept is an attractive idea in many ways: it takes advantage of technologist judgment, targets a review based on accurate instrument readings, and avoids the busy work of just reflexively banging out a 100-cell diff that may not add value to the report.

More significantly, a slide review reinforces what we already know: the instrument is much, much more accurate that we can ever be. Performing a manual differential to “prove” the instrument count is OK leads physicians down a path that doesn’t trust our technology. Performing a slide review sends the message that we are using the instrument to guide our workflow and look for abnormalities.

In my experience that’s been an easy sell for docs but much harder for staff. Almost all the techs I’ve known have had a knee jerk reaction that a slide review is more work. Why?

For one thing, after performing thousands of manual differentials it becomes second nature. It is comfortable, familiar, and repetitiously easy. Having said that, what is the first thing we all do when we finish? We compare the numbers to the automated differential. We might even repeat the manual differential if we think the numbers don’t match, which is really nutty behavior when you think about it. This begs the question, “What are we really doing?”

This could just be the shock of the new. We have to use new technology to change our diagnostic techniques. As suggested by the ISLH, part of this is slide reviews designed to look for what the instrument suggests. But can we teach old techs new tricks?

NEXT: Justifying Staff

Is PCR Ready for Small Labs?
May 2, 2016 6:36 AM by Scott Warner

What is true for big labs eventually becomes true for small labs, mostly because volume discounts drive affordability. This is most recently true for PCR, a technology that has arrived in small laboratories for two platforms, the Meridian Illumigene and Nanosphere Verigene.

But is PCR ready for small labs?

I’m intrigued by this technology, and I can easily imagine it playing a role in small laboratories. This kind of platform offers rapid, definitive testing for infectious agents such as C. diff, pertussis, and bacterial agents in blood culture specimens. Testing a stool specimen for the C. diff toxin can eliminate the need for GDH antigen testing that detects non-toxin producing strains. Identifying bacteria sooner in a blood culture can give the physician a 24-hour head start. In theory this faster turnaround time with better results will reduce length of stay and assist antibiotic stewardship programs.

I wonder.

One reason I’m on the fence is expense. It’s great to claim that a new instrument will reduce length of stay, but bean counters aren’t impressed by soft cost savings that can’t really be quantified. Unless one is testing for something completely new that changes a protocol - the introduction of BNP comes to mind - it can be hard to convince bean counters that a faster turnaround time equals fewer inpatient days.

Another reason is wondering how having these results will change treatment. Clinicians embrace technology at different rates. How many of you are still running cardiac troponin and CKMB, for example? How many are still reporting percentile differentials with absolutes? How many are still performing more ESRs than CRPs? It sounds great to me to give the docs a better, faster result, but clinicians have to buy into new technology and use it to its potential. As primary information consumers they always drive demand for technology.

Finally - and this is a big reason - new platforms with radically different technology require a lot of training and competency assessment to get off the ground. In a small lab staffing is often sparse, raising questions such as, “What happens on weekends?” and “What happens when a doc wants it done STAT in the middle of the night?” The advantages of rapid PCR testing imply STAT requests, after all. Labs with squeezed budgets and payroll will have a difficult time fitting anything new and different into their menus, no matter how wonderful. It all takes time, money, attention to detail, and bodies to run the tests. In the current healthcare climate it can be hard enough just to get the bread and butter tests done in a timely fashion as more and more labs adopt a rapid response model.

I don’t know the answer, but I don’t hear physicians beating down the door for in-house PCR just yet.

I’m interested in how this technology has affected your laboratory. Has it lived up to the hype, and what problems have you encountered? Is it worth the expense?

NEXT: Slide Review or Manual Diff?

Moving That Needle
April 21, 2016 3:15 PM by Scott Warner

One of the phrases I hear lately is “we need to move the needle,” meaning enough effort has to be put into change to not just make it stick, but change what matters. This might be customer satisfaction scores, test volumes, or cost containment.

If there’s one thing that change has taught me, it’s that no matter how much things change they seem to stay the same. The needle almost never moves.

I’ve witnessed countless alphabet soup campaigns, LEAN initiatives, customer service gimmicks, changed hours, changed protocols, and new technology. It isn’t so much that each change is filled with false promises or defeated by dashed hopes. In fact, we can all be easily convinced that any of these things can make a difference. They almost never do. Change is a constant in a world stubbornly set in the present rather than anticipating the future.


There are several good reasons. We are motivated by emotions, not numbers. Data is an excellent rationalization tool, but it won’t convince people to change how they feel about something. We each have a gut feeling about what works and what doesn’t based on the culture we work in. That is an extremely powerful force to try to overcome from within. Indeed, it may be impossible for a negative culture to change itself.

We are also motivated by leadership, a quality rare enough that we don’t just know it when we see it - we are surprised. I think leadership is a skill like many others that can be acquired with a good working knowledge of what it is. Without good leaders who can articulate a vision and make decisions that cement values in place, change is just pointless change that has no lasting effects.

Finally, each of us views change as something different. The classic management wisdom is “people hate change,” which is misleading. People love change if it means making their lives more convenient. What they don’t like about change varies enormously. Some don’t like change because they don’t trust the motives behind it; others don’t like it if they weren’t included in the decision making; still others have too much stress in their life outside of work to deal with one more change. Management frequently forgets that change happens everywhere in life, and often a workplace is the most stable environment a person has.

So, how do we move that needle?

For me why people are “resistant” hints at the answer. We need leaders who are not afraid to motivate people with emotions and who understand that our work lives are but a part of whatever we are going through. We need leaders who can articulate in plain, blunt language the stakes. We need the values to be articulated honestly and plainly enough to be supported by human interest stories that really motivate people. Leadership needs to walk the talk.

But that’s just me. What about your lab?

NEXT: Is PCR Ready for Small Labs?

Limit Input
April 13, 2016 4:59 PM by Scott Warner

One of the oft-quoted nuggets in The Elements of Style is “omit needless words.” I’ve seen this rule praised and criticized with equal fervor. As a writer, pruning and trimming prose seems like a necessary path to clarity for the reader. It is also intensely personal and driven by one’s own style. But in general, it’s a good rule of thumb. If it takes 4 words to say the same thing in 15, use 4. (Have politicians read Strunk and White?)

Unfortunately as bench techs we don’t have the option to rewrite and revise with time to mull over or forget what we’ve written. What we tell physicians has to be clear, concise, unequivocal, and reproducible. The more judgment that is involved in a test, the more difficult this can be and the more variation one encounters.

Variation dilutes quality, clinically significant or otherwise, and invites STAT abuse. If, for example, 95% of in-house chemistry tests are completed in a narrow time frame, there is less variation and the process creates consistent expectations for providers. They are less likely to order a test STAT to bump it to the top of a queue that has widely variable turnaround times. To a lesser extend, limiting variability in reporting non-numeric values builds similar expectations and can reduce telephone calls, repeats, and requests for further review.

I described a way to simplify urinalysis and reduce variation in a recent blog. While seemingly trivial, this kind of approach reduces the number of choices made by techs, another form of variation that increases the amount of work done.

One of the advantages information technology brings to the laboratory is the ability to reduce variation by defining result entry parameters. For numeric results this can mean entering a lower reportable limit in a test definition that causes the system to automatically change a value below the assay range to less than that range. For alphanumeric (text) values this can mean limiting the number of choices to a drop down list or selection box.

A simple example is limiting qualitative testing results to Positive or Negative. This can be improved in a few ways, perhaps. Positive can be offset with characters e.g. ***Positive, or these can be abbreviated to ***POS and NEG. One problem I have run into with fine-tuning what these look like on native reports is that usually formatting is lost in HL7 translation to another system such as Athena or the T-System. Something to think about.

Other examples: blood bank types, RBC morphology choices, urinalysis dipstick values, Gram stain morphology. Most information systems out there have this capability to limit input in the sense that only a few selections are available to result the test. From an IT perspective this makes sense when assigning SNOMED codes and uploading discrete data to central database systems. The fewer results listed in “comments,” the better.

What about your lab? Have you limited input, and does this improve quality?

NEXT: Moving That Needle

Are You Getting Paid?
April 4, 2016 4:53 PM by Scott Warner

The laboratory is a unique clinical department. The possible tests and their associated billable codes that are routinely ordered day in and day out can number in the thousands or even tens of thousands. The big moving target is referral lab testing, which may change according to where it is forwarded, reflex testing, or method changes. I spend a significant amount of time comparing bills to the charge master. If we don’t get paid, we can’t keep the doors open forever. Are you getting paid?

Some of the issues that make laboratory work complex include:

  • Associated charges that change. Some associated charges are routine, but many reflex or are conditional. This can be a real problem with referral or esoteric testing.
  • Charges that have to be bundled with other charges. Many physicians have patterns, but there are only so many patterns that can be learned. The sheer volume of testing and ordering tactics make this a constantly moving target.
  • Late charges. Many late charges are related to testing that is performed a day or two after collected, such as with microbiology cultures or pathology specimens. Still others are caused by invoices with additional charges.
  • Incorrect CPT codes. CPT codes change year to year, but they also change when methods change in laboratories. This is another constantly moving and often confusing target.

As a manager, you can tread dangerous ground stressing accurate billing. What we do is for the good of our patients. But if you don’t get paid, you can’t keep the doors open. And if you don’t charge for everything accurately, you won’t be able to justify the staff you have. Another lab with the same workload that bills for everything will look better.

I’ve puzzled over this idea of charge reconciliation - how do we know we have charged for everything? - for the past few months. It is not an easy problem for a laboratory to solve. Laboratory testing is complex.

One crucial strategy is to make sure your referral testing - easily the biggest target in terms of change - is accurate. Ask for an electronic version of your bill in Excel. From there it’s easy to compare it to a download from your information system to make sure you’re billing accurately.

Another strategy is to look at duplicates, missing tests, and things that don’t look right from day to day. This is extraordinarily difficult and time consuming in anything but the smallest labs, even more so when short staffed. I’ve created a SQL report that pulls out CPT codes and totals them by test per account number. Anything that could be a panel missing CPT codes or is a duplicate is flagged as an error. This is a great report, but I have a history of computer programming. I’ve no idea how any other lab would do such a thing. But we have to get paid. And if you don’t bill, you won’t get paid.

NEXT: Limit Input

Simplify Urinalysis
March 25, 2016 6:40 AM by Scott Warner

Urinalysis is one of the simpler screening tests laboratories perform. Modern dipstick readers have standardized and simplified the chemical analysis of urine. But what about the microscopic? Shouldn’t that also be simplified?

Beckman Coulter and Sysmex offer instrumentation that performs cell counting. For many if not most labs, urine sediment is examined under light microscopy. White and red blood cells are reported as an average per HPF.

Sounds simple enough. Count the number of white cells in 10 fields, move the decimal point, and report that.

In fact what labs do is much more complicated and more difficult to reproduce with any precision. Cellular elements are reported as a range of cells and are only accurate to the degree where counting or estimating becomes cumbersome. In recent data mining I discovered a wide variation that implies the method is far more precise than possible e.g. 10-12, 15-17, 20-25, 50-60, 70-80, etc. This reflects more tradition than accuracy. More importantly, does it meet the needs of the ordering physician?

At a meeting I asked physicians about clinical efficacy. One replied, “It’s pretty straightforward. I want to know are there a few, some, a lot, or too many to count?”

So we changed reporting as follows:

WBC = 0, <5, 5-20, 20-100, >100

RBC = 0, <3, 3-20, 20-100, >100

While a cutoff of 20 cells per HPF is arbitrary, it is also practical. It’s easy to count 20 per HPF by counting the average per quarter of a field. And while 100 per field can be estimated accurately but counting cells along an arc of the field, a guesstimate works just fine for clinical purposes.

It’s nice to have more reproducible results, even if some physicians may not notice. More importantly, consistent, reproducible results improve quality. To put it another way, when techs report microscopic elements with a wide degree of variation this dilutes quality. A physician may understand the imprecision of a method, but he or she will also guess that finer distinctions won’t matter. “A few, some, a lot, or too many” is what they will read when they see the report. Why not make that easier for them?

Finally, while it may seem that 20-100 is too broad, it is consistent with implied imprecision. Assuming these are probable ranges, 5-20 is an expected 30% coefficient of variation and 20-100 33%. Thus, what we claim is consistently precise; it is easier to estimate 5-20 cells than any significant interval between 20-100 e.g. 50-75. 5 and 3 are cutoffs related to reflex culture criteria and microhematuria respectively.

Of course, arbitrary limits are just that. An implication that 30, 40, 60, and 80 mean the same thing to a clinician may or may not be true. There is also great variation among doctors. That’s why I asked first.

NEXT: Are You Getting Paid?

Working Managers
March 16, 2016 6:38 AM by Scott Warner

A nursing paradigm has been the “working manager,” a nurse who also works the floors and pitches in when needed. This model has been used in ancillary departments such as radiology and respiratory therapy. It makes sense to have a manager with first hand knowledge of the day to day workflow.

I don’t know how common this is in laboratories, but I suspect not very. It seems odd in a small hospital laboratory to have a manager sitting in an office while three or four staff members process specimens, but it happens.

There are some good reasons for this separation of labor, however. One is the complexity of the laboratory. Managing a flow of information that includes ordering, contracts, quality control, proficiency testing, competency assessment, procedures, and a charge master ever in flux is time consuming. Add to that going to meetings, organizing meetings, scheduling, dealing with personnel issues, and handling complaints, and that’s a busy job for anyone.

I’ve been in both kinds of roles. I’ve been a manager when very little bench work was needed and when 1-3 days a week was needed. If one is competent at managing information and people, being a working manager is advantageous.

Advantages: a first-hand knowledge of bench issues, giving credibility to changes that are suggested; greater respect from peers; for veterans, a (mostly welcome) break from the stress that the office often brings.

Disadvantages: can be extremely difficult for rookie managers; often means longer hours to get caught up on office work; sometimes difficult to concentrate fully on bench work with projects in mind; difficult to maintain skills not done as often.

A less obvious advantage is that working the bench can force a manager to delegate many tasks to others to have more time to get critical projects done. These might include submitting quality control, submitting proficiency testing online, scheduling, and even payroll. There isn’t anything special about most office work, and spreading the wealth can make it all the easier when vacations roll around.

A political advantage that should not be overlooked is respect from management peers, especially those working managers who ask the obvious if you don’t work the bench, “What makes you special?”

I wonder: have staffing shortages and cutbacks in small hospitals created more working managers? Do hospital administrations look at other departments and expect laboratories to follow suit? And what happens when someone who has been a manager for 20 or more years is suddenly asked to work the bench? Keep in mind lab managers are asked to do more and more all the time.

NEXT: Simplify Urinalysis

Critical Time Limits
March 7, 2016 5:59 AM by Scott Warner

The Joint Commission National Patient Safety Goal 2 is to improve communication. For a laboratory, this means reporting critical results on a timely basis. Labs have to define critical results, by whom and to whom they are reported, and an acceptable length of time between “availability” and reporting.

Have computers lengthened our reporting window?

In all labs I’ve worked with an LIS, a critical value is called when the result appears on the screen across an interface. The results are on the screen, sometimes highlighted, and sometimes a hard stop forces the user to enter who was called, that a readback was performed, etc. That’s great.

But before computers, I never called after I finished handwriting the result on a lab slip. I called when the repeat result printed on the instrument. That’s where my feet were held to the fire, because that’s when the result was available. (And still is.)

A casual review of our lab has shown a similar perception of “availability” as when it is electronically available. It could be a critical in a stack of routine results, a CBC waiting for a smear review, or something else. Modern analyzers will often autorepeat critical values, further distancing a bench tech from the instrument.

This leads me to think that a critical results policy must define what “availability” means. In the case of transcribed or manual methods, this is often when the result is entered into the LIS - unless your lab still tracks everything on paper. If, for example, a positive C. diff test is a critical result, is it available when you write it on a log sheet or after you go to the computer, log in, navigate to the record, and record the result? Labs typically don’t write times on log sheets. But this is at least one good argument for getting rid of them and adopting real-time reporting.

For analyzers, availability is what it has always been: the time on the printout. As great as information systems are, it can sometimes take a while for results to appear in a queue. In the meantime, the result has been confirmed and available, just like in the days before computers. What I’m suggesting is that computer interfaces have conditioned us to rely on its electronic reporting as a redefinition of availability, when they are really a proxy reporting tool largely disconnected from the testing itself.

Of course many bench techs will telephone an extreme or critical result when it prints. But is this true for all critical results? I’m not sure, but the only way to know is to audit the printout to report time gap. Keep in mind when developing a policy that nursing also has a similar policy with a time limit; their clock starts ticking when they receive the call from the laboratory. If both times are at maximum defined by policy, this can add a significant delay to the treatment of the patient.

NEXT: Working Managers

Know What Docs Want
February 26, 2016 11:48 AM by Scott Warner

A lab tech’s job is very linear. Requisitions arrive, specimens are collected that match the tests that are checked off, the work is done, the tests are reported. We know what docs want. Right?

For the most part, yes. Most docs - especially younger ones - are not afraid to order lab tests. Many docs have favorite tests or favorite referral labs. But the sheer volume and variety of referral and esoteric testing is overwhelming. Even tests performed in-house can vary greatly from lab to lab in terms of sensitivity and specificity. Physicians are bombarded with information on all aspects of care and can’t be expected to keep up on every new development in diagnostic testing.

In other words, docs don’t always know what they want. Not exactly, anyway. That’s where we come in.

We forget how much docs rely on our experience and expertise that can be extended to making suggestions and recommendations. A physician once told me, “You guys run the tests and are the experts. I figure if you tell me a result is good, I have to believe you, because I don’t do what you do.” We need to remember that our specialized knowledge is unique in the healthcare setting. It’s true that docs interpret laboratory testing, but we often know what testing is best or what factors affect a result. As long as we are assertive in the best interests of the patient, the physician will not mind.

This goes clearly against the grain of the older, seasoned techs who learned to work in an environment where the physician was never questioned. It was a weird way to work, in which we were excluded from care plans and decision making and just left to perform the tests like robots.

These days the laboratory should be expected to take a much more active role. At my new position I attend an interdisciplinary team daily at which the hospitalist reviews inpatient cases with those in attendance. This includes nurses, social workers, physical therapy, dietary, pharmacy, and (since I have accepted this position) me. I have a sense it is unusual to have a lab person at these meetings. But I am often asked questions about cultures, sendouts, or blood bank. And if I can contribute, I will.

What’s interesting from a lab tech standpoint is how often the nurses suggest ideas to be supportive. A nurse might say, “Should we order a swallow eval on that patient?” and the doctor will say, “Oh, great idea, let’s do that.” It is apparent to me that docs simply can’t think of everything and remember everything at once. They appreciate help that comes from an interest in the best care of the patient.

This is completely different from the world I was trained in. Lab techs are not trained to be active participants in patient care, which is odd considering how much doctors and nurses rely on lab tests. But to know what the doctor wants, you have to pay attention, ask questions, and sometimes make suggestions.

NEXT: Critical Time Limits

Who Can You Follow?
February 18, 2016 6:29 PM by Scott Warner

According to one site, 1 in 40 adults (about 2.3% of the US population) suffer from obsessive-compulsive disorder. While it’s a serious condition that can be treated and is no laughing matter, I can’t help musing if the percentage of lab techs with OCD is higher. Most techs I’ve known joke about OCD, because the field requires a lot of learned behaviors that can appear compulsive.

On the Oscar to Felix scale of compulsiveness, for example, most lab techs are closer to Felix and the Oscars of the world drive them (OK, us) batty. Oscars make it difficult to step in and help on the bench, and even more difficult to follow work.

I know, I know. We are all different, we all have to get along, and one person’s chaos is another’s organization. I suppose. But it’s also a patient safety issue. If we don’t know what someone else has done, what someone else is working on, or what is pending, results can be delayed or incorrectly reported. Eventually, all techs get burned on an experience like this. We all know those on the bench we are happy to follow even if they leave work behind, and those we hope finish everything before they leave for the day.

Much of the compulsive behaviors I have observed involve checking pieces of paper or using paper to track a workflow, even though information systems are designed to do this with greater speed and accuracy. There isn’t anything wrong with all this paper, but it can create a chaotic mess on the bench. I’ve seen techs put paper in piles, leave it as a trail wherever they go, hang it on multiple clipboards or hooks, create numerous forms to track miniscule details, and order stickers alphabetically by patient name. Some techs are just-so, and some are not.

This isn’t a criticism. I understand the need to have a system that accurately tracks a large amount of information that must be accurate and timely. But our systems make it easier or harder to follow each other.

They are also time and space consuming. For example, if instrument tapes are compared to results in the information system, stickered for identification, saved with the day’s work, bundled at the end of the month, and bundled at the end of the year for storage, that is five times paper is handled (and additional expense to store for at least two years) when it can be thrown out if there is an electronic record. Many instruments allow results to be stored on a thumbdrive, capturing repeats, which allows all paper to be discarded.

We bristle at criticism, especially when our habits reassure us that we aren’t missing tests. But I’ve seen paper lost or misplaced countless times and work missed. It’s a big reason for why information systems are designed to alert users when work is overdue in one way or another. While a computer interface is consistent - unlike our paper interfaces for tracking - the next question is can you follow the computer?

NEXT: Know What Docs Want

Do Older Techs Really Hate Change?
February 5, 2016 5:33 PM by Scott Warner

Do older techs hate change? It’s a cliche that an old dog can’t be taught new tricks. When someone is “resistant to change,” ageism inevitably follows. “She’s old and close to retirement,” I’ll hear. Conversely, young people are assumed to be champions of change, malleable, and willing.

All of which may be true to an extent, but adopting such a view dismisses experience as a variable. The reason “older” techs hate change is because they’ve been there, done that, and have a closet full of tee shirts. It’s true that in time we all become tired of change for the sake of change, but it’s worse to be ignored or dismissed. Decades of experience working the bench includes decades of change, large and small. That counts for something, surely.

And younger techs naturally embrace change because it levels the playing field. A lack of experience prevents you from seeing unintended consequences of change. And all change has unintended consequences, hopefully more good than bad.

As a manager, I’ve always taken the approach that resistance is information. “Why would we do that?” and “That is a waste of time,” to mention two classics, should open a dialogue. And here is where it gets interesting.

Why would we do that? Decisions are based on values, whether we appreciate it or not. We value the patient experience, the support of employees, quality, the bottom line, etc. All managers approach their position with a set of values that overlap with those of the organization. Unpopular change involves unpopular or unknown values. Good managers are consistent in expressing their values through decisions. But when a decision doesn’t make sense or is out of synch, it’s natural for people to ask the question.

That is a waste of time. This is a clear clash of values and vision. As a manager you might see a change as important, and when challenged in this way there are several alternatives. You might, for example, attempt to explain why. You could invite the group to brainstorm alternatives to the change and possibly be surprised. Or you could just pull rank and tell everyone to suck it up. What you choose will be a reflection of your values.

The mantra in management training is that “resistance” is inevitable. I disagree. Like a trick, the acceptance is in the presentation. The last thing that any tech who works a hard 40 hours a week is that he or she will be asked to do more. Or less! Efficiency can be threatening to those worried about losing hours.

Do older techs hate change? I don’t think so, at least not in any meaningful sense. Experience hones instincts and brings wisdom, two qualities any manager needs to succeed. Ask why people hate a change, and they are likely to tell you. Opening this kind of dialogue is a first step to success.

NEXT: Who Can You Follow?



About this Blog

    Scott Warner, MLT(ASCP)
    Occupation: Laboratory Manager
    Setting: Critical Access Hospital
  • About Blog and Author

Keep Me Updated