Dr. Ayers introduces the purpose, structure, and limitations of safety metrics, emphasizing that metrics should help leaders understand system performance, predict future risk, and drive action—not simply generate reports. The episode stresses that many organizations misuse metrics by focusing on lagging indicators or treating numbers as goals instead of tools.
This episode sets the stage for the entire safety‑metrics series.
Dr. Ayers explains that effective safety metrics should:
Reveal system health, not just outcomes
Predict future risk, not just record past injuries
Guide decision‑making
Highlight weak processes
Support resource allocation
Drive continuous improvement
Metrics are diagnostic tools, not scorecards.
The episode critiques the overreliance on lagging indicators such as:
Total Recordable Incident Rate (TRIR)
Lost‑Time Injury Rate (LTIR)
Days Away, Restricted, or Transferred (DART)
These metrics:
Reflect past events, not current risk
Are influenced by reporting culture, not actual safety
Can be manipulated through classification decisions
Often drive fear‑based behaviors
Do not help leaders understand why incidents occur
Lagging indicators are necessary but not sufficient.
Dr. Ayers emphasizes the need for leading indicators—metrics that measure the inputs to safety, not the outputs.
Examples include:
Hazard identification
Hazard resolution
Training completion
Equipment maintenance
Field engagement
Risk assessments
Quality of controls
Leading indicators help leaders:
See risk before it becomes an incident
Identify weak processes
Strengthen systems proactively
Build trust with workers
According to the episode, strong metrics are:
They point to a specific behavior or process that can be improved.
Frontline workers and executives should interpret them the same way.
Data must be reliable and consistently collected.
Metrics must reflect real hazards and real work.
They should predict future performance, not just describe the past.
Dr. Ayers highlights several traps:
Using metrics as goals instead of tools (“We must hit zero injuries” creates fear and underreporting.)
Focusing on quantity instead of quality Counting inspections without evaluating their effectiveness.
Measuring what’s easy, not what matters Convenience often replaces relevance.
Failing to validate data Many organizations discover their numbers are inaccurate.
Ignoring context A high number of hazards found may indicate strong engagement, not poor safety.
Strong safety leaders:
Look for trends, not isolated numbers
Use metrics to ask better questions, not assign blame
Pair leading and lagging indicators for a full picture
Share metrics transparently with workers
Use metrics to prioritize resources
Treat metrics as conversation starters
Metrics should drive learning, not fear.
A site reports:
Zero injuries
Low hazard identification
Low training completion
Poor equipment maintenance
On paper, the site looks “safe,” but the leading indicators show a high‑risk environment with weak systems and low engagement.
This is why leading indicators matter.