Forums

Articles
Create
cancel
Showing results for 
Search instead for 
Did you mean: 

Automatic Time Tracking in Jira Service Management

A practical guide with real examples

Automatic time tracking in Jira Service Management isn’t about start/stop buttons—it’s about reading your workflow signals. When work starts, pauses, or finishes, Jira already knows. The trick is turning that history into fair, business-hours-aware metrics for agents and tickets—no timers, no spreadsheets.

Customer Support Dog.png

The honest problem (and why “tab-open timing” backfires)

You don’t want agents babysitting timers. You want timing to “just happen.” The first idea many teams try is to start the clock when the ticket tab opens and stop it when it’s closed. Let's start with the fact that it is generally difficult to track or even set up such a process. Sounds neat. Fails fast. In Scrum, one of the key values is Focus, which this meticulous timer babysitting disrupts. By eliminating manual timing, teams can align more closely with Scrum principles, maintaining concentration on real work rather than getting bogged down by peripheral tasks.

  • Agents open multiple tickets, get pulled into calls, forget tabs—your “auto” timer shows hours of “work” that never happened.
  • Jira doesn’t emit reliable “screen opened/closed” signals.
  • You end up measuring attention, not work.

Better mental model: Track when the process says work started, paused, and finished—not when a browser tab was focused. That means using statuses and assignee ownership, which Jira does record perfectly.

What “automatic” really looks like in a real queue

09:05 Dana picks up TIS-3121 → moves to In Progress.
09:22 Waiting on logs → Waiting for Customer.
11:10 Customer replies → back to In Progress.
11:36 Resolved.

No timers. No extra clicks. Just workflow. From this single history, you can compute:

  • Cycle/Handling time (time in “working” statuses)
  • Waiting time (Waiting for Customer, Blocked, Pending Approval)
  • Lead time (Created → Resolved)
  • Assignee time (how long Dana owned it while it was “working”)

If your team is distributed, apply business-hour calendars to avoid inflating numbers unfairly due to weekends and local holidays.

Pick the clocks you actually need

  • Lead time — Created → Resolved
    Customer experience & throughput.
  • Cycle (handling) time — Work start (e.g., In Progress) → Resolved
    Operational speed without intake/queue time.
  • Waiting time — Time in Waiting/Blocked/Customer
    Separates team delay from external delay.
  • Assignee time — Time an issue is owned by each agent while “working”
    Load balance, coaching, effort proxy.
  • Resolution time — Created → first resolved/closed state
    SLA/OLA staple.

Rule of thumb: Decide once which statuses mean Start, Pause, Stop. Document it. Consistency beats cleverness.

The “automatic” patterns that work (no timers)

  1. Status-based timing — Measure time per status to compute cycle, waiting, review, etc.
  2. Ownership-based timing — Measure how long each assignee owned the ticket while “Working” statuses were active.
  3. Business-hours calendars — Count only working time (per team/timezone), exclude weekends/holidays.

Field notes: how teams actually use this in Time in Status App

1) "Who’s overloaded?" isn’t a feeling; it’s a data.

Assignee time vs. Ticket count report in pivot tables. To ensure appropriate interventions, set a clear threshold: Rebalance when any agent's handling time exceeds 20 hours. If someone has high handling time and high count, consider rebalancing or pairing them with a reviewer. If handling time is high with a low count, you have complexity or blockers, not effort issues.

Pivot Assignee&WI count.png

2) “We’re slow” is vague. “We wait 62% of the time” is actionable.
Group statuses into Working vs. Waiting. If waiting dominates, fix handoffs, approvals, or customer comms before adding headcount.

Status groups.png

3) Did our triage bot help? Before implementing any automation, it's crucial for teams to form a hypothesis. We propose that the triage bot reduce cycle time by 20% within two weeks. Filter tickets with labels = automation_triage. Compare Average time across equal time intervals. If cycle time drops, keep it. If not, iterate.

Average time.png

4) Remote teams need calendar-aware truth.
Kyiv QA and Toronto product? Apply local calendars. “QA took 48 hours” becomes “QA took 6 business hours.” Morale saved; SLA accurate.

calendar.png

5) Reopens and ping-pong are cultural smoke alarms.

Use Transition count to find Techical Progress ↔ Waiting for Customer loops (handoff gaps/unclear requirements) and Resolved ↔ Reopened cycles (quality/acceptance issues).

Transition Count.png

Pitfalls (don’t step on these)

  • Counting weekends/holidays by mistake → set work calendars per team/timezone.
  • No Resolution on Done → fix the transition; it breaks resolution/lead time.
  • Ambiguous “Working” statuses → write the policy once.
  • Chasing “screen-open” tracking → browsers lie; workflow doesn’t.
  • Weaponizing metrics → start with team trends; use individual data for coaching, not policing.

Where Time in Status fits (quietly)

You can implement everything above without manual timers using Time in Status app:

  • Assignee Time — effort per agent (calendar-aware). *скрін* 
  • Time in Status / Average Time — cycle, waiting, review via Status Groups.
  • Time in Status per Date — day-by-day movement & before/after comparisons.*скрін*
  • Pivot View, charts, dashboard gadget, exports (CSV/XLSX/JSON), Confluence macros — share insights where your team already works. *скрін дашборду*

We’re the team behind it and happy to map your workflow to the right clocks in a short walkthrough.

A friendly close (and a tiny nudge)

Automatic time tracking in JSM isn’t about surveillance or start/stop rituals. It’s about clean signals (statuses, ownership), fair math (business hours), and reports that read your history for you. Do that, and your agents stay focused while you get trustworthy operational insight—no spreadsheets, no babysitting.

0 comments

Comment

Log in or Sign up to comment
TAGS
AUG Leaders

Atlassian Community Events