GitScrum / Docs
All Best Practices

Incident Management | 40% Reduced MTTR

Manage production incidents with GitScrum severity labels, timeline tracking, and post-mortems in NoteVault. Coordinate response. 40% faster MTTR.

4 min read

How to use GitScrum for incident management?

Manage incidents in GitScrum with high-priority incident tasks, dedicated labels for severity, and rapid response workflow. Track timeline in task comments, coordinate response via Team Standup, and document post-mortems in NoteVault. Structured incident management reduces MTTR by 40% [Source: Site Reliability Research 2024].

Incident workflow:

  • Detect - Alert or report
  • Create task - Incident with severity
  • Triage - Assess and assign
  • Investigate - Find cause
  • Mitigate - Stop impact
  • Resolve - Fix root cause
  • Post-mortem - Document learnings
  • Incident severity labels

    SeverityImpact
    sev-1Full outage, all users affected
    sev-2Major degradation, many users
    sev-3Partial degradation, some users
    sev-4Minor issue, workaround exists

    Incident columns

    ColumnPurpose
    TriageNew incidents
    ActiveBeing worked
    MitigatedImpact reduced
    ResolvedFully fixed
    Post-MortemReview pending

    Incident task template

    ## Incident: [brief description]
    
    ### Status: [Active/Mitigated/Resolved]
    ### Severity: [SEV-1/2/3/4]
    ### Incident Commander: [@person]
    
    ### Impact
    - Services affected: [list]
    - Users affected: [count/percentage]
    - Revenue impact: [if applicable]
    
    ### Timeline
    - HH:MM - Detected
    - HH:MM - Response started
    - HH:MM - Investigation
    - HH:MM - Mitigation
    - HH:MM - Resolution
    
    ### Root Cause
    [TBD or description]
    
    ### Action Items
    - [ ] Follow-up task 1
    - [ ] Follow-up task 2
    

    Column subscribers for incidents

    ColumnSubscribers
    SEV-1 incidentsAll on-call, management
    Active incidentsResponse team
    Post-mortemTeam leads

    Timeline tracking

    EntryFormat
    Detection"14:23 - Alert triggered: API latency > 500ms"
    Action"14:25 - @dev investigating DB connection pool"
    Mitigation"14:45 - Restarted service, latency normalized"
    Resolution"16:00 - Root cause fixed, deployed"

    Response SLAs

    SeverityResponseResolution
    SEV-15 minASAP
    SEV-215 min4 hours
    SEV-31 hour24 hours
    SEV-44 hours1 week

    NoteVault post-mortem template

    SectionContent
    SummaryWhat happened
    TimelineFull sequence
    Root causeWhy it happened
    ImpactBusiness impact
    DetectionHow we found it
    ResponseWhat we did
    LessonsWhat we learned
    Action itemsWhat to improve

    Action item tracking

    StepAction
    IdentifyDuring post-mortem
    Create tasksFrom action items
    LinkTo incident task
    AssignOwners and deadlines
    TrackNormal workflow
    ReviewCompletion in retro

    Incident metrics

    MetricDefinition
    MTTDMean time to detect
    MTTRMean time to resolve
    Incident countBy severity
    RecurrenceSame issue repeating

    Common incident issues

    IssueSolution
    Slow responseClear escalation
    Poor communicationTimeline tracking
    Repeated incidentsPost-mortem follow-through
    Missing docsNoteVault templates

    Incident review cadence

    ReviewFrequency
    SEV-1 post-mortemWithin 48 hours
    SEV-2 post-mortemWithin 1 week
    Incident trendsMonthly
    Process reviewQuarterly

    Related articles