What You'll Learn
- Distinguish between four escalation tiers: handle yourself, standard, priority, and emergency
- Identify the specific conditions that mandate immediate escalation — active exfiltration, ransomware, compromised privileged accounts, and multi-host attacks
- Construct a structured escalation brief that gives Tier 2 everything they need without a follow-up call
- Apply SLA-driven response timelines (P1–P4) and understand the consequences of missed SLAs
- Execute a complete shift handoff that maintains investigative continuity across analyst rotations
- Avoid the three most common escalation failures: under-escalating out of fear, over-escalating to avoid work, and escalating without investigation notes
The Cost of Getting Escalation Wrong
At 02:47 AM on a Tuesday, a Tier 1 analyst sees Wazuh rule 5712 fire — multiple failed SSH logons on linux-web-01 from an external IP. The analyst has seen this pattern hundreds of times. Usually it is a bot scanning the internet. They mark it as a false positive and move to the next alert.
What the analyst did not check: the same IP successfully authenticated via SSH three minutes later (Event ID 5715 — successful logon after brute force). By 03:15, the attacker had elevated to root, installed a cryptominer, and began lateral movement to dns-server-01. The overnight shift found the compromise at 07:30 when CPU alerts fired across the subnet.
The root cause was not a bad detection — Wazuh caught the brute force. The root cause was a failed escalation decision. The analyst had the right alert, had the tools to investigate, but classified and closed without checking for follow-on activity. A single pivot — checking for successful logons from the same source IP — would have changed the outcome from a 5-hour undetected breach to a 3-minute escalation.
Escalation is not passing the buck. Escalation is the mechanism that connects detection to response. Get it wrong in either direction — escalating too little or too much — and the SOC breaks down.
The Escalation Spectrum
Not every alert needs the same response. Escalation exists on a spectrum from "handle it yourself" to "wake the CISO." Your job is to place each alert at the correct level on that spectrum — quickly and accurately.
Level 0: Handle Yourself
Most of your queue falls here. These are alerts you can fully investigate, classify, and close without involving anyone else.
Criteria:
- Confirmed false positive with documented evidence
- Low-severity isolated event on a non-critical asset
- Known vulnerability scan from an authorized scanner (verify against the scan schedule)
- Alert matches an existing allowlist entry or known-good pattern
- Single-host event with no indicators of lateral movement
Example: Wazuh rule 18152 fires for multiple failed logons on WIN-SERVER-01. You check the source — it is the svc-backup service account, the failures occurred during the scheduled maintenance window at 02:00, and the account successfully authenticated 30 seconds later. This is a password rotation during maintenance. Document your findings, classify as FP, close the ticket.
Your responsibility: Even at Level 0, you document what you checked and why you closed it. If you cannot explain your classification decision in two sentences, you did not investigate enough.
Level 1: Standard Escalation
The alert is real or you cannot determine whether it is real after reasonable investigation. It requires skills, access, or tools beyond your current capability.
Criteria:
- Confirmed true positive but containment requires Tier 2 access (firewall changes, endpoint isolation)
- Suspicious activity you cannot classify after 10 minutes of investigation
- Alert involves a Tier 2 asset (business-critical server, email infrastructure)
- Potential policy violation requiring HR or legal coordination
- Unusual pattern that does not match any known FP category from Lesson 4.1
Response time expectation: Acknowledge within 15 minutes, begin investigation within 1 hour.
Example: A Wazuh FIM alert shows /etc/shadow was modified on linux-web-01 outside of any maintenance window. No authorized change request exists. You check the user — it is www-data, which should never modify password files. This is a likely post-exploitation action. You cannot isolate the host (Tier 1 lacks firewall access). Standard escalation to Tier 2 with your investigation notes.
Level 2: Priority Escalation
Active threat confirmed. Impact is significant. Multiple indicators suggest a coordinated attack, but containment is still possible if response begins within minutes.
Criteria:
- Multiple correlated alerts across two or more hosts (lateral movement)
- Confirmed malware execution on a business-critical system
- Evidence of credential theft or pass-the-hash activity
- Active C2 communication detected (Suricata + Wazuh correlation)
- Insider threat indicators (data staging, bulk downloads from sensitive shares)
Response time expectation: Immediate notification (phone call, not just a ticket). Tier 2/3 begins response within 15 minutes.
Example: Suricata alerts show C2 beaconing from 10.0.5.22 to an external IP at 60-second intervals. Wazuh shows a new service registered on the same host 20 minutes earlier, and the parent process was powershell.exe executing a Base64-encoded command. A second host (WIN-SERVER-01) shows an RDP logon from 10.0.5.22 using an admin account that does not belong to that machine. This is active lateral movement with C2 — priority escalation. Pick up the phone.
Level 3: Emergency Escalation
Organizational-level impact. Active data destruction, confirmed ransomware, or compromise of identity infrastructure. Response requires leadership-level decisions (isolate network segments, notify legal, activate incident response retainer).
Criteria:
- Active ransomware execution (file encryption in progress)
- Confirmed data exfiltration to external destination
- Domain controller or certificate authority compromise
- Confirmed APT indicators (tooling, TTPs matching known threat actor profiles)
- Any event requiring external notification (law enforcement, regulators, customers)
Response time expectation: All-hands. Incident commander activated. CISO notified. Response within minutes, not hours.
Example: Wazuh FIM alerts are firing in rapid succession across 14 hosts — file extensions changing to .encrypted. Windows Event ID 7045 shows a new service named svc-update on the domain controller, and the binary path points to C:\Windows\Temp\locker.exe. Suricata shows outbound connections to three different TOR exit nodes from the domain controller. This is active ransomware with domain-level access. Emergency escalation — this is not a ticket, this is a phone call to the incident commander and CISO simultaneously.
Under-escalation kills. An analyst who sees ransomware indicators and files a standard ticket instead of making a phone call may cost the organization hours of additional encryption time. In incident response, minutes matter. A P1 event treated as P3 is an organizational failure, not an individual mistake — but the individual decision is where the failure starts.
The Anatomy of a Good Escalation
A bad escalation creates more work for Tier 2 than the original alert. A good escalation gives them everything they need to begin response immediately. The difference is structure.
Every escalation brief must contain five elements:
1. Alert Details
What fired, when, and on what. Include the rule ID, severity, source/destination IPs, hostname, and the exact timestamp.
Alert: Wazuh Rule 5712 — SSH brute force
Severity: High (Level 10)
Timestamp: 2026-02-23 02:47:18 UTC
Host: linux-web-01 (10.0.2.15)
Source IP: 185.220.101.42
Agent: linux-web-01 (Agent ID 003)
2. What You Investigated
List every pivot and data source you checked. This prevents Tier 2 from duplicating your work.
Investigation Steps:
- Checked source IP 185.220.101.42 in VirusTotal → flagged by 8/92 vendors as malicious (TOR exit node)
- Searched Wazuh for all events from 185.220.101.42 in last 24h → 847 failed SSH attempts in 5 min window
- Checked for successful logon (Rule 5715) from same IP → FOUND: successful auth at 02:50:03 UTC
- Checked for post-auth activity → sudo commands from www-data user at 02:51:xx
- Checked FIM alerts on linux-web-01 → /etc/shadow modified at 02:52:14 UTC
3. What You Found
State your findings clearly. Do not hedge or bury the lead. Start with the conclusion.
Findings:
Confirmed compromise of linux-web-01 via SSH brute force.
Attacker achieved root access and modified /etc/shadow.
Post-exploitation activity ongoing — lateral movement to dns-server-01 detected (RDP logon at 02:55).
4. Current Impact
What is affected right now? How many hosts, what data is at risk, is the attacker still active?
Impact:
- 2 hosts confirmed compromised: linux-web-01, dns-server-01
- Root access on linux-web-01 (password file modified)
- Active lateral movement in progress
- Web application data at risk (linux-web-01 serves customer portal)
- Attacker source is TOR — attribution unlikely
5. Recommended Next Step
Tell Tier 2 what you think should happen. You might be wrong — that is fine. The recommendation shows you have thought about impact and context, not just forwarded an alert.
Recommendation:
- Immediately isolate linux-web-01 and dns-server-01 from the network
- Reset credentials for all accounts that authenticated to either host in last 24h
- Check for additional lateral movement targets (firewall logs for outbound from 10.0.2.15)
- Classify as Priority (P2) — active compromise with lateral movement, no data exfil confirmed yet
The 30-second test: If Tier 2 can read your escalation brief and begin response within 30 seconds without asking a single follow-up question, you wrote a good escalation. If they have to call you back for basic details, you wasted time for both of you.
Bad Escalations vs Good Escalations
The difference between a useful escalation and a useless one is always the same: specificity. Here are two escalations for the exact same incident. One is useless. One is actionable.
Bad Escalation
Subject: Suspicious activity on linux-web-01
Hey, something looks weird on linux-web-01. Wazuh is showing some alerts. Can someone take a look? Might be nothing but figured I'd flag it.
This escalation tells Tier 2 nothing. What alerts? What did you check? What is the severity? Is this happening now or did it happen last week? Tier 2 now has to start from scratch — find the alerts, investigate from zero, and context-switch from whatever they were doing. You have cost the SOC 15-30 minutes of duplicated investigation.
Good Escalation
Subject: [P2 — PRIORITY] Confirmed SSH brute-force compromise — linux-web-01 — lateral movement to dns-server-01
Alert: Wazuh Rule 5712 (SSH brute force) at 02:47 UTC on linux-web-01 from 185.220.101.42 (TOR exit node, 8/92 VT detections)
Investigated: 847 failed SSH attempts → successful auth at 02:50 → sudo by www-data → /etc/shadow modified → RDP to dns-server-01 at 02:55
Finding: Confirmed root compromise of linux-web-01 with active lateral movement. Attacker has domain presence on 2 hosts.
Impact: 2 hosts compromised, customer portal at risk, attacker active NOW
Recommendation: Isolate both hosts, reset all creds for hosts in same subnet, check for additional lateral targets
This escalation can be acted on immediately. Tier 2 knows the severity, the scope, the timeline, what was already checked, and what needs to happen next.
Ticket subject lines matter. Most ticketing systems show only the subject in queue views. A subject that says "Suspicious activity" is invisible in a queue of 50 tickets. A subject that says "[P2 — PRIORITY] Confirmed SSH brute-force compromise — linux-web-01" gets immediate attention. Always include the priority level, the classification, and the affected asset in the subject line.
SLA Awareness: The Clock Is Always Running
Every SOC operates under Service Level Agreements (SLAs) that define how quickly you must acknowledge, investigate, and respond to alerts at each severity level. Missing SLAs is not just a performance metric — it has contractual and regulatory consequences.
The Four Priority Levels
| Priority | Name | Acknowledge | Investigate | Resolve | Example |
|---|---|---|---|---|---|
| P1 | Critical | 5 minutes | Immediate | 4 hours | Active ransomware, domain controller compromise, confirmed data exfiltration |
| P2 | High | 15 minutes | 30 minutes | 8 hours | Confirmed compromise with lateral movement, malware on business-critical system |
| P3 | Medium | 1 hour | 4 hours | 24 hours | Isolated TP on standard asset, policy violation, suspicious but contained activity |
| P4 | Low | 4 hours | 24 hours | 72 hours | Confirmed FP requiring rule tuning, informational findings, vulnerability notifications |
What Happens When You Miss SLAs
- P1 SLA miss: The incident continues uncontained. Every minute of delay during active ransomware or exfiltration directly increases damage. A 30-minute delay in isolating a ransomware-infected host can mean the difference between 1 host encrypted and 200.
- P2 SLA miss: The attacker has more time to establish persistence, stage data, and move laterally. A compromise that could have been contained to one host expands to a subnet.
- P3/P4 SLA miss: Regulatory risk. Many frameworks (SOC 2, HIPAA, PCI-DSS) require documented response times. Missed SLAs appear in audit findings and can result in compliance failures.
- Customer SLA miss: For MSSPs, SLA breaches trigger contractual penalties and erode trust. Repeated misses lead to lost clients.
SLAs are not arbitrary. They are derived from attacker dwell time research. The MITRE ATT&CK framework documents that most attackers achieve their objectives within hours of initial access. SLA timelines are calibrated to interrupt the attack chain before the attacker reaches their goal — typically data exfiltration or impact. When you miss a P1 SLA, you are giving the attacker exactly what they need: time.
Who to Escalate To
Knowing that you need to escalate is half the problem. Knowing who to call is the other half. Every SOC should have an escalation matrix posted where every analyst can see it. If yours does not, build one on your first day.
| Condition | Escalate To | Channel | When |
|---|---|---|---|
| Cannot classify after 10 minutes | Tier 2 Analyst | Ticket + chat | During shift |
| Confirmed TP requiring containment | Tier 2 Analyst on call | Ticket + phone | Anytime |
| Multi-host attack, lateral movement | Tier 3 / Hunt Team | Phone + ticket | Immediately |
| Active exfiltration or ransomware | Incident Commander | Phone (direct) | Immediately |
| Domain controller / CA compromise | Incident Commander + CISO | Phone (direct) | Immediately |
| External notification required | CISO + Legal | Via incident commander | After IC confirmation |
Communication Channels by Urgency
- Ticket only: P4 and some P3 events. Asynchronous is fine — no one is losing sleep over a rule tuning request.
- Ticket + chat (Slack/Teams): P3 events during business hours. The ticket creates the record; the chat gets attention.
- Ticket + phone call: P2 events. The phone call ensures the right person sees it immediately. Do not rely on chat notifications for priority events — people mute channels.
- Phone call (direct) + ticket after: P1 events. Call first, document after. If the incident commander does not answer, call the backup. If the backup does not answer, call the CISO. P1 events do not wait for callbacks.
Never escalate a P1 by ticket alone. A ticket is a record, not a notification. If your P1 escalation was a ticket submitted at 02:47 AM and Tier 2 did not see it until 07:00 AM, the SLA was missed by 4 hours and the attacker had 4 uninterrupted hours in your environment. Phone calls are the only reliable escalation channel for emergencies.
The 10-Minute Rule
Here is a practical heuristic that prevents both under-escalation and analysis paralysis:
If you cannot confidently classify an alert within 10 minutes of focused investigation, escalate with your notes.
This rule works because:
- Most FPs are obvious within 2-3 minutes. Scheduled task, known scanner, maintenance window — the evidence is clear and the pattern is familiar. If you are still uncertain after 10 minutes, the alert is non-trivial.
- 10 minutes of notes is valuable. Even if you could not classify the alert, you checked sources, ran queries, and eliminated possibilities. Those 10 minutes of work are not wasted — they are the foundation of Tier 2's investigation. Without your notes, Tier 2 starts from zero.
- The cost of late escalation exceeds the cost of unnecessary escalation. A Tier 2 analyst who receives a well-documented escalation that turns out to be a FP loses 5 minutes reviewing your notes and closing the ticket. An attacker who gets an extra 30 minutes because you hesitated to escalate can establish persistence, steal credentials, and move laterally. The math is clear.
10-Minute Escalation Checklist:
□ Checked source/dest IPs in threat intel (VT, AbuseIPDB)
□ Searched for related events (same src IP, same host, same user, ±1 hour)
□ Checked asset value in CMDB or asset inventory
□ Verified whether a maintenance window or change request covers this activity
□ Checked user role and baseline behavior for the account involved
□ Looked for indicators of lateral movement to other hosts
→ Still uncertain? Escalate now with these notes attached.
The 10-minute rule also protects you. If an incident is later reviewed and the question comes up "why did the L1 analyst not escalate sooner?" — your answer is: "I investigated for 10 minutes, documented my findings, and escalated when I could not classify. Here are my notes." That is a defensible decision. "I kept investigating for 45 minutes because I thought I could figure it out" is not.
The Shift Handoff
A 24/7 SOC runs in shifts. When your shift ends, your open investigations do not pause — the attacker does not stop because your shift is over. The shift handoff is the most underrated escalation skill because it is not escalating to a senior analyst — it is escalating to yourself in the future, represented by the next analyst on shift.
What a Handoff Must Include
| Section | Content | Example |
|---|---|---|
| Open Tickets | Ticket IDs, current status, what is pending | "TICK-4421: SSH brute force on linux-web-01, awaiting Tier 2 review" |
| Active Investigations | What you were looking at, what you found, what is left | "Investigating 185.220.101.42 — confirmed TOR exit, 847 SSH failures, checking for post-auth" |
| Pending Actions | Tasks that need follow-up on the next shift | "Need to verify if svc-backup password rotation completed (check with IT ops at 09:00)" |
| Environment Notes | Anything unusual about the current environment | "Nessus scan running until 04:00 — expect rule 18100 FPs on subnet 10.0.5.0/24" |
| Queue Status | Unacknowledged alert count, oldest unacked alert | "Queue: 12 unacked, oldest is 45 min (P4 info-level)" |
Handoff Anti-Patterns
"Everything is fine" — The handoff that says nothing. If you worked an 8-hour shift and have nothing to hand off, you were either in the quietest SOC in history or you are not documenting your work.
"Check the tickets" — Forcing the incoming analyst to read every ticket you touched during your shift instead of summarizing the critical items. Your job is to highlight what matters, not create a reading assignment.
"I was looking at something weird but I forgot what" — The result of not taking notes during investigation. Every pivot, every query, every finding goes in the ticket in real-time. Not at the end of the shift. Not from memory.
In Lab 4.4 (Alert Queue Challenge), you will practice a realistic shift handoff. After triaging 50 alerts under time pressure, you will produce a structured handoff document that includes open tickets, active investigations, and recommended priority for the incoming shift. The handoff deliverable is graded — because in a real SOC, a bad handoff is an operational failure.
Three Escalation Walkthroughs
Let's apply the framework to three realistic scenarios using Wazuh alerts. Each scenario requires a different escalation decision.
Scenario 1: Handle Yourself (Level 0)
Alert: Wazuh Rule 18152 — Multiple Windows logon failures
Host: WIN-SERVER-01 (Agent ID 002)
Source: 10.0.5.50 (internal workstation)
Time: Monday 09:15 UTC
Details: 12 failed logon attempts for user jsmith over 3 minutes, followed by a successful logon at 09:18.
Your investigation:
- Check user
jsmith— finance department, standard user, assigned to workstation 10.0.5.50 - Check the time — Monday morning, 09:15 AM, start of business
- Check for related events — no other suspicious activity from
jsmithor 10.0.5.50 in last 24 hours - Check logon type — Type 10 (RDP).
jsmithnormally uses RDP to access WIN-SERVER-01 for the finance application - Pattern recognition — Monday morning, user forgot password over the weekend, multiple attempts before success
Decision: Handle yourself. This is a textbook FP — a user struggling with a Monday-morning password. Document findings: "12 failed RDP logons for jsmith from assigned workstation. Successful auth at 09:18. Normal Monday-morning pattern, no lateral indicators. Classified FP. Closed."
Scenario 2: Standard Escalation (Level 1)
Alert: Wazuh Rule 550 — File integrity monitoring — checksum changed
Host: linux-web-01 (Agent ID 003)
File: /usr/lib/systemd/system/sshd-helper.service
Time: Wednesday 14:22 UTC
Details: New file detected. No matching change request.
Your investigation:
- Check the file path —
sshd-helper.serviceis not a standard systemd unit. The legitimate SSH service file issshd.service. The name mimics a legitimate service. - Check the change management calendar — no authorized changes to linux-web-01 today
- Check who created the file — Wazuh audit log shows the file was created by user
www-data. A web service user should never create systemd service files. - Check for related events — Wazuh shows
www-dataexecutedsudo cpat 14:21 to place the file. Rule 5402 (sudo command) was also triggered. - Check asset tier — linux-web-01 is Tier 2 (customer-facing web application)
Decision: Standard escalation. This is a confirmed TP — a suspicious systemd service installed by a web user outside of change management. You cannot isolate the host (Tier 1 access limitation). Escalate to Tier 2 with full investigation notes.
Your escalation:
[P3 — CONFIRMED TP] Suspicious systemd service on linux-web-01
Alert: Wazuh Rule 550 (FIM) — new file /usr/lib/systemd/system/sshd-helper.service
Time: 2026-02-19 14:22 UTC | Host: linux-web-01 (Agent 003)
Investigated:
- sshd-helper.service is not a standard systemd unit (legitimate is sshd.service)
- Created by www-data via sudo cp at 14:21 (Rule 5402)
- No change request for linux-web-01 today
- linux-web-01 is Tier 2 (customer-facing)
Finding: Likely persistence mechanism — web service account created a mimicked SSH service file via sudo.
Impact: Single host, Tier 2 asset, no lateral movement detected yet.
Recommendation: Inspect sshd-helper.service binary, check for additional persistence, consider isolating linux-web-01 pending review.
Scenario 3: Emergency Escalation (Level 3)
Alert: Multiple simultaneous alerts across agents Host(s): WIN-SERVER-01 (002), linux-web-01 (003), dns-server-01 (004), fw-edge-01 (005) Time: Saturday 03:10 UTC
What you see in the queue (first 60 seconds):
- Rule 554 (FIM): File renamed with
.lockedextension — WIN-SERVER-01 — 03:10 - Rule 554 (FIM): File renamed with
.lockedextension — WIN-SERVER-01 — 03:10 (2nd file) - Rule 554 (FIM): File renamed with
.lockedextension — WIN-SERVER-01 — 03:10 (3rd file) - Rule 7045 (New service):
update-svcinstalled on WIN-SERVER-01 — 03:08 - Rule 5715 (SSH success after brute): linux-web-01 from 10.0.5.22 — 03:05
- Rule 550 (FIM):
/var/www/html/index.htmlmodified — linux-web-01 — 03:09 - Rule 5402 (sudo):
www-dataranchmod 777 /tmp/crypton linux-web-01 — 03:09 - Alerts still arriving every few seconds...
Decision: Emergency escalation (Level 3). Do not investigate further — the pattern is unmistakable. Files being renamed to .locked across a server, a new service installed minutes before, and evidence of cross-host compromise means active ransomware with lateral movement. Call the incident commander immediately.
You will have time to document afterward. Right now, every second of investigation is a second of continued encryption. The correct action sequence is:
- Call the incident commander — 30 seconds
- Call the Tier 2/3 on-call — 30 seconds
- Open a P1 ticket with what you see (bullet points, not a novel) — 60 seconds
- If you have the access, isolate affected hosts immediately — do not wait for authorization on a P1
Time is the enemy during emergency escalation. Every minute you spend perfecting your escalation notes is a minute the ransomware is encrypting more files. For P1 events, the phone call comes FIRST. The documentation comes AFTER you have alerted the response team. This is the one exception to "document everything before escalating."
Common Escalation Mistakes
1. Under-Escalation: Fear of Being Wrong
New analysts often hesitate to escalate because they fear looking incompetent. "What if it is a false positive and I wasted Tier 2's time?" This fear is understandable and completely wrong.
The reality: Every Tier 2 and Tier 3 analyst was once a Tier 1 who escalated things that turned out to be false positives. It is expected. It is how you learn. A well-documented escalation that turns out to be a FP costs Tier 2 five minutes. A missed true positive that you sat on for 45 minutes costs the organization potentially millions.
The fix: Use the 10-minute rule. Escalate with notes. No one will criticize you for escalating a well-investigated alert that turns out to be benign. People will criticize you for not escalating a real attack because you were afraid of being wrong.
2. Over-Escalation: Passing the Buck
The opposite failure. Some analysts escalate everything because it is easier than investigating. Every alert becomes "Hey, can someone look at this?" with no context, no investigation, and no analysis.
The impact: Tier 2 becomes Tier 1. The escalation queue backs up. Response times degrade across the board. Eventually, Tier 2 starts ignoring escalations from the analyst who cries wolf — and that is when the real attack gets missed.
The fix: Before escalating, ask yourself: "Did I check the five elements? Can I explain what I found?" If not, you have not investigated yet — you have just forwarded an email.
3. Escalating Without Investigation Notes
The most common mistake across all experience levels. The analyst investigates, draws a conclusion, but only escalates the conclusion without the evidence trail.
Bad: "I think linux-web-01 is compromised. Can Tier 2 investigate?"
Good: "linux-web-01 shows FIM alerts for new systemd service (not in change management), created by www-data via sudo, service name mimics sshd. I checked for lateral movement — none detected. Recommend host inspection and isolation."
The fix: Document as you investigate, not after. Every query, every result, every dead end goes into the ticket in real-time. When you escalate, the notes are already there.
Connecting It All: Module 4 in Review
This lesson completes Module 4 — Alert Triage. Over five lessons, you have built the complete L1 analyst triage workflow:
- Lesson 4.1: Classify alerts as TP or FP using pattern recognition and the confusion matrix
- Lesson 4.2: Add context — asset value, user role, temporal patterns, geography, behavior — to make classification decisions
- Lesson 4.3: Follow a structured investigation workflow: alert → pivot → correlate → decide
- Lesson 4.4: Decode obfuscated payloads with CyberChef to understand what an alert is actually doing
- Lesson 4.5: Escalate correctly — know when to handle, when to pass, and how to hand off
In Lab 4.4, you will put all of these skills together. You will face a queue of 50 alerts of varying severity, investigate them under time pressure, make escalation decisions, and produce a shift handoff document. The lab is graded on accuracy (did you correctly classify TP vs FP?), efficiency (did you meet SLA targets?), and handoff quality (could the next analyst continue your work without starting over?).
Key Takeaways
- Escalation exists on a spectrum: handle yourself → standard → priority → emergency. Each level has specific criteria and response timelines.
- A good escalation brief has five elements: alert details, what you investigated, what you found, current impact, and recommended next step. If Tier 2 needs to call you back for details, your brief failed.
- The 10-minute rule prevents both under-escalation and analysis paralysis: if you cannot classify an alert after 10 minutes of focused work, escalate with your investigation notes.
- SLAs are not suggestions. P1 events require phone calls and immediate response. P4 events can wait for the next business day. Misclassifying severity wastes resources or lets attackers operate unchecked.
- Never escalate a P1 by ticket alone. Phone calls are the only reliable notification channel for emergencies.
- Shift handoffs are escalations to the next analyst. Open tickets, active investigations, pending actions, and queue status — if you do not hand these off, investigative continuity breaks.
- Document as you investigate, not after. Real-time notes in the ticket are the foundation of every escalation, every handoff, and every post-incident review.
What's Next
Module 4 gave you the triage framework. You can classify alerts, add context, investigate, decode payloads, and escalate correctly. But triage is reactive — you respond to what your SIEM shows you.
In Module 5: Threat Intelligence, you shift from reactive to proactive. You will learn how threat feeds, IOCs, and intelligence-sharing platforms like MISP transform your triage decisions. Instead of asking "is this alert suspicious?" you will ask "does this match a known threat actor campaign?" — and the answer will change everything about how you prioritize your queue.
Knowledge Check: Escalation — When and How
10 questions · 70% to pass
A Wazuh FIM alert fires showing /etc/shadow was modified on linux-web-01 at 14:22 UTC by the www-data user. No change request exists for today. What is the correct escalation level?
Which of the following is NOT one of the five required elements of a structured escalation brief?
An analyst sees Wazuh alerts showing files being renamed with a .locked extension across WIN-SERVER-01, with a new service installed two minutes earlier. What should they do FIRST?
In Lab 4.4 (Alert Queue Challenge), students produce a shift handoff document after triaging 50 alerts. Which of the following would be an anti-pattern in that handoff?
What is the primary purpose of the 10-minute rule in alert triage?
A P1 incident occurs at 02:47 AM. The analyst submits a detailed ticket but does not make a phone call. Tier 2 sees the ticket at 07:00 AM. What is the consequence?
An analyst investigates an alert for 10 minutes and escalates with notes. Tier 2 reviews and determines it was a false positive. What is the correct assessment of the analyst's action?
In the CyberBlue lab environment, which Wazuh agent would you expect to generate Windows Event ID 7045 (new service installed) alerts?
A SOC operates with P1-P4 SLAs. An analyst classifies a confirmed compromise with active lateral movement across two hosts as P3 (Medium) instead of P2 (High). What is the primary risk?
During Lab 4.4, you triage 50 alerts and must prioritize them. Which combination of factors should determine whether an alert is handled as P2 (High) versus P3 (Medium)?
0/10 answered