Classify evidence types: best evidence, corroborative evidence, and indirect (circumstantial) evidence
Interpret OS logs (Windows Event IDs, Linux syslog/auditd), SIEM correlation, and SOAR playbooks to identify security events
Analyze output from malware analysis tools including sandboxes and detonation chambers
Recognize IOC indicators from file hashes, URLs, network artifacts, and system artifacts; understand STIX/TAXII distribution
Section 1: Evidence Types and Classification
Digital evidence is any binary data — stored on or transmitted by a computing device — that can be used in an investigation. Its sources span hard drive images, packet captures, audit logs, deleted file fragments, and memory dumps. Evidence is not equally useful: courts apply a classification hierarchy that determines how much weight each piece carries and what documentation is required.
The Three Core Evidence Classes
Best Evidence
Best evidence is the original, unaltered record. In digital forensics, best evidence includes:
An unmodified disk image captured with a write-blocker
Original log files with intact timestamps and SHA-256-verified checksums
Unedited CCTV or body-camera footage
A forensic memory dump taken directly from a live system
A SHA-256 hash taken at acquisition time and re-verified at every subsequent step is the mechanism by which analysts prove a file has not changed.
Corroborative Evidence
Corroborative evidence supports and reinforces primary evidence. It does not prove the main fact on its own, but it makes the primary evidence more credible. Examples:
Browser history that corroborates the timeline established by server-side access logs
Firewall logs that confirm traffic was blocked at the same time an alert fired in the SIEM
IP geolocation records that support claims about connection origin
Indirect (Circumstantial) Evidence
Indirect evidence does not directly prove a fact — instead, it allows a reasonable inference about what happened. Examples:
Deleted files recovered from unallocated disk space (imply files once existed)
Prefetch records suggesting a program was executed even if the executable was removed
Residual registry keys suggesting a program was installed then uninstalled
Outbound DNS queries to an unusual domain suggesting C2 communication
Evidence Type
Definition
Digital Example
Strength
Best Evidence
Original, unaltered record
SHA-256-verified disk image
Highest — direct proof
Corroborative
Supports and reinforces primary evidence
Firewall log matching SIEM alert
Medium — strengthens primary
Indirect/Circumstantial
Allows inference about a fact
Deleted temp files suggesting execution
Variable — builds cumulatively
Chain of Custody
Chain of custody is the documented, unbroken record of who collected evidence and when, how it was transported and stored, who accessed it and for what purpose, and what transformations were performed. A practical form records the SHA-256 hash of every artifact at the moment of collection. Before analysis, the hash is re-verified. If hashes do not match, the evidence may be inadmissible.
Rules of Evidence
For digital evidence to be accepted in legal proceedings, it must satisfy four properties:
Authentic — it is what it claims to be (proven via hash verification and documentation)
Complete — it tells the whole story, not a cherry-picked fragment
Reliable — collected using sound, repeatable methods
Believable — can be explained to a judge or jury in plain terms
FIGURE 7.1 — Evidence Classification and Chain of Custody Flow
Key Points — Evidence Types and Classification
Best evidence is the original artifact with a hash-verified chain of custody — the gold standard in investigations and legal proceedings.
Corroborative evidence strengthens the primary case by independently confirming the same facts from a different source.
Indirect evidence alone is weak, but ten pieces pointing in the same direction create a compelling circumstantial case.
A broken chain of custody can render even perfect evidence inadmissible — document every collection, transfer, and access event.
SHA-256 is the required hash algorithm for forensic evidence verification; re-hash before every analysis step.
1. An analyst captures a disk image with a write-blocker and records its SHA-256 hash. What class of evidence is this?
2. Firewall logs that independently confirm the same blocked connection recorded in a SIEM alert are an example of what?
3. Prefetch records suggesting a malicious executable ran, even though the executable was deleted, are classified as:
Section 2: Log Analysis and Event Identification
Logs are the primary narrative of any security incident. Every operating system, application, network device, and security platform generates log records as it operates. The investigator's job is to correlate these records across sources and reconstruct a coherent timeline.
Windows Event Logs
Windows Event Logs are stored in .evtx format under C:\Windows\System32\winevt\Logs\. The most security-relevant event IDs:
Log Channel
Event ID
What It Records
Security
4624
Successful logon
Security
4625
Failed logon — watch for brute force patterns
Security
4648
Logon using explicit credentials (pass-the-hash indicator)
Security
4688
Process creation — records command execution
Security
4698 / 4702
Scheduled task created / modified (persistence)
Security
4720 / 4732
User account created / added to privileged group
System
7045
New service installed — common malware persistence method
Security
4104
PowerShell Script Block Logging — full script content
Pattern: Lateral Movement Detection
Event 4625 (failed logon) for administrator from 192.168.10.45 — 47 times in 90 seconds
Event 4624 (successful logon) from the same IP — logon type 3 (network logon)
Event 4688 (process creation) — cmd.exe spawned by services.exe
This sequence — brute force → network logon → command shell from a service process — is a classic indicator of credential stuffing and remote command execution via a compromised service account.
Linux / syslog
File
Contents
/var/log/auth.log
SSH logins, sudo usage, PAM authentication
/var/log/syslog
General system messages
/var/log/kern.log
Kernel messages — driver errors, network issues
/var/log/secure (RHEL/CentOS)
Authentication events
/var/log/audit/audit.log
Auditd events — file access, syscalls
A burst of "Failed password" entries from a single IP followed by "Accepted password" is the syslog signature of a successful brute-force attack.
SIEM: Aggregation, Correlation, and Alerting
A SIEM ingests log data from hundreds of sources, normalizes it into a common schema, and applies correlation rules. A single failed login is noise. Ten thousand failed logins across fifty accounts from one IP in five minutes is an alert. Common platforms: Splunk (SPL), Microsoft Sentinel (KQL), IBM QRadar, ELK Stack.
SIEM workflow:
Collection — agents push logs to the SIEM
Normalization — raw data parsed into structured fields
Indexing — records stored for search
Correlation — rules match patterns across multiple events
Alerting — correlation rule fires; alert routed to SOC queue
Investigation — analysts query raw logs surrounding the alert
index=windows EventCode=4656 Object_Name="*lsass*" Access_Mask="0x1410"
| stats count by ComputerName, Account_Name, Process_Name
| where count > 1
This Splunk SPL query detects attempts to open a handle to lsass.exe with read permission — the hallmark of credential dumping tools like Mimikatz.
SOAR: Automated Response Playbooks
SOAR platforms take SIEM alerts and execute automated response workflows (playbooks). A typical phishing SOAR playbook:
Receive alert: "User clicked suspicious URL"
Automatically query VirusTotal for URL hash score
Pull email headers, extract sender, reply-to, originating IP
Check originating IP against threat intelligence feeds
If malicious: quarantine inbox, block URL at proxy, create ITSM ticket
Notify SOC analyst with pre-populated summary for human review
SOAR reduces mean-time-to-respond (MTTR) from hours to minutes. Platforms include Palo Alto XSOAR, Splunk SOAR, and IBM Security SOAR.
is a strong indicator of obfuscation — decoding the Base64 payload in a sandbox often reveals a reverse shell or dropper.
FIGURE 7.2 — SIEM Collection and SOAR Response Pipeline
Key Points — Log Analysis and Event Identification
Windows Event ID 4624/4625 (successful/failed logon) and 4688 (process creation) are the three most critical security log events for detecting intrusions.
SIEM correlation transforms raw log noise into actionable alerts by matching patterns across multiple log sources simultaneously.
SOAR playbooks automate high-volume, well-understood responses — reducing MTTR from hours to minutes while freeing analysts for complex investigations.
PowerShell Event ID 4104 (Script Block Logging) is the highest-value single log source for detecting PowerShell-based attacks; Base64-encoded commands indicate obfuscation.
Web server logs with cmd= parameters in URLs, non-browser user-agents, and HTTP 200 responses to admin paths are strong command injection indicators.
sequenceDiagram
participant SIEM as SIEM
participant SOAR as SOAR Playbook
participant VT as VirusTotal API
participant TI as Threat Intel Feed
participant FW as Proxy / Firewall
participant ITSM as ITSM (Ticket)
participant SOC as SOC Analyst
SIEM->>SOAR: Alert — "User clicked suspicious URL"
SOAR->>VT: Query URL hash score
VT-->>SOAR: Score: Malicious (87/90 engines)
SOAR->>TI: Check originating IP reputation
TI-->>SOAR: IP listed — known C2 infrastructure
SOAR->>FW: Block URL at proxy
SOAR->>ITSM: Create incident ticket (pre-populated)
SOAR->>SOC: Notify — summary report for human review
SOC->>SOC: Validate, escalate, or close
Post-Check — Section 2: Log Analysis
4. Which Windows Event ID records a new service being installed — a common malware persistence method?
5. An analyst sees 47 Event ID 4625 entries from 192.168.10.45 in 90 seconds, followed immediately by a 4624 (network logon) and then a 4688 with cmd.exe spawned by services.exe. What does this pattern indicate?
6. What is the primary purpose of a SOAR playbook compared to a SIEM?
7. A PowerShell command line contains -EncodedCommand JABjAGwAaQBlAG4AdA.... What does this indicate?
Section 3: Malware Analysis Techniques
Analyzing malware presents a fundamental problem: to understand what it does, you must run it — but running it risks infecting your analysis environment. The solution combines two complementary approaches: static analysis (examining code without executing it) and dynamic analysis (executing it in a controlled, isolated environment).
Static Analysis
Static analysis examines a malware sample without executing it. Key techniques:
For Windows PE format executables, tools like PEStudio examine:
Import Address Table (IAT) — API functions the binary calls. CreateRemoteThread, VirtualAllocEx, WriteProcessMemory indicate process injection.
Section entropy — packed/encrypted sections have high entropy (close to 8.0). Normal code runs 5–6.
Digital signature — is the file signed? Is the signature valid or revoked?
Compilation timestamp — can provide context (though easily falsified)
Disassembly and Decompilation
Ghidra (NIST-released, free) and IDA Pro disassemble binary code into assembly language or pseudo-C, allowing analysts to trace execution logic and identify anti-analysis techniques.
Dynamic Analysis: Sandboxes and Detonation Chambers
When a sample is heavily packed, obfuscated, or uses runtime decryption, dynamic analysis in a sandbox or detonation chamber is required.
Feature
Sandbox
Detonation Chamber
Primary purpose
Automated behavioral analysis
Deep investigation, often manual
Implementation
Cloud-based VM, automated
Full emulation or dedicated hardware
Speed
Minutes per sample
Minutes to hours
Output
Automated report + IOC extraction
Detailed forensic artifacts
Examples
Cuckoo, Any.run, Hybrid Analysis
FireEye Malware Analysis, Falcon X
The Sandbox Detonation Workflow
File submission — analyst uploads sample via UI or API
Pre-filter — quick signature check; known malware flagged immediately
Detonation — sample executes in isolated VM (Windows, Linux, or Android)
Event logging — sandbox monitors all system activity during execution
Report generation — analysis returned in human-readable format, typically within minutes
What the Sandbox Monitors
Artifact Category
Specific Observations
Network communications
DNS queries, HTTP/HTTPS requests, C2 beacon patterns
File system changes
Files created, modified, or deleted; dropped payloads
Registry modifications
Persistence keys added (Run, RunOnce, Services)
Process activity
Child processes spawned, process injection, code hollowing
Static analysis (hash, strings, PE header) is fast and safe; dynamic analysis in a sandbox reveals runtime behavior that static methods cannot — use both together.
High section entropy (close to 8.0) in a PE file indicates packing or encryption — a strong indicator that dynamic analysis is required.
Sandbox reports map every observed behavior to MITRE ATT&CK technique IDs, providing immediate operational context for SIEM rule creation.
Malware evasion techniques (sleep delays, VM detection, human interaction checks) are countered by sandboxes with clock acceleration, bare-metal environments, and simulated user activity.
Vssadmin.exe deleting shadow copies (delete shadows /all /quiet) is the definitive behavioral indicator of ransomware — T1490 Inhibit System Recovery.
Post-Check — Section 3: Malware Analysis
8. A PE file's .text section shows an entropy value of 7.9. What does this indicate?
9. During sandbox detonation, a sample executes vssadmin.exe delete shadows /all /quiet. Which MITRE ATT&CK technique does this map to?
10. What does the presence of CreateRemoteThread, VirtualAllocEx, and WriteProcessMemory in a PE file's Import Address Table strongly indicate?
11. A malware sample detects a mouse cursor that hasn't moved in 3 minutes and refuses to execute its payload. What evasion technique is this?
Section 4: IOC Recognition and Threat Intelligence
An Indicator of Compromise (IOC) is a forensic artifact that, when observed in a system or network, indicates with high confidence that a security breach has occurred or is in progress. IOCs are the actionable outputs of malware analysis and incident investigation — they answer: "What observable evidence can we use to detect this threat?"
File-Level IOCs: Hash Values
The most precise IOC for a specific file is its cryptographic hash. SHA-256 is the current standard:
Algorithm
Output Length
Current Status
Use Case
MD5
128-bit (32 hex chars)
Deprecated — collision-prone
Legacy systems only
SHA-1
160-bit (40 hex chars)
Deprecated
Legacy compatibility
SHA-256
256-bit (64 hex chars)
Preferred standard
All modern forensic and IOC use
SHA-512
512-bit (128 hex chars)
High security
Sensitive data integrity
ssdeep
Variable (fuzzy hash)
Active use
Similarity matching between variants
ssdeep (fuzzy hashing) identifies malware variants that share significant code regions — useful for tracking malware families even after recompilation.
Network-Level IOCs
IP Addresses
IP IOCs have a short shelf life — attackers frequently rotate infrastructure. Validation rules:
Dropped executables in C:\Users\Public\, %TEMP%, or %APPDATA%\
Modified legitimate system binaries (DLL sideloading)
Processes with no associated executable on disk (process hollowing / fileless malware)
Unexpected parent process chains (explorer.exe → cmd.exe → powershell.exe)
STIX and TAXII: Sharing Intelligence
STIX (Structured Threat Information eXpression) is a JSON-based language for describing cyber threat intelligence objects:
STIX Object Type
Represents
indicator
An IOC with detection pattern (hashes, IPs, domains)
malware
A malware family — behaviors, capabilities
threat-actor
An APT group or threat actor
campaign
A coordinated series of attacks
attack-pattern
A MITRE ATT&CK technique
relationship
Links between objects (malware used by actor)
TAXII (Trusted Automated eXchange of Indicator Information) defines how STIX bundles are distributed. TAXII 2.1 uses REST API endpoints with Collections (named repositories of STIX objects) and Channels (pub/sub for real-time distribution). TAXII consumers (SIEMs, EDR platforms) automatically pull new indicators and create detection rules.
The Pyramid of Pain
David Bianco's Pyramid of Pain ranks IOC types by how difficult they are for attackers to change once defenders start detecting them:
FIGURE 7.4 — IOC Sources, STIX Packaging, and TAXII Distribution
Key Points — IOC Recognition and Threat Intelligence
SHA-256 is the required standard for IOC file hashes — MD5 and SHA-1 are deprecated due to collision vulnerabilities.
Network IOCs are ranked by stability: domains outlast IPs (days vs. minutes to change), and TTPs outlast all artifact-level IOCs.
Windows persistence registry keys under HKCU\...\Run and HKLM\...\Run are primary system artifact IOCs — any unexpected entry pointing to temp directories is a red flag.
STIX 2.1 is the JSON format for describing threat intelligence objects; TAXII 2.1 is the REST API transport that distributes them to consumers.
The Pyramid of Pain: detecting TTPs (top) forces attackers to completely retrain — far more valuable than detecting hashes (bottom) which attackers can evade in seconds by recompiling.
Post-Check — Section 4: IOC Recognition and Threat Intelligence
12. An analyst identifies a malicious file and wants to share its IOC with partner organizations. Which hash algorithm should they use?
13. According to the Pyramid of Pain, which IOC type is hardest for an attacker to change once defenders start detecting it?
14. What is the relationship between STIX and TAXII?
15. A SIEM alert fires on an unexpected registry value at HKCU\Software\Microsoft\Windows\CurrentVersion\Run\svchost32 pointing to %APPDATA%\temp\update.exe. This is an example of which type of IOC?