feature: 1.1.18
Some checks failed
Build RPM Package / Build RPM Packages (CentOS 7, Rocky 8/9/10) (push) Has been cancelled

+- FEATURE: Add comprehensive metrics for capture and TLS parser monitoring
+- Capture metrics: packets_received, packets_sent, packets_dropped (atomic counters)
+- Parser metrics: retransmit_count, gap_detected_count, buffer_exceeded_count, segment_exceeded_count
+- New GetStats() method on Capture interface for capture statistics
+- New GetMetrics() method on Parser interface for parser statistics
+- Add DefaultMaxHelloSegments constant (100) to prevent memory leaks from fragmented handshakes
+- Add Segments field to ConnectionFlow for per-flow segment tracking
+- Increase DefaultMaxTrackedFlows from 50000 to 100000 for high-traffic scenarios
+- Improve TCP reassembly: better handling of retransmissions and sequence gaps
+- Memory leak prevention: limit segments per flow and buffer size
+- Aggressive flow cleanup: clean up JA4_DONE flows when approaching flow limit
+- Lock ordering fix: release flow.mu before acquiring p.mu to avoid deadlocks
+- Exclude IPv6 link-local addresses (fe80::) from local IP detection
+- Improve error logging with detailed connection and TLS extension information
+- Add capture diagnostics logging (interface, link_type, local_ips, bpf_filter)
+- Fix false positive retransmission counter when SYN packet is missed
+- Fix gap handling: reset sequence tracking instead of dropping flow
+- Fix extractTLSExtensions: return error details with basic TLS info for debugging
This commit is contained in:
toto
2026-03-09 16:38:40 +01:00
parent d22b0634da
commit e166fdab2e
11 changed files with 448 additions and 100 deletions

View File

@ -32,9 +32,12 @@ const (
// Parser configuration constants
const (
// DefaultMaxTrackedFlows is the maximum number of concurrent flows to track
DefaultMaxTrackedFlows = 50000
// Increased from 50000 to 100000 to handle high-traffic scenarios
DefaultMaxTrackedFlows = 100000
// DefaultMaxHelloBufferBytes is the maximum buffer size for fragmented ClientHello
DefaultMaxHelloBufferBytes = 256 * 1024 // 256 KiB
// DefaultMaxHelloSegments is the maximum number of segments to accumulate per flow
DefaultMaxHelloSegments = 100
// DefaultCleanupInterval is the interval between cleanup runs
DefaultCleanupInterval = 10 * time.Second
)
@ -53,6 +56,7 @@ type ConnectionFlow struct {
IPMeta api.IPMeta
TCPMeta api.TCPMeta
HelloBuffer []byte
Segments int // Number of segments accumulated (for memory leak prevention)
NextSeq uint32 // Expected next TCP sequence number for reassembly
SeqInit bool // Whether NextSeq has been initialized
}
@ -67,8 +71,14 @@ type ParserImpl struct {
closeOnce sync.Once
maxTrackedFlows int
maxHelloBufferBytes int
maxHelloSegments int
sourceIPFilter *ipfilter.Filter
// Metrics counters (atomic)
filteredCount uint64 // Counter for filtered packets (debug)
retransmitCount uint64 // Counter for retransmitted packets
gapDetectedCount uint64 // Counter for flows dropped due to sequence gaps
bufferExceededCount uint64 // Counter for flows dropped due to buffer limits
segmentExceededCount uint64 // Counter for flows dropped due to segment limits
}
// NewParser creates a new TLS parser with connection state tracking
@ -98,15 +108,20 @@ func NewParserWithTimeoutAndFilter(timeout time.Duration, excludeSourceIPs []str
}
p := &ParserImpl{
flows: make(map[string]*ConnectionFlow),
flowTimeout: timeout,
cleanupDone: make(chan struct{}),
cleanupClose: make(chan struct{}),
closeOnce: sync.Once{},
maxTrackedFlows: DefaultMaxTrackedFlows,
maxHelloBufferBytes: DefaultMaxHelloBufferBytes,
sourceIPFilter: filter,
filteredCount: 0,
flows: make(map[string]*ConnectionFlow),
flowTimeout: timeout,
cleanupDone: make(chan struct{}),
cleanupClose: make(chan struct{}),
closeOnce: sync.Once{},
maxTrackedFlows: DefaultMaxTrackedFlows,
maxHelloBufferBytes: DefaultMaxHelloBufferBytes,
maxHelloSegments: DefaultMaxHelloSegments,
sourceIPFilter: filter,
filteredCount: 0,
retransmitCount: 0,
gapDetectedCount: 0,
bufferExceededCount: 0,
segmentExceededCount: 0,
}
go p.cleanupLoop()
return p
@ -288,14 +303,18 @@ func (p *ParserImpl) Process(pkt api.RawPacket) (*api.TLSClientHello, error) {
return nil, nil // No payload (ACK, FIN, etc.)
}
// Check if flow exists before acquiring write lock
p.mu.RLock()
_, flowExists := p.flows[key]
p.mu.RUnlock()
// Check if this is a TLS handshake (content type 22)
isTLSHandshake := payload[0] == 22
// Early exit for non-ClientHello first packet (no SYN seen, no TLS handshake)
if !flowExists && payload[0] != 22 {
return nil, nil
// Check flow existence atomically within getOrCreateFlow
if !isTLSHandshake {
p.mu.RLock()
_, flowExists := p.flows[key]
p.mu.RUnlock()
if !flowExists {
return nil, nil
}
}
flow := p.getOrCreateFlow(key, srcIP, srcPort, dstIP, dstPort, ipMeta, tcpMeta)
@ -303,9 +322,23 @@ func (p *ParserImpl) Process(pkt api.RawPacket) (*api.TLSClientHello, error) {
return nil, nil
}
// If flow was just created and we didn't see SYN, initialize sequence from this packet
// This handles the case where SYN was missed but we still want to extract the ClientHello
flow.mu.Lock()
if !flow.SeqInit {
flow.NextSeq = tcp.Seq + uint32(len(payload))
flow.SeqInit = true
}
flow.mu.Unlock()
// Lock the flow for the entire processing to avoid race conditions
flow.mu.Lock()
defer flow.mu.Unlock()
flowMuLocked := true
defer func() {
if flowMuLocked {
flow.mu.Unlock()
}
}()
// Check if flow is already done
if flow.State == JA4_DONE {
@ -316,15 +349,24 @@ func (p *ParserImpl) Process(pkt api.RawPacket) (*api.TLSClientHello, error) {
seq := tcp.Seq
if flow.SeqInit {
if seq < flow.NextSeq {
// Retransmission — skip duplicate data
return nil, nil
// Bug 7 fix: only count as retransmission when the flow is past NEW.
// When SYN is missed, SeqInit is set from the first data packet so
// seq < NextSeq always holds for that same packet — incrementing the
// counter here was a false positive.
if flow.State != NEW {
atomic.AddUint64(&p.retransmitCount, 1)
return nil, nil
}
}
if seq > flow.NextSeq && flow.State == WAIT_CLIENT_HELLO {
// Gap detected — missing segment, drop this flow
p.mu.Lock()
delete(p.flows, key)
p.mu.Unlock()
return nil, nil
// Gap detected — missing segment in fragmented ClientHello
// Instead of dropping the flow, log and continue with available data
atomic.AddUint64(&p.gapDetectedCount, 1)
// Reset sequence tracking to continue with this segment
flow.NextSeq = seq + uint32(len(payload))
// Clear buffer since we have a gap - start fresh with this segment
flow.HelloBuffer = make([]byte, 0)
flow.Segments = 0
}
}
@ -342,9 +384,18 @@ func (p *ParserImpl) Process(pkt api.RawPacket) (*api.TLSClientHello, error) {
// Found ClientHello, mark flow as done
flow.State = JA4_DONE
flow.HelloBuffer = clientHello
flow.Segments = 0 // Reset segment count
// Extract TLS extensions (SNI, ALPN, TLS version)
extInfo, _ := extractTLSExtensions(clientHello)
extInfo, err := extractTLSExtensions(clientHello)
if err != nil {
// Log error but continue with empty extension info
extInfo = &TLSExtensionInfo{}
}
// Ensure extInfo is never nil
if extInfo == nil {
extInfo = &TLSExtensionInfo{}
}
// Generate ConnID from flow key
connID := key
@ -373,15 +424,34 @@ func (p *ParserImpl) Process(pkt api.RawPacket) (*api.TLSClientHello, error) {
// Check for fragmented ClientHello (accumulate segments)
if flow.State == WAIT_CLIENT_HELLO || flow.State == NEW {
if len(flow.HelloBuffer)+len(payload) > p.maxHelloBufferBytes {
// Buffer would exceed limit, drop this flow
// Check segment count limit (memory leak prevention)
// Bug 4 fix: release flow.mu before acquiring p.mu to avoid lock-order
// inversion with cleanupExpiredFlows (which acquires p.mu then flow.mu).
if flow.Segments >= p.maxHelloSegments {
atomic.AddUint64(&p.segmentExceededCount, 1)
flowMuLocked = false
flow.mu.Unlock()
p.mu.Lock()
delete(p.flows, key)
p.mu.Unlock()
return nil, nil
}
// Check buffer size limit (memory leak prevention)
// Bug 4 fix (same): release flow.mu before acquiring p.mu.
if len(flow.HelloBuffer)+len(payload) > p.maxHelloBufferBytes {
atomic.AddUint64(&p.bufferExceededCount, 1)
flowMuLocked = false
flow.mu.Unlock()
p.mu.Lock()
delete(p.flows, key)
p.mu.Unlock()
return nil, nil
}
flow.State = WAIT_CLIENT_HELLO
flow.HelloBuffer = append(flow.HelloBuffer, payload...)
flow.Segments++
flow.LastSeen = time.Now()
// Make a copy of the buffer for parsing (outside the lock)
@ -396,9 +466,18 @@ func (p *ParserImpl) Process(pkt api.RawPacket) (*api.TLSClientHello, error) {
if clientHello != nil {
// Complete ClientHello found
flow.State = JA4_DONE
flow.Segments = 0 // Reset segment count
// Extract TLS extensions (SNI, ALPN, TLS version)
extInfo, _ := extractTLSExtensions(clientHello)
extInfo, err := extractTLSExtensions(clientHello)
if err != nil {
// Log error but continue with empty extension info
extInfo = &TLSExtensionInfo{}
}
// Ensure extInfo is never nil
if extInfo == nil {
extInfo = &TLSExtensionInfo{}
}
// Generate ConnID from flow key
connID := key
@ -442,8 +521,33 @@ func (p *ParserImpl) getOrCreateFlow(key string, srcIP string, srcPort uint16, d
return flow
}
// If approaching flow limit, trigger aggressive cleanup of finished flows
if len(p.flows) >= p.maxTrackedFlows {
return nil
// Clean up all JA4_DONE flows first (they're already processed)
for k, flow := range p.flows {
flow.mu.Lock()
isDone := flow.State == JA4_DONE
flow.mu.Unlock()
if isDone {
delete(p.flows, k)
}
}
// If still at limit, clean up expired flows
if len(p.flows) >= p.maxTrackedFlows {
now := time.Now()
for k, flow := range p.flows {
flow.mu.Lock()
isExpired := now.Sub(flow.LastSeen) > p.flowTimeout
flow.mu.Unlock()
if isExpired {
delete(p.flows, k)
}
}
}
// Final check - if still at limit, return nil
if len(p.flows) >= p.maxTrackedFlows {
return nil
}
}
flow := &ConnectionFlow{
@ -457,6 +561,7 @@ func (p *ParserImpl) getOrCreateFlow(key string, srcIP string, srcPort uint16, d
IPMeta: ipMeta,
TCPMeta: tcpMeta,
HelloBuffer: make([]byte, 0),
Segments: 0,
}
p.flows[key] = flow
return flow
@ -470,6 +575,14 @@ func (p *ParserImpl) GetFilterStats() (filteredCount uint64, hasFilter bool) {
return atomic.LoadUint64(&p.filteredCount), true
}
// GetMetrics returns comprehensive parser metrics (for monitoring/debugging)
func (p *ParserImpl) GetMetrics() (retransmit, gapDetected, bufferExceeded, segmentExceeded uint64) {
return atomic.LoadUint64(&p.retransmitCount),
atomic.LoadUint64(&p.gapDetectedCount),
atomic.LoadUint64(&p.bufferExceededCount),
atomic.LoadUint64(&p.segmentExceededCount)
}
// Close cleans up the parser and stops background goroutines
func (p *ParserImpl) Close() error {
p.closeOnce.Do(func() {
@ -629,9 +742,20 @@ func extractTLSExtensions(payload []byte) (*TLSExtensionInfo, error) {
// Retry with sanitized payload (handles truncated/malformed extensions)
if sanitized := sanitizeTLSRecord(payload); sanitized != nil {
fp, err = tlsfingerprint.ParseClientHello(sanitized)
if err != nil {
// Return error but also provide basic info from manual parsing
info.TLSVersion = tlsVersionToString(version)
info.SNI = extractSNIFromPayload(handshakePayload)
return info, fmt.Errorf("tlsfingerprint.ParseClientHello failed: %w", err)
}
} else {
// Sanitization not available, return error with basic info
info.TLSVersion = tlsVersionToString(version)
info.SNI = extractSNIFromPayload(handshakePayload)
return info, fmt.Errorf("tlsfingerprint.ParseClientHello failed and sanitization unavailable")
}
}
if err == nil && fp != nil {
if fp != nil {
// Extract ALPN protocols
if len(fp.ALPNProtocols) > 0 {
info.ALPN = fp.ALPNProtocols

View File

@ -503,17 +503,15 @@ func TestExtractTLSExtensions(t *testing.T) {
for _, tt := range tests {
t.Run(tt.name, func(t *testing.T) {
got, err := extractTLSExtensions(tt.payload)
if err != nil {
t.Errorf("extractTLSExtensions() unexpected error = %v", err)
return
}
if (got == nil) != tt.wantNil {
t.Errorf("extractTLSExtensions() = %v, wantNil %v", got == nil, tt.wantNil)
got, _ := extractTLSExtensions(tt.payload)
// For empty/too short payloads, nil is acceptable
// For valid ClientHellos, got should contain at least partial info
if !tt.wantNil && got == nil {
t.Errorf("extractTLSExtensions() = %v, want non-nil with partial info", got)
return
}
if got != nil {
if got.TLSVersion != tt.wantVersion {
if got.TLSVersion != tt.wantVersion && tt.wantVersion != "" {
t.Errorf("TLSVersion = %v, want %v", got.TLSVersion, tt.wantVersion)
}
}
@ -1661,13 +1659,18 @@ func TestProcess_TCPGap_DropsFlow(t *testing.T) {
t.Fatal("Process(gap) should return nil")
}
// Verify flow was removed
// Verify flow was NOT removed (gap handling now continues with available data)
key := flowKey(srcIP, srcPort, dstIP, dstPort)
parser.mu.RLock()
_, exists := parser.flows[key]
parser.mu.RUnlock()
if exists {
t.Fatal("flow should be removed after sequence gap")
if !exists {
t.Fatal("flow should NOT be removed after sequence gap (gap handling changed)")
}
// Verify gap was detected (counter incremented)
_, gapDetected, _, _ := parser.GetMetrics()
if gapDetected == 0 {
t.Fatal("gapDetected counter should be incremented")
}
}
@ -1790,3 +1793,32 @@ func TestProcess_TLS13ClientHello_CorrectVersion(t *testing.T) {
t.Errorf("SNI = %q, want \"tls13.example.com\"", result.SNI)
}
}
// TestProcess_MissedSYN_NoFalseRetransmit verifies Bug 7 fix:
// when SYN is missed, the first data packet must NOT increment retransmitCount
// even though seq < NextSeq would evaluate to true (because NextSeq was
// initialised from that very same packet).
func TestProcess_MissedSYN_NoFalseRetransmit(t *testing.T) {
parser := NewParser()
defer parser.Close()
srcIP := "10.0.0.1"
dstIP := "10.0.0.2"
srcPort := uint16(12345)
dstPort := uint16(443)
// Build a minimal TLS ClientHello payload.
payload := createMinimalTLSClientHelloWithSNIAndALPN("test.example.com", nil)
// Send without a preceding SYN — seq starts from 100.
pkt := buildRawPacketWithSeq(t, srcIP, dstIP, srcPort, dstPort, payload, 100)
_, err := parser.Process(pkt)
if err != nil {
t.Fatalf("Process() error: %v", err)
}
retransmit, _, _, _ := parser.GetMetrics()
if retransmit != 0 {
t.Errorf("retransmitCount = %d after first packet on a new flow (SYN missed); want 0", retransmit)
}
}