package main import ( "database/sql" "encoding/json" "fmt" "log" "math" "sort" "strconv" "strings" "sync" "sync/atomic" "time" ) // payloadTypeNames maps payload_type int → human-readable name (firmware-standard). var payloadTypeNames = map[int]string{ 0: "REQ", 1: "RESPONSE", 2: "TXT_MSG", 3: "ACK", 4: "ADVERT", 5: "GRP_TXT", 7: "ANON_REQ", 8: "PATH", 9: "TRACE", 11: "CONTROL", } // StoreTx is an in-memory transmission with embedded observations. type StoreTx struct { ID int RawHex string Hash string FirstSeen string RouteType *int PayloadType *int DecodedJSON string Observations []*StoreObs ObservationCount int // Display fields from longest-path observation ObserverID string ObserverName string SNR *float64 RSSI *float64 PathJSON string Direction string // Cached parsed fields (set once, read many) parsedPath []string // cached parsePathJSON result pathParsed bool // whether parsedPath has been set } // StoreObs is a lean in-memory observation (no duplication of transmission fields). type StoreObs struct { ID int TransmissionID int ObserverID string ObserverName string Direction string SNR *float64 RSSI *float64 Score *int PathJSON string Timestamp string } // PacketStore holds all transmissions in memory with indexes for fast queries. type PacketStore struct { mu sync.RWMutex db *DB packets []*StoreTx // sorted by first_seen DESC byHash map[string]*StoreTx // hash → *StoreTx byTxID map[int]*StoreTx // transmission_id → *StoreTx byObsID map[int]*StoreObs // observation_id → *StoreObs byObserver map[string][]*StoreObs // observer_id → observations byNode map[string][]*StoreTx // pubkey → transmissions nodeHashes map[string]map[string]bool // pubkey → Set byPayloadType map[int][]*StoreTx // payload_type → transmissions loaded bool totalObs int insertCount int64 queryCount int64 // Response caches (separate mutex to avoid contention with store RWMutex) cacheMu sync.Mutex rfCache map[string]*cachedResult // region → cached RF result topoCache map[string]*cachedResult // region → cached topology result hashCache map[string]*cachedResult // region → cached hash-sizes result chanCache map[string]*cachedResult // region → cached channels result rfCacheTTL time.Duration cacheHits int64 cacheMisses int64 // Cached node list + prefix map (rebuilt on demand, shared across analytics) nodeCache []nodeInfo nodePM *prefixMap nodeCacheTime time.Time } type cachedResult struct { data map[string]interface{} expiresAt time.Time } // NewPacketStore creates a new empty packet store backed by db. func NewPacketStore(db *DB) *PacketStore { return &PacketStore{ db: db, packets: make([]*StoreTx, 0, 65536), byHash: make(map[string]*StoreTx, 65536), byTxID: make(map[int]*StoreTx, 65536), byObsID: make(map[int]*StoreObs, 65536), byObserver: make(map[string][]*StoreObs), byNode: make(map[string][]*StoreTx), nodeHashes: make(map[string]map[string]bool), byPayloadType: make(map[int][]*StoreTx), rfCache: make(map[string]*cachedResult), topoCache: make(map[string]*cachedResult), hashCache: make(map[string]*cachedResult), chanCache: make(map[string]*cachedResult), rfCacheTTL: 15 * time.Second, } } // Load reads all transmissions + observations from SQLite into memory. func (s *PacketStore) Load() error { s.mu.Lock() defer s.mu.Unlock() t0 := time.Now() var loadSQL string if s.db.isV3 { loadSQL = `SELECT t.id, t.raw_hex, t.hash, t.first_seen, t.route_type, t.payload_type, t.payload_version, t.decoded_json, o.id, obs.id, obs.name, o.direction, o.snr, o.rssi, o.score, o.path_json, datetime(o.timestamp, 'unixepoch') FROM transmissions t LEFT JOIN observations o ON o.transmission_id = t.id LEFT JOIN observers obs ON obs.rowid = o.observer_idx ORDER BY t.first_seen DESC, o.timestamp DESC` } else { loadSQL = `SELECT t.id, t.raw_hex, t.hash, t.first_seen, t.route_type, t.payload_type, t.payload_version, t.decoded_json, o.id, o.observer_id, o.observer_name, o.direction, o.snr, o.rssi, o.score, o.path_json, o.timestamp FROM transmissions t LEFT JOIN observations o ON o.transmission_id = t.id ORDER BY t.first_seen DESC, o.timestamp DESC` } rows, err := s.db.conn.Query(loadSQL) if err != nil { return err } defer rows.Close() for rows.Next() { var txID int var rawHex, hash, firstSeen, decodedJSON sql.NullString var routeType, payloadType, payloadVersion sql.NullInt64 var obsID sql.NullInt64 var observerID, observerName, direction, pathJSON, obsTimestamp sql.NullString var snr, rssi sql.NullFloat64 var score sql.NullInt64 if err := rows.Scan(&txID, &rawHex, &hash, &firstSeen, &routeType, &payloadType, &payloadVersion, &decodedJSON, &obsID, &observerID, &observerName, &direction, &snr, &rssi, &score, &pathJSON, &obsTimestamp); err != nil { log.Printf("[store] scan error: %v", err) continue } hashStr := nullStrVal(hash) tx := s.byHash[hashStr] if tx == nil { tx = &StoreTx{ ID: txID, RawHex: nullStrVal(rawHex), Hash: hashStr, FirstSeen: nullStrVal(firstSeen), RouteType: nullIntPtr(routeType), PayloadType: nullIntPtr(payloadType), DecodedJSON: nullStrVal(decodedJSON), } s.byHash[hashStr] = tx s.packets = append(s.packets, tx) s.byTxID[txID] = tx s.indexByNode(tx) if tx.PayloadType != nil { pt := *tx.PayloadType s.byPayloadType[pt] = append(s.byPayloadType[pt], tx) } } if obsID.Valid { oid := int(obsID.Int64) obsIDStr := nullStrVal(observerID) obsPJ := nullStrVal(pathJSON) // Dedup: skip if same observer + same path already loaded isDupe := false for _, existing := range tx.Observations { if existing.ObserverID == obsIDStr && existing.PathJSON == obsPJ { isDupe = true break } } if isDupe { continue } obs := &StoreObs{ ID: oid, TransmissionID: txID, ObserverID: obsIDStr, ObserverName: nullStrVal(observerName), Direction: nullStrVal(direction), SNR: nullFloatPtr(snr), RSSI: nullFloatPtr(rssi), Score: nullIntPtr(score), PathJSON: obsPJ, Timestamp: nullStrVal(obsTimestamp), } tx.Observations = append(tx.Observations, obs) tx.ObservationCount++ s.byObsID[oid] = obs if obsIDStr != "" { s.byObserver[obsIDStr] = append(s.byObserver[obsIDStr], obs) } s.totalObs++ } } // Post-load: pick best observation (longest path) for each transmission for _, tx := range s.packets { pickBestObservation(tx) } s.loaded = true elapsed := time.Since(t0) estMB := (len(s.packets)*450 + s.totalObs*100) / (1024 * 1024) log.Printf("[store] Loaded %d transmissions (%d observations) in %v (~%dMB est)", len(s.packets), s.totalObs, elapsed, estMB) return nil } // pickBestObservation selects the observation with the longest path // and sets it as the transmission's display observation. func pickBestObservation(tx *StoreTx) { if len(tx.Observations) == 0 { return } best := tx.Observations[0] bestLen := pathLen(best.PathJSON) for _, obs := range tx.Observations[1:] { l := pathLen(obs.PathJSON) if l > bestLen { best = obs bestLen = l } } tx.ObserverID = best.ObserverID tx.ObserverName = best.ObserverName tx.SNR = best.SNR tx.RSSI = best.RSSI tx.PathJSON = best.PathJSON tx.Direction = best.Direction tx.pathParsed = false // invalidate cached parsed path } func pathLen(pathJSON string) int { if pathJSON == "" { return 0 } var hops []interface{} if json.Unmarshal([]byte(pathJSON), &hops) != nil { return 0 } return len(hops) } // indexByNode extracts pubkeys from decoded_json and indexes the transmission. func (s *PacketStore) indexByNode(tx *StoreTx) { if tx.DecodedJSON == "" { return } var decoded map[string]interface{} if json.Unmarshal([]byte(tx.DecodedJSON), &decoded) != nil { return } for _, field := range []string{"pubKey", "destPubKey", "srcPubKey"} { if v, ok := decoded[field].(string); ok && v != "" { if s.nodeHashes[v] == nil { s.nodeHashes[v] = make(map[string]bool) } if s.nodeHashes[v][tx.Hash] { continue } s.nodeHashes[v][tx.Hash] = true s.byNode[v] = append(s.byNode[v], tx) } } } // QueryPackets returns filtered, paginated packets from memory. func (s *PacketStore) QueryPackets(q PacketQuery) *PacketResult { atomic.AddInt64(&s.queryCount, 1) s.mu.RLock() defer s.mu.RUnlock() if q.Limit <= 0 { q.Limit = 50 } if q.Order == "" { q.Order = "DESC" } results := s.filterPackets(q) total := len(results) if q.Order == "ASC" { sorted := make([]*StoreTx, len(results)) copy(sorted, results) sort.Slice(sorted, func(i, j int) bool { return sorted[i].FirstSeen < sorted[j].FirstSeen }) results = sorted } // Paginate start := q.Offset if start >= len(results) { return &PacketResult{Packets: []map[string]interface{}{}, Total: total} } end := start + q.Limit if end > len(results) { end = len(results) } packets := make([]map[string]interface{}, 0, end-start) for _, tx := range results[start:end] { packets = append(packets, txToMap(tx)) } return &PacketResult{Packets: packets, Total: total} } // QueryGroupedPackets returns transmissions grouped by hash (already 1:1). func (s *PacketStore) QueryGroupedPackets(q PacketQuery) *PacketResult { atomic.AddInt64(&s.queryCount, 1) s.mu.RLock() defer s.mu.RUnlock() if q.Limit <= 0 { q.Limit = 50 } results := s.filterPackets(q) // Build grouped output sorted by latest observation DESC type groupEntry struct { tx *StoreTx latest string } entries := make([]groupEntry, len(results)) for i, tx := range results { latest := tx.FirstSeen for _, obs := range tx.Observations { if obs.Timestamp > latest { latest = obs.Timestamp } } entries[i] = groupEntry{tx: tx, latest: latest} } sort.Slice(entries, func(i, j int) bool { return entries[i].latest > entries[j].latest }) total := len(entries) start := q.Offset if start >= total { return &PacketResult{Packets: []map[string]interface{}{}, Total: total} } end := start + q.Limit if end > total { end = total } packets := make([]map[string]interface{}, 0, end-start) for _, e := range entries[start:end] { tx := e.tx observerCount := 0 seen := make(map[string]bool) for _, obs := range tx.Observations { if obs.ObserverID != "" && !seen[obs.ObserverID] { seen[obs.ObserverID] = true observerCount++ } } packets = append(packets, map[string]interface{}{ "hash": strOrNil(tx.Hash), "first_seen": strOrNil(tx.FirstSeen), "count": tx.ObservationCount, "observer_count": observerCount, "observation_count": tx.ObservationCount, "latest": strOrNil(e.latest), "observer_id": strOrNil(tx.ObserverID), "observer_name": strOrNil(tx.ObserverName), "path_json": strOrNil(tx.PathJSON), "payload_type": intPtrOrNil(tx.PayloadType), "route_type": intPtrOrNil(tx.RouteType), "raw_hex": strOrNil(tx.RawHex), "decoded_json": strOrNil(tx.DecodedJSON), "snr": floatPtrOrNil(tx.SNR), "rssi": floatPtrOrNil(tx.RSSI), }) } return &PacketResult{Packets: packets, Total: total} } // GetStoreStats returns aggregate counts (packet data from memory, node/observer from DB). func (s *PacketStore) GetStoreStats() (*Stats, error) { s.mu.RLock() txCount := len(s.packets) obsCount := s.totalObs s.mu.RUnlock() st := &Stats{ TotalTransmissions: txCount, TotalPackets: txCount, TotalObservations: obsCount, } sevenDaysAgo := time.Now().Add(-7 * 24 * time.Hour).Format(time.RFC3339) s.db.conn.QueryRow("SELECT COUNT(*) FROM nodes WHERE last_seen > ?", sevenDaysAgo).Scan(&st.TotalNodes) s.db.conn.QueryRow("SELECT COUNT(*) FROM nodes").Scan(&st.TotalNodesAllTime) s.db.conn.QueryRow("SELECT COUNT(*) FROM observers").Scan(&st.TotalObservers) oneHourAgo := time.Now().Add(-1 * time.Hour).Unix() s.db.conn.QueryRow("SELECT COUNT(*) FROM observations WHERE timestamp > ?", oneHourAgo).Scan(&st.PacketsLastHour) return st, nil } // GetPerfStoreStats returns packet store statistics for /api/perf. func (s *PacketStore) GetPerfStoreStats() map[string]interface{} { s.mu.RLock() totalLoaded := len(s.packets) totalObs := s.totalObs hashIdx := len(s.byHash) txIdx := len(s.byTxID) obsIdx := len(s.byObsID) observerIdx := len(s.byObserver) nodeIdx := len(s.byNode) ptIdx := len(s.byPayloadType) // Count distinct pubkeys with ADVERT observations (matches Node.js _advertByObserver.size) advertByObsCount := 0 if adverts, ok := s.byPayloadType[4]; ok { seen := make(map[string]bool) for _, tx := range adverts { if tx.DecodedJSON == "" { continue } var d map[string]interface{} if json.Unmarshal([]byte(tx.DecodedJSON), &d) != nil { continue } pk := "" if v, ok := d["pubKey"].(string); ok { pk = v } else if v, ok := d["public_key"].(string); ok { pk = v } if pk != "" && !seen[pk] { seen[pk] = true advertByObsCount++ } } } s.mu.RUnlock() // Rough estimate: ~430 bytes per packet + ~200 per observation estimatedMB := math.Round(float64(totalLoaded*430+totalObs*200)/1048576*10) / 10 return map[string]interface{}{ "totalLoaded": totalLoaded, "totalObservations": totalObs, "evicted": 0, "inserts": atomic.LoadInt64(&s.insertCount), "queries": atomic.LoadInt64(&s.queryCount), "inMemory": totalLoaded, "sqliteOnly": false, "maxPackets": 2386092, "estimatedMB": estimatedMB, "maxMB": 1024, "indexes": map[string]interface{}{ "byHash": hashIdx, "byTxID": txIdx, "byObsID": obsIdx, "byObserver": observerIdx, "byNode": nodeIdx, "byPayloadType": ptIdx, "advertByObserver": advertByObsCount, }, } } // GetCacheStats returns RF cache hit/miss statistics. func (s *PacketStore) GetCacheStats() map[string]interface{} { s.cacheMu.Lock() size := len(s.rfCache) + len(s.topoCache) + len(s.hashCache) + len(s.chanCache) hits := s.cacheHits misses := s.cacheMisses s.cacheMu.Unlock() var hitRate float64 if hits+misses > 0 { hitRate = math.Round(float64(hits)/float64(hits+misses)*1000) / 10 } return map[string]interface{}{ "size": size, "hits": hits, "misses": misses, "staleHits": 0, "recomputes": misses, "hitRate": hitRate, } } // GetTransmissionByID returns a transmission by its DB ID, formatted as a map. func (s *PacketStore) GetTransmissionByID(id int) map[string]interface{} { s.mu.RLock() defer s.mu.RUnlock() tx := s.byTxID[id] if tx == nil { return nil } return txToMap(tx) } // GetPacketByHash returns a transmission by content hash. func (s *PacketStore) GetPacketByHash(hash string) map[string]interface{} { s.mu.RLock() defer s.mu.RUnlock() tx := s.byHash[strings.ToLower(hash)] if tx == nil { return nil } return txToMap(tx) } // GetPacketByID returns an observation (enriched with transmission fields) by observation ID. func (s *PacketStore) GetPacketByID(id int) map[string]interface{} { s.mu.RLock() defer s.mu.RUnlock() obs := s.byObsID[id] if obs == nil { return nil } return s.enrichObs(obs) } // GetObservationsForHash returns all observations for a hash, enriched with transmission fields. func (s *PacketStore) GetObservationsForHash(hash string) []map[string]interface{} { s.mu.RLock() defer s.mu.RUnlock() tx := s.byHash[strings.ToLower(hash)] if tx == nil { return []map[string]interface{}{} } result := make([]map[string]interface{}, 0, len(tx.Observations)) for _, obs := range tx.Observations { result = append(result, s.enrichObs(obs)) } return result } // GetTimestamps returns transmission first_seen timestamps after since, in ASC order. func (s *PacketStore) GetTimestamps(since string) []string { s.mu.RLock() defer s.mu.RUnlock() // packets sorted newest first — scan from start until older than since var result []string for _, tx := range s.packets { if tx.FirstSeen <= since { break } result = append(result, tx.FirstSeen) } // Reverse to get ASC order for i, j := 0, len(result)-1; i < j; i, j = i+1, j-1 { result[i], result[j] = result[j], result[i] } return result } // QueryMultiNodePackets filters packets matching any of the given pubkeys. func (s *PacketStore) QueryMultiNodePackets(pubkeys []string, limit, offset int, order, since, until string) *PacketResult { s.mu.RLock() defer s.mu.RUnlock() if len(pubkeys) == 0 { return &PacketResult{Packets: []map[string]interface{}{}, Total: 0} } if limit <= 0 { limit = 50 } resolved := make([]string, len(pubkeys)) for i, pk := range pubkeys { resolved[i] = s.db.resolveNodePubkey(pk) } var filtered []*StoreTx for _, tx := range s.packets { if tx.DecodedJSON == "" { continue } match := false for _, pk := range resolved { if strings.Contains(tx.DecodedJSON, pk) { match = true break } } if !match { continue } if since != "" && tx.FirstSeen < since { continue } if until != "" && tx.FirstSeen > until { continue } filtered = append(filtered, tx) } total := len(filtered) if order == "ASC" { sort.Slice(filtered, func(i, j int) bool { return filtered[i].FirstSeen < filtered[j].FirstSeen }) } if offset >= total { return &PacketResult{Packets: []map[string]interface{}{}, Total: total} } end := offset + limit if end > total { end = total } packets := make([]map[string]interface{}, 0, end-offset) for _, tx := range filtered[offset:end] { packets = append(packets, txToMap(tx)) } return &PacketResult{Packets: packets, Total: total} } // IngestNewFromDB loads new transmissions from SQLite into memory and returns // broadcast-ready maps plus the new max transmission ID. func (s *PacketStore) IngestNewFromDB(sinceID, limit int) ([]map[string]interface{}, int) { if limit <= 0 { limit = 100 } var querySQL string if s.db.isV3 { querySQL = `SELECT t.id, t.raw_hex, t.hash, t.first_seen, t.route_type, t.payload_type, t.payload_version, t.decoded_json, o.id, obs.id, obs.name, o.direction, o.snr, o.rssi, o.score, o.path_json, datetime(o.timestamp, 'unixepoch') FROM transmissions t LEFT JOIN observations o ON o.transmission_id = t.id LEFT JOIN observers obs ON obs.rowid = o.observer_idx WHERE t.id > ? ORDER BY t.id ASC, o.timestamp DESC` } else { querySQL = `SELECT t.id, t.raw_hex, t.hash, t.first_seen, t.route_type, t.payload_type, t.payload_version, t.decoded_json, o.id, o.observer_id, o.observer_name, o.direction, o.snr, o.rssi, o.score, o.path_json, o.timestamp FROM transmissions t LEFT JOIN observations o ON o.transmission_id = t.id WHERE t.id > ? ORDER BY t.id ASC, o.timestamp DESC` } rows, err := s.db.conn.Query(querySQL, sinceID) if err != nil { log.Printf("[store] ingest query error: %v", err) return nil, sinceID } defer rows.Close() // Scan into temp structures type tempRow struct { txID int rawHex, hash, firstSeen, decodedJSON string routeType, payloadType *int obsID *int observerID, observerName, direction, pathJSON, obsTS string snr, rssi *float64 score *int } var tempRows []tempRow txCount := 0 lastTxID := sinceID for rows.Next() { var txID int var rawHex, hash, firstSeen, decodedJSON sql.NullString var routeType, payloadType, payloadVersion sql.NullInt64 var obsIDVal sql.NullInt64 var observerID, observerName, direction, pathJSON, obsTimestamp sql.NullString var snrVal, rssiVal sql.NullFloat64 var scoreVal sql.NullInt64 if err := rows.Scan(&txID, &rawHex, &hash, &firstSeen, &routeType, &payloadType, &payloadVersion, &decodedJSON, &obsIDVal, &observerID, &observerName, &direction, &snrVal, &rssiVal, &scoreVal, &pathJSON, &obsTimestamp); err != nil { continue } if txID != lastTxID { txCount++ if txCount > limit { break } lastTxID = txID } tr := tempRow{ txID: txID, rawHex: nullStrVal(rawHex), hash: nullStrVal(hash), firstSeen: nullStrVal(firstSeen), decodedJSON: nullStrVal(decodedJSON), routeType: nullIntPtr(routeType), payloadType: nullIntPtr(payloadType), observerID: nullStrVal(observerID), observerName: nullStrVal(observerName), direction: nullStrVal(direction), pathJSON: nullStrVal(pathJSON), obsTS: nullStrVal(obsTimestamp), snr: nullFloatPtr(snrVal), rssi: nullFloatPtr(rssiVal), score: nullIntPtr(scoreVal), } if obsIDVal.Valid { oid := int(obsIDVal.Int64) tr.obsID = &oid } tempRows = append(tempRows, tr) } if len(tempRows) == 0 { return nil, sinceID } // Now lock and merge into store s.mu.Lock() defer s.mu.Unlock() newMaxID := sinceID broadcastTxs := make(map[int]*StoreTx) // track new transmissions for broadcast var broadcastOrder []int for _, r := range tempRows { if r.txID > newMaxID { newMaxID = r.txID } tx := s.byHash[r.hash] if tx == nil { tx = &StoreTx{ ID: r.txID, RawHex: r.rawHex, Hash: r.hash, FirstSeen: r.firstSeen, RouteType: r.routeType, PayloadType: r.payloadType, DecodedJSON: r.decodedJSON, } s.byHash[r.hash] = tx // Prepend (newest first) s.packets = append([]*StoreTx{tx}, s.packets...) s.byTxID[r.txID] = tx s.indexByNode(tx) if tx.PayloadType != nil { pt := *tx.PayloadType s.byPayloadType[pt] = append(s.byPayloadType[pt], tx) } if _, exists := broadcastTxs[r.txID]; !exists { broadcastTxs[r.txID] = tx broadcastOrder = append(broadcastOrder, r.txID) } } if r.obsID != nil { oid := *r.obsID // Dedup isDupe := false for _, existing := range tx.Observations { if existing.ObserverID == r.observerID && existing.PathJSON == r.pathJSON { isDupe = true break } } if isDupe { continue } obs := &StoreObs{ ID: oid, TransmissionID: r.txID, ObserverID: r.observerID, ObserverName: r.observerName, Direction: r.direction, SNR: r.snr, RSSI: r.rssi, Score: r.score, PathJSON: r.pathJSON, Timestamp: r.obsTS, } tx.Observations = append(tx.Observations, obs) tx.ObservationCount++ s.byObsID[oid] = obs if r.observerID != "" { s.byObserver[r.observerID] = append(s.byObserver[r.observerID], obs) } s.totalObs++ } } // Pick best observation for new transmissions for _, tx := range broadcastTxs { pickBestObservation(tx) } // Build broadcast maps (same shape as Node.js WS broadcast) result := make([]map[string]interface{}, 0, len(broadcastOrder)) for _, txID := range broadcastOrder { tx := broadcastTxs[txID] // Build decoded object with header.payloadTypeName for live.js decoded := map[string]interface{}{ "header": map[string]interface{}{ "payloadTypeName": resolvePayloadTypeName(tx.PayloadType), }, } if tx.DecodedJSON != "" { var payload map[string]interface{} if json.Unmarshal([]byte(tx.DecodedJSON), &payload) == nil { decoded["payload"] = payload } } result = append(result, map[string]interface{}{ "id": tx.ID, "raw_hex": strOrNil(tx.RawHex), "hash": strOrNil(tx.Hash), "first_seen": strOrNil(tx.FirstSeen), "route_type": intPtrOrNil(tx.RouteType), "payload_type": intPtrOrNil(tx.PayloadType), "decoded_json": strOrNil(tx.DecodedJSON), "decoded": decoded, "observer_id": strOrNil(tx.ObserverID), "observer_name": strOrNil(tx.ObserverName), "snr": floatPtrOrNil(tx.SNR), "rssi": floatPtrOrNil(tx.RSSI), "path_json": strOrNil(tx.PathJSON), "observation_count": tx.ObservationCount, }) } // Invalidate analytics caches since new data was ingested if len(result) > 0 { s.cacheMu.Lock() s.rfCache = make(map[string]*cachedResult) s.topoCache = make(map[string]*cachedResult) s.hashCache = make(map[string]*cachedResult) s.chanCache = make(map[string]*cachedResult) s.cacheMu.Unlock() } return result, newMaxID } // MaxTransmissionID returns the highest transmission ID in the store. func (s *PacketStore) MaxTransmissionID() int { s.mu.RLock() defer s.mu.RUnlock() maxID := 0 for id := range s.byTxID { if id > maxID { maxID = id } } return maxID } // --- Internal filter/query helpers --- // filterPackets applies PacketQuery filters to the in-memory packet list. func (s *PacketStore) filterPackets(q PacketQuery) []*StoreTx { // Fast path: single-key index lookups if q.Hash != "" && q.Type == nil && q.Route == nil && q.Observer == "" && q.Region == "" && q.Node == "" && q.Since == "" && q.Until == "" { h := strings.ToLower(q.Hash) tx := s.byHash[h] if tx == nil { return nil } return []*StoreTx{tx} } if q.Observer != "" && q.Type == nil && q.Route == nil && q.Region == "" && q.Node == "" && q.Hash == "" && q.Since == "" && q.Until == "" { return s.transmissionsForObserver(q.Observer, nil) } results := s.packets if q.Type != nil { t := *q.Type results = filterTxSlice(results, func(tx *StoreTx) bool { return tx.PayloadType != nil && *tx.PayloadType == t }) } if q.Route != nil { r := *q.Route results = filterTxSlice(results, func(tx *StoreTx) bool { return tx.RouteType != nil && *tx.RouteType == r }) } if q.Observer != "" { results = s.transmissionsForObserver(q.Observer, results) } if q.Hash != "" { h := strings.ToLower(q.Hash) results = filterTxSlice(results, func(tx *StoreTx) bool { return tx.Hash == h }) } if q.Since != "" { results = filterTxSlice(results, func(tx *StoreTx) bool { return tx.FirstSeen > q.Since }) } if q.Until != "" { results = filterTxSlice(results, func(tx *StoreTx) bool { return tx.FirstSeen < q.Until }) } if q.Region != "" { regionObservers := s.resolveRegionObservers(q.Region) if len(regionObservers) > 0 { results = filterTxSlice(results, func(tx *StoreTx) bool { for _, obs := range tx.Observations { if regionObservers[obs.ObserverID] { return true } } return false }) } else { results = nil } } if q.Node != "" { pk := s.db.resolveNodePubkey(q.Node) // Use node index if available if indexed, ok := s.byNode[pk]; ok && results == nil { results = indexed } else { results = filterTxSlice(results, func(tx *StoreTx) bool { if tx.DecodedJSON == "" { return false } return strings.Contains(tx.DecodedJSON, pk) || strings.Contains(tx.DecodedJSON, q.Node) }) } } return results } // transmissionsForObserver returns unique transmissions for an observer. func (s *PacketStore) transmissionsForObserver(observerID string, from []*StoreTx) []*StoreTx { if from != nil { return filterTxSlice(from, func(tx *StoreTx) bool { for _, obs := range tx.Observations { if obs.ObserverID == observerID { return true } } return false }) } // Use byObserver index observations := s.byObserver[observerID] if len(observations) == 0 { return nil } seen := make(map[int]bool, len(observations)) var result []*StoreTx for _, obs := range observations { if seen[obs.TransmissionID] { continue } seen[obs.TransmissionID] = true tx := s.byTxID[obs.TransmissionID] if tx != nil { result = append(result, tx) } } return result } // resolveRegionObservers returns a set of observer IDs for a given IATA region. func (s *PacketStore) resolveRegionObservers(region string) map[string]bool { ids, err := s.db.GetObserverIdsForRegion(region) if err != nil || len(ids) == 0 { return nil } m := make(map[string]bool, len(ids)) for _, id := range ids { m[id] = true } return m } // enrichObs returns a map with observation fields + transmission fields. func (s *PacketStore) enrichObs(obs *StoreObs) map[string]interface{} { tx := s.byTxID[obs.TransmissionID] m := map[string]interface{}{ "id": obs.ID, "timestamp": strOrNil(obs.Timestamp), "observer_id": strOrNil(obs.ObserverID), "observer_name": strOrNil(obs.ObserverName), "direction": strOrNil(obs.Direction), "snr": floatPtrOrNil(obs.SNR), "rssi": floatPtrOrNil(obs.RSSI), "score": intPtrOrNil(obs.Score), "path_json": strOrNil(obs.PathJSON), } if tx != nil { m["hash"] = strOrNil(tx.Hash) m["raw_hex"] = strOrNil(tx.RawHex) m["payload_type"] = intPtrOrNil(tx.PayloadType) m["route_type"] = intPtrOrNil(tx.RouteType) m["decoded_json"] = strOrNil(tx.DecodedJSON) } return m } // --- Conversion helpers --- // txToMap converts a StoreTx to the map shape matching scanTransmissionRow output. func txToMap(tx *StoreTx) map[string]interface{} { return map[string]interface{}{ "id": tx.ID, "raw_hex": strOrNil(tx.RawHex), "hash": strOrNil(tx.Hash), "first_seen": strOrNil(tx.FirstSeen), "timestamp": strOrNil(tx.FirstSeen), "route_type": intPtrOrNil(tx.RouteType), "payload_type": intPtrOrNil(tx.PayloadType), "decoded_json": strOrNil(tx.DecodedJSON), "observation_count": tx.ObservationCount, "observer_id": strOrNil(tx.ObserverID), "observer_name": strOrNil(tx.ObserverName), "snr": floatPtrOrNil(tx.SNR), "rssi": floatPtrOrNil(tx.RSSI), "path_json": strOrNil(tx.PathJSON), "direction": strOrNil(tx.Direction), } } func strOrNil(s string) interface{} { if s == "" { return nil } return s } func intPtrOrNil(p *int) interface{} { if p == nil { return nil } return *p } func floatPtrOrNil(p *float64) interface{} { if p == nil { return nil } return *p } func nullIntPtr(ni sql.NullInt64) *int { if ni.Valid { v := int(ni.Int64) return &v } return nil } func nullFloatPtr(nf sql.NullFloat64) *float64 { if nf.Valid { return &nf.Float64 } return nil } // resolvePayloadTypeName returns the firmware-standard name for a payload_type. func resolvePayloadTypeName(pt *int) string { if pt == nil { return "UNKNOWN" } if name, ok := payloadTypeNames[*pt]; ok { return name } return fmt.Sprintf("UNK(%d)", *pt) } // txGetParsedPath returns cached parsed path hops, parsing on first call. func txGetParsedPath(tx *StoreTx) []string { if tx.pathParsed { return tx.parsedPath } tx.parsedPath = parsePathJSON(tx.PathJSON) tx.pathParsed = true return tx.parsedPath } func filterTxSlice(s []*StoreTx, fn func(*StoreTx) bool) []*StoreTx { var result []*StoreTx for _, tx := range s { if fn(tx) { result = append(result, tx) } } return result } // GetChannels returns channel list from in-memory packets (payload_type 5, decoded type CHAN). func (s *PacketStore) GetChannels(region string) []map[string]interface{} { s.mu.RLock() defer s.mu.RUnlock() var regionObs map[string]bool if region != "" { regionObs = s.resolveRegionObservers(region) } type chanInfo struct { Hash string Name string LastMessage interface{} LastSender interface{} MessageCount int LastActivity string } type decodedGrp struct { Type string `json:"type"` Channel string `json:"channel"` Text string `json:"text"` Sender string `json:"sender"` } channelMap := map[string]*chanInfo{} grpTxts := s.byPayloadType[5] for _, tx := range grpTxts { // Region filter: check if any observation is from a regional observer if regionObs != nil { match := false for _, obs := range tx.Observations { if regionObs[obs.ObserverID] { match = true break } } if !match { continue } } var decoded decodedGrp if json.Unmarshal([]byte(tx.DecodedJSON), &decoded) != nil { continue } if decoded.Type != "CHAN" { continue } channelName := decoded.Channel if channelName == "" { channelName = "unknown" } key := channelName ch := channelMap[key] if ch == nil { ch = &chanInfo{ Hash: key, Name: channelName, LastActivity: tx.FirstSeen, } channelMap[key] = ch } ch.MessageCount++ if tx.FirstSeen >= ch.LastActivity { ch.LastActivity = tx.FirstSeen if decoded.Text != "" { idx := strings.Index(decoded.Text, ": ") if idx > 0 { ch.LastMessage = decoded.Text[idx+2:] } else { ch.LastMessage = decoded.Text } if decoded.Sender != "" { ch.LastSender = decoded.Sender } } } } channels := make([]map[string]interface{}, 0, len(channelMap)) for _, ch := range channelMap { channels = append(channels, map[string]interface{}{ "hash": ch.Hash, "name": ch.Name, "lastMessage": ch.LastMessage, "lastSender": ch.LastSender, "messageCount": ch.MessageCount, "lastActivity": ch.LastActivity, }) } return channels } // GetChannelMessages returns deduplicated messages for a channel from in-memory packets. func (s *PacketStore) GetChannelMessages(channelHash string, limit, offset int) ([]map[string]interface{}, int) { s.mu.RLock() defer s.mu.RUnlock() if limit <= 0 { limit = 100 } type msgEntry struct { Data map[string]interface{} Repeats int Observers []string } msgMap := map[string]*msgEntry{} var msgOrder []string // Iterate type-5 packets oldest-first (byPayloadType is in load order = newest first) type decodedMsg struct { Type string `json:"type"` Channel string `json:"channel"` Text string `json:"text"` Sender string `json:"sender"` SenderTimestamp interface{} `json:"sender_timestamp"` PathLen int `json:"path_len"` } grpTxts := s.byPayloadType[5] for i := len(grpTxts) - 1; i >= 0; i-- { tx := grpTxts[i] if tx.DecodedJSON == "" { continue } var decoded decodedMsg if json.Unmarshal([]byte(tx.DecodedJSON), &decoded) != nil { continue } if decoded.Type != "CHAN" { continue } ch := decoded.Channel if ch == "" { ch = "unknown" } if ch != channelHash { continue } text := decoded.Text sender := decoded.Sender if sender == "" && text != "" { idx := strings.Index(text, ": ") if idx > 0 && idx < 50 { sender = text[:idx] } } dedupeKey := sender + ":" + tx.Hash if existing, ok := msgMap[dedupeKey]; ok { existing.Repeats++ existing.Data["repeats"] = existing.Repeats // Add observer if new obsName := tx.ObserverName if obsName == "" { obsName = tx.ObserverID } if obsName != "" { found := false for _, o := range existing.Observers { if o == obsName { found = true break } } if !found { existing.Observers = append(existing.Observers, obsName) existing.Data["observers"] = existing.Observers } } } else { displaySender := sender displayText := text if text != "" { idx := strings.Index(text, ": ") if idx > 0 && idx < 50 { displaySender = text[:idx] displayText = text[idx+2:] } } hops := pathLen(tx.PathJSON) var snrVal interface{} if tx.SNR != nil { snrVal = *tx.SNR } senderTs := decoded.SenderTimestamp observers := []string{} obsName := tx.ObserverName if obsName == "" { obsName = tx.ObserverID } if obsName != "" { observers = []string{obsName} } entry := &msgEntry{ Data: map[string]interface{}{ "sender": displaySender, "text": displayText, "timestamp": strOrNil(tx.FirstSeen), "sender_timestamp": senderTs, "packetId": tx.ID, "packetHash": strOrNil(tx.Hash), "repeats": 1, "observers": observers, "hops": hops, "snr": snrVal, }, Repeats: 1, Observers: observers, } msgMap[dedupeKey] = entry msgOrder = append(msgOrder, dedupeKey) } } total := len(msgOrder) // Return latest messages (tail) start := total - limit - offset if start < 0 { start = 0 } end := total - offset if end < 0 { end = 0 } if end > total { end = total } messages := make([]map[string]interface{}, 0, end-start) for i := start; i < end; i++ { messages = append(messages, msgMap[msgOrder[i]].Data) } return messages, total } // GetAnalyticsChannels returns full channel analytics computed from in-memory packets. func (s *PacketStore) GetAnalyticsChannels(region string) map[string]interface{} { s.cacheMu.Lock() if cached, ok := s.chanCache[region]; ok && time.Now().Before(cached.expiresAt) { s.cacheHits++ s.cacheMu.Unlock() return cached.data } s.cacheMisses++ s.cacheMu.Unlock() result := s.computeAnalyticsChannels(region) s.cacheMu.Lock() s.chanCache[region] = &cachedResult{data: result, expiresAt: time.Now().Add(s.rfCacheTTL)} s.cacheMu.Unlock() return result } func (s *PacketStore) computeAnalyticsChannels(region string) map[string]interface{} { s.mu.RLock() defer s.mu.RUnlock() var regionObs map[string]bool if region != "" { regionObs = s.resolveRegionObservers(region) } type decodedGrp struct { Type string `json:"type"` Channel string `json:"channel"` ChannelHash interface{} `json:"channelHash"` ChannelHash2 string `json:"channel_hash"` Text string `json:"text"` Sender string `json:"sender"` } // Convert channelHash (number or string in JSON) to string chHashStr := func(v interface{}) string { if v == nil { return "" } switch val := v.(type) { case string: return val case float64: return strconv.FormatFloat(val, 'f', -1, 64) default: return fmt.Sprintf("%v", val) } } type chanInfo struct { Hash string Name string Messages int Senders map[string]bool LastActivity string Encrypted bool } channelMap := map[string]*chanInfo{} senderCounts := map[string]int{} msgLengths := make([]int, 0) timeline := map[string]int{} // hour|channelName → count grpTxts := s.byPayloadType[5] for _, tx := range grpTxts { if regionObs != nil { match := false for _, obs := range tx.Observations { if regionObs[obs.ObserverID] { match = true break } } if !match { continue } } var decoded decodedGrp if json.Unmarshal([]byte(tx.DecodedJSON), &decoded) != nil { continue } hash := chHashStr(decoded.ChannelHash) if hash == "" { hash = decoded.ChannelHash2 } if hash == "" { hash = "?" } name := decoded.Channel if name == "" { name = "ch" + hash } encrypted := decoded.Text == "" && decoded.Sender == "" // Use hash as key for grouping (matches Node.js String(hash)) chKey := hash if decoded.Type == "CHAN" && decoded.Channel != "" { chKey = hash + "_" + decoded.Channel } ch := channelMap[chKey] if ch == nil { ch = &chanInfo{Hash: hash, Name: name, Senders: map[string]bool{}, LastActivity: tx.FirstSeen, Encrypted: encrypted} channelMap[chKey] = ch } ch.Messages++ ch.LastActivity = tx.FirstSeen if !encrypted { ch.Encrypted = false } if decoded.Sender != "" { ch.Senders[decoded.Sender] = true senderCounts[decoded.Sender]++ } if decoded.Text != "" { msgLengths = append(msgLengths, len(decoded.Text)) } // Timeline if len(tx.FirstSeen) >= 13 { hr := tx.FirstSeen[:13] key := hr + "|" + name timeline[key]++ } } channelList := make([]map[string]interface{}, 0, len(channelMap)) decryptable := 0 for _, c := range channelMap { if !c.Encrypted { decryptable++ } channelList = append(channelList, map[string]interface{}{ "hash": c.Hash, "name": c.Name, "messages": c.Messages, "senders": len(c.Senders), "lastActivity": c.LastActivity, "encrypted": c.Encrypted, }) } sort.Slice(channelList, func(i, j int) bool { return channelList[i]["messages"].(int) > channelList[j]["messages"].(int) }) // Top senders type senderEntry struct { name string count int } senderList := make([]senderEntry, 0, len(senderCounts)) for n, c := range senderCounts { senderList = append(senderList, senderEntry{n, c}) } sort.Slice(senderList, func(i, j int) bool { return senderList[i].count > senderList[j].count }) topSenders := make([]map[string]interface{}, 0) for i, e := range senderList { if i >= 15 { break } topSenders = append(topSenders, map[string]interface{}{"name": e.name, "count": e.count}) } // Channel timeline type tlEntry struct { hour, channel string count int } var tlList []tlEntry for key, count := range timeline { parts := strings.SplitN(key, "|", 2) if len(parts) == 2 { tlList = append(tlList, tlEntry{parts[0], parts[1], count}) } } sort.Slice(tlList, func(i, j int) bool { return tlList[i].hour < tlList[j].hour }) channelTimeline := make([]map[string]interface{}, 0, len(tlList)) for _, e := range tlList { channelTimeline = append(channelTimeline, map[string]interface{}{ "hour": e.hour, "channel": e.channel, "count": e.count, }) } return map[string]interface{}{ "activeChannels": len(channelList), "decryptable": decryptable, "channels": channelList, "topSenders": topSenders, "channelTimeline": channelTimeline, "msgLengths": msgLengths, } } // GetAnalyticsRF returns full RF analytics computed from in-memory observations. func (s *PacketStore) GetAnalyticsRF(region string) map[string]interface{} { s.cacheMu.Lock() if cached, ok := s.rfCache[region]; ok && time.Now().Before(cached.expiresAt) { s.cacheHits++ s.cacheMu.Unlock() return cached.data } s.cacheMisses++ s.cacheMu.Unlock() result := s.computeAnalyticsRF(region) s.cacheMu.Lock() s.rfCache[region] = &cachedResult{data: result, expiresAt: time.Now().Add(s.rfCacheTTL)} s.cacheMu.Unlock() return result } func (s *PacketStore) computeAnalyticsRF(region string) map[string]interface{} { s.mu.RLock() defer s.mu.RUnlock() ptNames := payloadTypeNames var regionObs map[string]bool if region != "" { regionObs = s.resolveRegionObservers(region) } // Collect all observations matching the region estCap := s.totalObs if estCap > 2000000 { estCap = 2000000 } snrVals := make([]float64, 0, estCap/2) rssiVals := make([]float64, 0, estCap/2) packetSizes := make([]int, 0, len(s.packets)) seenSizeHashes := make(map[string]bool, len(s.packets)) seenTypeHashes := make(map[string]bool, len(s.packets)) typeBuckets := map[int]int{} hourBuckets := map[string]int{} snrByType := map[string]*struct{ vals []float64 }{} sigTime := map[string]*struct { snrs []float64 count int }{} scatterAll := make([]struct{ snr, rssi float64 }, 0, estCap/4) totalObs := 0 regionalHashes := make(map[string]bool, len(s.packets)) var minTimestamp, maxTimestamp string if regionObs != nil { // Regional: iterate observations from matching observers for obsID := range regionObs { obsList := s.byObserver[obsID] for _, obs := range obsList { totalObs++ tx := s.byTxID[obs.TransmissionID] hash := "" if tx != nil { hash = tx.Hash } if hash != "" { regionalHashes[hash] = true } ts := obs.Timestamp if ts != "" { if minTimestamp == "" || ts < minTimestamp { minTimestamp = ts } if ts > maxTimestamp { maxTimestamp = ts } } // SNR/RSSI if obs.SNR != nil { snrVals = append(snrVals, *obs.SNR) typeName := "UNK" if tx != nil && tx.PayloadType != nil { if n, ok := ptNames[*tx.PayloadType]; ok { typeName = n } else { typeName = fmt.Sprintf("UNK(%d)", *tx.PayloadType) } } if snrByType[typeName] == nil { snrByType[typeName] = &struct{ vals []float64 }{} } snrByType[typeName].vals = append(snrByType[typeName].vals, *obs.SNR) if obs.RSSI != nil { scatterAll = append(scatterAll, struct{ snr, rssi float64 }{*obs.SNR, *obs.RSSI}) } // Signal over time if len(ts) >= 13 { hr := ts[:13] if sigTime[hr] == nil { sigTime[hr] = &struct { snrs []float64 count int }{} } sigTime[hr].snrs = append(sigTime[hr].snrs, *obs.SNR) sigTime[hr].count++ } } if obs.RSSI != nil { rssiVals = append(rssiVals, *obs.RSSI) } // Packets per hour if len(ts) >= 13 { hr := ts[:13] hourBuckets[hr]++ } // Packet sizes (unique by hash) if hash != "" && !seenSizeHashes[hash] && tx != nil && tx.RawHex != "" { seenSizeHashes[hash] = true packetSizes = append(packetSizes, len(tx.RawHex)/2) } // Payload type distribution (unique by hash) if hash != "" && !seenTypeHashes[hash] && tx != nil && tx.PayloadType != nil { seenTypeHashes[hash] = true typeBuckets[*tx.PayloadType]++ } } } } else { // No region: iterate all transmissions and their observations for _, tx := range s.packets { hash := tx.Hash if hash != "" { regionalHashes[hash] = true if !seenSizeHashes[hash] && tx.RawHex != "" { seenSizeHashes[hash] = true packetSizes = append(packetSizes, len(tx.RawHex)/2) } if !seenTypeHashes[hash] && tx.PayloadType != nil { seenTypeHashes[hash] = true typeBuckets[*tx.PayloadType]++ } } // Pre-resolve type name once per transmission typeName := "UNK" if tx.PayloadType != nil { if n, ok := ptNames[*tx.PayloadType]; ok { typeName = n } else { typeName = fmt.Sprintf("UNK(%d)", *tx.PayloadType) } } if len(tx.Observations) > 0 { for _, obs := range tx.Observations { totalObs++ ts := obs.Timestamp if ts != "" { if minTimestamp == "" || ts < minTimestamp { minTimestamp = ts } if ts > maxTimestamp { maxTimestamp = ts } } if obs.SNR != nil { snr := *obs.SNR snrVals = append(snrVals, snr) entry := snrByType[typeName] if entry == nil { entry = &struct{ vals []float64 }{} snrByType[typeName] = entry } entry.vals = append(entry.vals, snr) if obs.RSSI != nil { scatterAll = append(scatterAll, struct{ snr, rssi float64 }{snr, *obs.RSSI}) } if len(ts) >= 13 { hr := ts[:13] st := sigTime[hr] if st == nil { st = &struct { snrs []float64 count int }{} sigTime[hr] = st } st.snrs = append(st.snrs, snr) st.count++ } } if obs.RSSI != nil { rssiVals = append(rssiVals, *obs.RSSI) } if len(ts) >= 13 { hourBuckets[ts[:13]]++ } } } else { // Legacy: transmission without observations totalObs++ if tx.SNR != nil { snrVals = append(snrVals, *tx.SNR) } if tx.RSSI != nil { rssiVals = append(rssiVals, *tx.RSSI) } ts := tx.FirstSeen if ts != "" { if minTimestamp == "" || ts < minTimestamp { minTimestamp = ts } if ts > maxTimestamp { maxTimestamp = ts } } if len(ts) >= 13 { hourBuckets[ts[:13]]++ } } } } // Stats helpers sortedF64 := func(arr []float64) []float64 { c := make([]float64, len(arr)) copy(c, arr) sort.Float64s(c) return c } medianF64 := func(arr []float64) float64 { s := sortedF64(arr) if len(s) == 0 { return 0 } return s[len(s)/2] } stddevF64 := func(arr []float64, avg float64) float64 { if len(arr) == 0 { return 0 } sum := 0.0 for _, v := range arr { d := v - avg sum += d * d } return math.Sqrt(sum / float64(len(arr))) } minF64 := func(arr []float64) float64 { if len(arr) == 0 { return 0 } m := arr[0] for _, v := range arr[1:] { if v < m { m = v } } return m } maxF64 := func(arr []float64) float64 { if len(arr) == 0 { return 0 } m := arr[0] for _, v := range arr[1:] { if v > m { m = v } } return m } minInt := func(arr []int) int { if len(arr) == 0 { return 0 } m := arr[0] for _, v := range arr[1:] { if v < m { m = v } } return m } maxInt := func(arr []int) int { if len(arr) == 0 { return 0 } m := arr[0] for _, v := range arr[1:] { if v > m { m = v } } return m } snrAvg := 0.0 if len(snrVals) > 0 { sum := 0.0 for _, v := range snrVals { sum += v } snrAvg = sum / float64(len(snrVals)) } rssiAvg := 0.0 if len(rssiVals) > 0 { sum := 0.0 for _, v := range rssiVals { sum += v } rssiAvg = sum / float64(len(rssiVals)) } // Packets per hour type hourCount struct { Hour string `json:"hour"` Count int `json:"count"` } hourKeys := make([]string, 0, len(hourBuckets)) for k := range hourBuckets { hourKeys = append(hourKeys, k) } sort.Strings(hourKeys) packetsPerHour := make([]hourCount, len(hourKeys)) for i, k := range hourKeys { packetsPerHour[i] = hourCount{Hour: k, Count: hourBuckets[k]} } // Payload types type ptEntry struct { Type int `json:"type"` Name string `json:"name"` Count int `json:"count"` } payloadTypes := make([]ptEntry, 0, len(typeBuckets)) for t, c := range typeBuckets { name := ptNames[t] if name == "" { name = fmt.Sprintf("UNK(%d)", t) } payloadTypes = append(payloadTypes, ptEntry{Type: t, Name: name, Count: c}) } sort.Slice(payloadTypes, func(i, j int) bool { return payloadTypes[i].Count > payloadTypes[j].Count }) // SNR by type type snrTypeEntry struct { Name string `json:"name"` Count int `json:"count"` Avg float64 `json:"avg"` Min float64 `json:"min"` Max float64 `json:"max"` } snrByTypeArr := make([]snrTypeEntry, 0, len(snrByType)) for name, d := range snrByType { sum := 0.0 for _, v := range d.vals { sum += v } snrByTypeArr = append(snrByTypeArr, snrTypeEntry{ Name: name, Count: len(d.vals), Avg: sum / float64(len(d.vals)), Min: minF64(d.vals), Max: maxF64(d.vals), }) } sort.Slice(snrByTypeArr, func(i, j int) bool { return snrByTypeArr[i].Count > snrByTypeArr[j].Count }) // Signal over time type sigTimeEntry struct { Hour string `json:"hour"` Count int `json:"count"` AvgSnr float64 `json:"avgSnr"` } sigKeys := make([]string, 0, len(sigTime)) for k := range sigTime { sigKeys = append(sigKeys, k) } sort.Strings(sigKeys) signalOverTime := make([]sigTimeEntry, len(sigKeys)) for i, k := range sigKeys { d := sigTime[k] sum := 0.0 for _, v := range d.snrs { sum += v } signalOverTime[i] = sigTimeEntry{Hour: k, Count: d.count, AvgSnr: sum / float64(d.count)} } // Scatter (downsample to 500) type scatterPoint struct { SNR float64 `json:"snr"` RSSI float64 `json:"rssi"` } scatterStep := 1 if len(scatterAll) > 500 { scatterStep = len(scatterAll) / 500 } scatterData := make([]scatterPoint, 0, 500) for i, p := range scatterAll { if i%scatterStep == 0 { scatterData = append(scatterData, scatterPoint{SNR: p.snr, RSSI: p.rssi}) } } // Histograms buildHistogramF64 := func(values []float64, bins int) map[string]interface{} { if len(values) == 0 { return map[string]interface{}{"bins": []interface{}{}, "min": 0, "max": 0} } mn, mx := minF64(values), maxF64(values) rng := mx - mn if rng == 0 { rng = 1 } binWidth := rng / float64(bins) counts := make([]int, bins) for _, v := range values { idx := int((v - mn) / binWidth) if idx >= bins { idx = bins - 1 } counts[idx]++ } binArr := make([]map[string]interface{}, bins) for i, c := range counts { binArr[i] = map[string]interface{}{"x": mn + float64(i)*binWidth, "w": binWidth, "count": c} } return map[string]interface{}{"bins": binArr, "min": mn, "max": mx} } buildHistogramInt := func(values []int, bins int) map[string]interface{} { if len(values) == 0 { return map[string]interface{}{"bins": []interface{}{}, "min": 0, "max": 0} } mn, mx := float64(minInt(values)), float64(maxInt(values)) rng := mx - mn if rng == 0 { rng = 1 } binWidth := rng / float64(bins) counts := make([]int, bins) for _, v := range values { idx := int((float64(v) - mn) / binWidth) if idx >= bins { idx = bins - 1 } counts[idx]++ } binArr := make([]map[string]interface{}, bins) for i, c := range counts { binArr[i] = map[string]interface{}{"x": mn + float64(i)*binWidth, "w": binWidth, "count": c} } return map[string]interface{}{"bins": binArr, "min": mn, "max": mx} } snrHistogram := buildHistogramF64(snrVals, 20) rssiHistogram := buildHistogramF64(rssiVals, 20) sizeHistogram := buildHistogramInt(packetSizes, 25) // Time span from min/max timestamps tracked during first pass timeSpanHours := 0.0 if minTimestamp != "" && maxTimestamp != "" && minTimestamp != maxTimestamp { // Parse only 2 timestamps instead of 1.2M parseTS := func(ts string) (time.Time, bool) { t, err := time.Parse("2006-01-02 15:04:05", ts) if err != nil { t, err = time.Parse(time.RFC3339, ts) } if err != nil { return time.Time{}, false } return t, true } if tMin, ok := parseTS(minTimestamp); ok { if tMax, ok := parseTS(maxTimestamp); ok { timeSpanHours = float64(tMax.UnixMilli()-tMin.UnixMilli()) / 3600000.0 } } } // Avg packet size avgPktSize := 0 if len(packetSizes) > 0 { sum := 0 for _, v := range packetSizes { sum += v } avgPktSize = sum / len(packetSizes) } snrStats := map[string]interface{}{"min": 0.0, "max": 0.0, "avg": 0.0, "median": 0.0, "stddev": 0.0} if len(snrVals) > 0 { snrStats = map[string]interface{}{ "min": minF64(snrVals), "max": maxF64(snrVals), "avg": snrAvg, "median": medianF64(snrVals), "stddev": stddevF64(snrVals, snrAvg), } } rssiStats := map[string]interface{}{"min": 0.0, "max": 0.0, "avg": 0.0, "median": 0.0, "stddev": 0.0} if len(rssiVals) > 0 { rssiStats = map[string]interface{}{ "min": minF64(rssiVals), "max": maxF64(rssiVals), "avg": rssiAvg, "median": medianF64(rssiVals), "stddev": stddevF64(rssiVals, rssiAvg), } } return map[string]interface{}{ "totalPackets": len(snrVals), "totalAllPackets": totalObs, "totalTransmissions": len(regionalHashes), "snr": snrStats, "rssi": rssiStats, "snrValues": snrHistogram, "rssiValues": rssiHistogram, "packetSizes": sizeHistogram, "minPacketSize": minInt(packetSizes), "maxPacketSize": maxInt(packetSizes), "avgPacketSize": avgPktSize, "packetsPerHour": packetsPerHour, "payloadTypes": payloadTypes, "snrByType": snrByTypeArr, "signalOverTime": signalOverTime, "scatterData": scatterData, "timeSpanHours": timeSpanHours, } } // --- Topology Analytics --- type nodeInfo struct { PublicKey string Name string Role string Lat float64 Lon float64 HasGPS bool } func (s *PacketStore) getAllNodes() []nodeInfo { rows, err := s.db.conn.Query("SELECT public_key, name, role, lat, lon FROM nodes") if err != nil { return nil } defer rows.Close() var nodes []nodeInfo for rows.Next() { var pk string var name, role sql.NullString var lat, lon sql.NullFloat64 rows.Scan(&pk, &name, &role, &lat, &lon) n := nodeInfo{PublicKey: pk, Name: nullStrVal(name), Role: nullStrVal(role)} if lat.Valid && lon.Valid { n.Lat = lat.Float64 n.Lon = lon.Float64 n.HasGPS = !(n.Lat == 0 && n.Lon == 0) } nodes = append(nodes, n) } return nodes } type prefixMap struct { m map[string][]nodeInfo } func buildPrefixMap(nodes []nodeInfo) *prefixMap { pm := &prefixMap{m: make(map[string][]nodeInfo, len(nodes)*10)} for _, n := range nodes { pk := strings.ToLower(n.PublicKey) for l := 2; l <= len(pk); l++ { pfx := pk[:l] pm.m[pfx] = append(pm.m[pfx], n) } } return pm } // getCachedNodesAndPM returns cached node list and prefix map, rebuilding if stale. // Must be called with s.mu held (RLock or Lock). func (s *PacketStore) getCachedNodesAndPM() ([]nodeInfo, *prefixMap) { s.cacheMu.Lock() if s.nodeCache != nil && time.Since(s.nodeCacheTime) < 30*time.Second { nodes, pm := s.nodeCache, s.nodePM s.cacheMu.Unlock() return nodes, pm } s.cacheMu.Unlock() nodes := s.getAllNodes() pm := buildPrefixMap(nodes) s.cacheMu.Lock() s.nodeCache = nodes s.nodePM = pm s.nodeCacheTime = time.Now() s.cacheMu.Unlock() return nodes, pm } func (pm *prefixMap) resolve(hop string) *nodeInfo { h := strings.ToLower(hop) candidates := pm.m[h] if len(candidates) == 0 { return nil } if len(candidates) == 1 { return &candidates[0] } // Multiple candidates: prefer one with GPS for i := range candidates { if candidates[i].HasGPS { return &candidates[i] } } return &candidates[0] } func parsePathJSON(pathJSON string) []string { if pathJSON == "" || pathJSON == "[]" { return nil } var hops []string if json.Unmarshal([]byte(pathJSON), &hops) != nil { return nil } return hops } func (s *PacketStore) GetAnalyticsTopology(region string) map[string]interface{} { s.cacheMu.Lock() if cached, ok := s.topoCache[region]; ok && time.Now().Before(cached.expiresAt) { s.cacheHits++ s.cacheMu.Unlock() return cached.data } s.cacheMisses++ s.cacheMu.Unlock() result := s.computeAnalyticsTopology(region) s.cacheMu.Lock() s.topoCache[region] = &cachedResult{data: result, expiresAt: time.Now().Add(s.rfCacheTTL)} s.cacheMu.Unlock() return result } func (s *PacketStore) computeAnalyticsTopology(region string) map[string]interface{} { s.mu.RLock() defer s.mu.RUnlock() var regionObs map[string]bool if region != "" { regionObs = s.resolveRegionObservers(region) } allNodes, pm := s.getCachedNodesAndPM() _ = allNodes // only pm is needed for topology hopCache := make(map[string]*nodeInfo) resolveHop := func(hop string) *nodeInfo { if cached, ok := hopCache[hop]; ok { return cached } r := pm.resolve(hop) hopCache[hop] = r return r } hopCounts := map[int]int{} var allHopsList []int hopSnr := map[int][]float64{} hopFreq := map[string]int{} pairFreq := map[string]int{} observerMap := map[string]string{} // observer_id → observer_name perObserver := map[string]map[string]*struct{ minDist, maxDist, count int }{} for _, tx := range s.packets { hops := txGetParsedPath(tx) if len(hops) == 0 { continue } if regionObs != nil { match := false for _, obs := range tx.Observations { if regionObs[obs.ObserverID] { match = true break } } if !match { continue } } n := len(hops) hopCounts[n]++ allHopsList = append(allHopsList, n) if tx.SNR != nil { hopSnr[n] = append(hopSnr[n], *tx.SNR) } for _, h := range hops { hopFreq[h]++ } for i := 0; i < len(hops)-1; i++ { a, b := hops[i], hops[i+1] if a > b { a, b = b, a } pairFreq[a+"|"+b]++ } obsID := tx.ObserverID if obsID != "" { observerMap[obsID] = tx.ObserverName } if _, ok := perObserver[obsID]; !ok { perObserver[obsID] = map[string]*struct{ minDist, maxDist, count int }{} } for i, h := range hops { dist := n - i entry := perObserver[obsID][h] if entry == nil { entry = &struct{ minDist, maxDist, count int }{dist, dist, 0} perObserver[obsID][h] = entry } if dist < entry.minDist { entry.minDist = dist } if dist > entry.maxDist { entry.maxDist = dist } entry.count++ } } // Hop distribution hopDist := make([]map[string]interface{}, 0) for h, c := range hopCounts { if h <= 25 { hopDist = append(hopDist, map[string]interface{}{"hops": h, "count": c}) } } sort.Slice(hopDist, func(i, j int) bool { return hopDist[i]["hops"].(int) < hopDist[j]["hops"].(int) }) avgHops := 0.0 if len(allHopsList) > 0 { sum := 0 for _, v := range allHopsList { sum += v } avgHops = float64(sum) / float64(len(allHopsList)) } medianHops := 0 if len(allHopsList) > 0 { sorted := make([]int, len(allHopsList)) copy(sorted, allHopsList) sort.Ints(sorted) medianHops = sorted[len(sorted)/2] } maxHops := 0 for _, v := range allHopsList { if v > maxHops { maxHops = v } } // Top repeaters type freqEntry struct { hop string count int } freqList := make([]freqEntry, 0, len(hopFreq)) for h, c := range hopFreq { freqList = append(freqList, freqEntry{h, c}) } sort.Slice(freqList, func(i, j int) bool { return freqList[i].count > freqList[j].count }) topRepeaters := make([]map[string]interface{}, 0) for i, e := range freqList { if i >= 20 { break } r := resolveHop(e.hop) entry := map[string]interface{}{"hop": e.hop, "count": e.count, "name": nil, "pubkey": nil} if r != nil { entry["name"] = r.Name entry["pubkey"] = r.PublicKey } topRepeaters = append(topRepeaters, entry) } // Top pairs pairList := make([]freqEntry, 0, len(pairFreq)) for p, c := range pairFreq { pairList = append(pairList, freqEntry{p, c}) } sort.Slice(pairList, func(i, j int) bool { return pairList[i].count > pairList[j].count }) topPairs := make([]map[string]interface{}, 0) for i, e := range pairList { if i >= 15 { break } parts := strings.SplitN(e.hop, "|", 2) rA := resolveHop(parts[0]) rB := resolveHop(parts[1]) entry := map[string]interface{}{ "hopA": parts[0], "hopB": parts[1], "count": e.count, "nameA": nil, "nameB": nil, "pubkeyA": nil, "pubkeyB": nil, } if rA != nil { entry["nameA"] = rA.Name entry["pubkeyA"] = rA.PublicKey } if rB != nil { entry["nameB"] = rB.Name entry["pubkeyB"] = rB.PublicKey } topPairs = append(topPairs, entry) } // Hops vs SNR hopsVsSnr := make([]map[string]interface{}, 0) for h, snrs := range hopSnr { if h > 20 { continue } sum := 0.0 for _, v := range snrs { sum += v } hopsVsSnr = append(hopsVsSnr, map[string]interface{}{ "hops": h, "count": len(snrs), "avgSnr": sum / float64(len(snrs)), }) } sort.Slice(hopsVsSnr, func(i, j int) bool { return hopsVsSnr[i]["hops"].(int) < hopsVsSnr[j]["hops"].(int) }) // Observers list observers := make([]map[string]interface{}, 0) for id, name := range observerMap { n := name if n == "" { n = id } observers = append(observers, map[string]interface{}{"id": id, "name": n}) } // Per-observer reachability perObserverReach := map[string]interface{}{} for obsID, nodes := range perObserver { obsName := observerMap[obsID] if obsName == "" { obsName = obsID } byDist := map[int][]map[string]interface{}{} for hop, data := range nodes { d := data.minDist if d > 15 { continue } r := resolveHop(hop) entry := map[string]interface{}{ "hop": hop, "name": nil, "pubkey": nil, "count": data.count, "distRange": nil, } if r != nil { entry["name"] = r.Name entry["pubkey"] = r.PublicKey } if data.minDist != data.maxDist { entry["distRange"] = fmt.Sprintf("%d-%d", data.minDist, data.maxDist) } byDist[d] = append(byDist[d], entry) } rings := make([]map[string]interface{}, 0) for dist, nodeList := range byDist { sort.Slice(nodeList, func(i, j int) bool { return nodeList[i]["count"].(int) > nodeList[j]["count"].(int) }) rings = append(rings, map[string]interface{}{"hops": dist, "nodes": nodeList}) } sort.Slice(rings, func(i, j int) bool { return rings[i]["hops"].(int) < rings[j]["hops"].(int) }) perObserverReach[obsID] = map[string]interface{}{ "observer_name": obsName, "rings": rings, } } // Cross-observer: build from perObserver crossObserver := map[string][]map[string]interface{}{} bestPath := map[string]map[string]interface{}{} for obsID, nodes := range perObserver { obsName := observerMap[obsID] if obsName == "" { obsName = obsID } for hop, data := range nodes { crossObserver[hop] = append(crossObserver[hop], map[string]interface{}{ "observer_id": obsID, "observer_name": obsName, "minDist": data.minDist, "count": data.count, }) if bp, ok := bestPath[hop]; !ok || data.minDist < bp["minDist"].(int) { bestPath[hop] = map[string]interface{}{ "minDist": data.minDist, "observer_id": obsID, "observer_name": obsName, } } } } // Multi-observer nodes multiObsNodes := make([]map[string]interface{}, 0) for hop, obs := range crossObserver { if len(obs) <= 1 { continue } sort.Slice(obs, func(i, j int) bool { return obs[i]["minDist"].(int) < obs[j]["minDist"].(int) }) r := resolveHop(hop) entry := map[string]interface{}{ "hop": hop, "name": nil, "pubkey": nil, "observers": obs, } if r != nil { entry["name"] = r.Name entry["pubkey"] = r.PublicKey } multiObsNodes = append(multiObsNodes, entry) } sort.Slice(multiObsNodes, func(i, j int) bool { return len(multiObsNodes[i]["observers"].([]map[string]interface{})) > len(multiObsNodes[j]["observers"].([]map[string]interface{})) }) if len(multiObsNodes) > 50 { multiObsNodes = multiObsNodes[:50] } // Best path list bestPathList := make([]map[string]interface{}, 0, len(bestPath)) for hop, data := range bestPath { r := resolveHop(hop) entry := map[string]interface{}{ "hop": hop, "name": nil, "pubkey": nil, "minDist": data["minDist"], "observer_id": data["observer_id"], "observer_name": data["observer_name"], } if r != nil { entry["name"] = r.Name entry["pubkey"] = r.PublicKey } bestPathList = append(bestPathList, entry) } sort.Slice(bestPathList, func(i, j int) bool { return bestPathList[i]["minDist"].(int) < bestPathList[j]["minDist"].(int) }) if len(bestPathList) > 50 { bestPathList = bestPathList[:50] } // Use DB 7-day active node count (matches /api/stats totalNodes) uniqueNodes := 0 if s.db != nil { if stats, err := s.db.GetStats(); err == nil { uniqueNodes = stats.TotalNodes } } return map[string]interface{}{ "uniqueNodes": uniqueNodes, "avgHops": avgHops, "medianHops": medianHops, "maxHops": maxHops, "hopDistribution": hopDist, "topRepeaters": topRepeaters, "topPairs": topPairs, "hopsVsSnr": hopsVsSnr, "observers": observers, "perObserverReach": perObserverReach, "multiObsNodes": multiObsNodes, "bestPathList": bestPathList, } } // --- Distance Analytics --- func haversineKm(lat1, lon1, lat2, lon2 float64) float64 { const R = 6371.0 dLat := (lat2 - lat1) * math.Pi / 180 dLon := (lon2 - lon1) * math.Pi / 180 a := math.Sin(dLat/2)*math.Sin(dLat/2) + math.Cos(lat1*math.Pi/180)*math.Cos(lat2*math.Pi/180)* math.Sin(dLon/2)*math.Sin(dLon/2) return R * 2 * math.Atan2(math.Sqrt(a), math.Sqrt(1-a)) } func (s *PacketStore) GetAnalyticsDistance(region string) map[string]interface{} { s.mu.RLock() defer s.mu.RUnlock() var regionObs map[string]bool if region != "" { regionObs = s.resolveRegionObservers(region) } allNodes, pm := s.getCachedNodesAndPM() hopCache := make(map[string]*nodeInfo) resolveHop := func(hop string) *nodeInfo { if cached, ok := hopCache[hop]; ok { return cached } r := pm.resolve(hop) hopCache[hop] = r return r } repeaterSet := make(map[string]bool) nodeByPk := make(map[string]*nodeInfo, len(allNodes)) for i := range allNodes { n := &allNodes[i] nodeByPk[n.PublicKey] = n if strings.Contains(strings.ToLower(n.Role), "repeater") { repeaterSet[n.PublicKey] = true } } type hopRecord struct { FromName, FromPk, ToName, ToPk string Dist float64 Type string SNR interface{} Hash, Timestamp string } type pathRecord struct { Hash string TotalDist float64 HopCount int Timestamp string Hops []map[string]interface{} } var allHops []hopRecord var pathTotals []pathRecord catDists := map[string][]float64{"R↔R": {}, "C↔R": {}, "C↔C": {}} distByHour := map[string][]float64{} for _, tx := range s.packets { hops := txGetParsedPath(tx) if len(hops) == 0 { continue } if regionObs != nil { match := false for _, obs := range tx.Observations { if regionObs[obs.ObserverID] { match = true break } } if !match { continue } } // Resolve all hops to nodes resolved := make([]*nodeInfo, len(hops)) for i, h := range hops { resolved[i] = resolveHop(h) } // Resolve sender from decoded_json var senderNode *nodeInfo if tx.DecodedJSON != "" { var dec map[string]interface{} if json.Unmarshal([]byte(tx.DecodedJSON), &dec) == nil { if pk, ok := dec["pubKey"].(string); ok && pk != "" { senderNode = nodeByPk[pk] } } } // Build chain of GPS-located nodes chain := make([]*nodeInfo, 0, len(hops)+1) if senderNode != nil && senderNode.HasGPS { chain = append(chain, senderNode) } for _, r := range resolved { if r != nil && r.HasGPS { chain = append(chain, r) } } if len(chain) < 2 { continue } hourBucket := "" if tx.FirstSeen != "" { if len(tx.FirstSeen) >= 13 { hourBucket = tx.FirstSeen[:13] } } pathDist := 0.0 var pathHops []map[string]interface{} for i := 0; i < len(chain)-1; i++ { a, b := chain[i], chain[i+1] dist := haversineKm(a.Lat, a.Lon, b.Lat, b.Lon) if dist > 300 { continue } aRep := repeaterSet[a.PublicKey] bRep := repeaterSet[b.PublicKey] var hopType string if aRep && bRep { hopType = "R↔R" } else if !aRep && !bRep { hopType = "C↔C" } else { hopType = "C↔R" } roundedDist := math.Round(dist*100) / 100 var snrVal interface{} if tx.SNR != nil { snrVal = *tx.SNR } allHops = append(allHops, hopRecord{ FromName: a.Name, FromPk: a.PublicKey, ToName: b.Name, ToPk: b.PublicKey, Dist: roundedDist, Type: hopType, SNR: snrVal, Hash: tx.Hash, Timestamp: tx.FirstSeen, }) catDists[hopType] = append(catDists[hopType], dist) pathDist += dist pathHops = append(pathHops, map[string]interface{}{ "fromName": a.Name, "fromPk": a.PublicKey, "toName": b.Name, "toPk": b.PublicKey, "dist": roundedDist, }) if hourBucket != "" { distByHour[hourBucket] = append(distByHour[hourBucket], dist) } } if len(pathHops) > 0 { pathTotals = append(pathTotals, pathRecord{ Hash: tx.Hash, TotalDist: math.Round(pathDist*100) / 100, HopCount: len(pathHops), Timestamp: tx.FirstSeen, Hops: pathHops, }) } } // Sort and pick top hops/paths sort.Slice(allHops, func(i, j int) bool { return allHops[i].Dist > allHops[j].Dist }) topHops := make([]map[string]interface{}, 0) for i, h := range allHops { if i >= 50 { break } topHops = append(topHops, map[string]interface{}{ "fromName": h.FromName, "fromPk": h.FromPk, "toName": h.ToName, "toPk": h.ToPk, "dist": h.Dist, "type": h.Type, "snr": h.SNR, "hash": h.Hash, "timestamp": h.Timestamp, }) } sort.Slice(pathTotals, func(i, j int) bool { return pathTotals[i].TotalDist > pathTotals[j].TotalDist }) topPaths := make([]map[string]interface{}, 0) for i, p := range pathTotals { if i >= 20 { break } topPaths = append(topPaths, map[string]interface{}{ "hash": p.Hash, "totalDist": p.TotalDist, "hopCount": p.HopCount, "timestamp": p.Timestamp, "hops": p.Hops, }) } // Category stats medianF := func(arr []float64) float64 { if len(arr) == 0 { return 0 } c := make([]float64, len(arr)) copy(c, arr) sort.Float64s(c) return c[len(c)/2] } minF := func(arr []float64) float64 { if len(arr) == 0 { return 0 } m := arr[0] for _, v := range arr[1:] { if v < m { m = v } } return m } maxF := func(arr []float64) float64 { if len(arr) == 0 { return 0 } m := arr[0] for _, v := range arr[1:] { if v > m { m = v } } return m } catStats := map[string]interface{}{} for cat, dists := range catDists { if len(dists) == 0 { catStats[cat] = map[string]interface{}{"count": 0, "avg": 0, "median": 0, "min": 0, "max": 0} continue } sum := 0.0 for _, v := range dists { sum += v } avg := sum / float64(len(dists)) catStats[cat] = map[string]interface{}{ "count": len(dists), "avg": math.Round(avg*100) / 100, "median": math.Round(medianF(dists)*100) / 100, "min": math.Round(minF(dists)*100) / 100, "max": math.Round(maxF(dists)*100) / 100, } } // Distance histogram var distHistogram interface{} = []interface{}{} allDists := make([]float64, len(allHops)) for i, h := range allHops { allDists[i] = h.Dist } if len(allDists) > 0 { hMin, hMax := minF(allDists), maxF(allDists) binCount := 25 binW := (hMax - hMin) / float64(binCount) if binW == 0 { binW = 1 } bins := make([]int, binCount) for _, d := range allDists { idx := int(math.Floor((d - hMin) / binW)) if idx >= binCount { idx = binCount - 1 } if idx < 0 { idx = 0 } bins[idx]++ } binArr := make([]map[string]interface{}, binCount) for i, c := range bins { binArr[i] = map[string]interface{}{ "x": math.Round((hMin+float64(i)*binW)*10) / 10, "w": math.Round(binW*10) / 10, "count": c, } } distHistogram = map[string]interface{}{"bins": binArr, "min": hMin, "max": hMax} } // Distance over time timeKeys := make([]string, 0, len(distByHour)) for k := range distByHour { timeKeys = append(timeKeys, k) } sort.Strings(timeKeys) distOverTime := make([]map[string]interface{}, 0, len(timeKeys)) for _, hour := range timeKeys { dists := distByHour[hour] sum := 0.0 for _, v := range dists { sum += v } distOverTime = append(distOverTime, map[string]interface{}{ "hour": hour, "avg": math.Round(sum/float64(len(dists))*100) / 100, "count": len(dists), }) } // Summary summary := map[string]interface{}{ "totalHops": len(allHops), "totalPaths": len(pathTotals), "avgDist": 0.0, "maxDist": 0.0, } if len(allDists) > 0 { sum := 0.0 for _, v := range allDists { sum += v } summary["avgDist"] = math.Round(sum/float64(len(allDists))*100) / 100 summary["maxDist"] = math.Round(maxF(allDists)*100) / 100 } return map[string]interface{}{ "summary": summary, "topHops": topHops, "topPaths": topPaths, "catStats": catStats, "distHistogram": distHistogram, "distOverTime": distOverTime, } } // --- Hash Sizes Analytics --- func (s *PacketStore) GetAnalyticsHashSizes(region string) map[string]interface{} { s.cacheMu.Lock() if cached, ok := s.hashCache[region]; ok && time.Now().Before(cached.expiresAt) { s.cacheHits++ s.cacheMu.Unlock() return cached.data } s.cacheMisses++ s.cacheMu.Unlock() result := s.computeAnalyticsHashSizes(region) s.cacheMu.Lock() s.hashCache[region] = &cachedResult{data: result, expiresAt: time.Now().Add(s.rfCacheTTL)} s.cacheMu.Unlock() return result } func (s *PacketStore) computeAnalyticsHashSizes(region string) map[string]interface{} { s.mu.RLock() defer s.mu.RUnlock() var regionObs map[string]bool if region != "" { regionObs = s.resolveRegionObservers(region) } _, pm := s.getCachedNodesAndPM() distribution := map[string]int{"1": 0, "2": 0, "3": 0} byHour := map[string]map[string]int{} byNode := map[string]map[string]interface{}{} uniqueHops := map[string]map[string]interface{}{} total := 0 for _, tx := range s.packets { if tx.RawHex == "" { continue } hops := txGetParsedPath(tx) if len(hops) == 0 { continue } if regionObs != nil { match := false for _, obs := range tx.Observations { if regionObs[obs.ObserverID] { match = true break } } if !match { continue } } // Parse header and path byte if len(tx.RawHex) < 4 { continue } header, err := strconv.ParseUint(tx.RawHex[:2], 16, 8) if err != nil { continue } routeType := header & 0x03 pathByteIdx := 1 if routeType == 0 || routeType == 3 { pathByteIdx = 5 } hexStart := pathByteIdx * 2 hexEnd := hexStart + 2 if hexEnd > len(tx.RawHex) { continue } actualPathByte, err := strconv.ParseUint(tx.RawHex[hexStart:hexEnd], 16, 8) if err != nil { continue } hashSize := int((actualPathByte>>6)&0x3) + 1 if hashSize > 3 { continue } total++ sizeKey := strconv.Itoa(hashSize) distribution[sizeKey]++ // Hourly buckets if len(tx.FirstSeen) >= 13 { hour := tx.FirstSeen[:13] if byHour[hour] == nil { byHour[hour] = map[string]int{"1": 0, "2": 0, "3": 0} } byHour[hour][sizeKey]++ } // Track unique hops with their sizes for _, hop := range hops { if uniqueHops[hop] == nil { hopLower := strings.ToLower(hop) candidates := pm.m[hopLower] var matchName, matchPk interface{} if len(candidates) > 0 { matchName = candidates[0].Name matchPk = candidates[0].PublicKey } uniqueHops[hop] = map[string]interface{}{ "size": (len(hop) + 1) / 2, "count": 0, "name": matchName, "pubkey": matchPk, } } uniqueHops[hop]["count"] = uniqueHops[hop]["count"].(int) + 1 } // Track originator from advert packets if tx.PayloadType != nil && *tx.PayloadType == 4 && tx.DecodedJSON != "" { var d map[string]interface{} if json.Unmarshal([]byte(tx.DecodedJSON), &d) == nil { name := "" if n, ok := d["name"].(string); ok { name = n } if name == "" { if pk, ok := d["pubKey"].(string); ok && pk != "" { name = pk[:8] } else if pk, ok := d["public_key"].(string); ok && pk != "" { name = pk[:8] } } if name != "" { if byNode[name] == nil { var pubkey interface{} if pk, ok := d["pubKey"].(string); ok { pubkey = pk } else if pk, ok := d["public_key"].(string); ok { pubkey = pk } byNode[name] = map[string]interface{}{ "hashSize": hashSize, "packets": 0, "lastSeen": tx.FirstSeen, "pubkey": pubkey, } } byNode[name]["packets"] = byNode[name]["packets"].(int) + 1 byNode[name]["hashSize"] = hashSize byNode[name]["lastSeen"] = tx.FirstSeen } } } } // Sort hourly data hourKeys := make([]string, 0, len(byHour)) for k := range byHour { hourKeys = append(hourKeys, k) } sort.Strings(hourKeys) hourly := make([]map[string]interface{}, 0, len(hourKeys)) for _, hour := range hourKeys { sizes := byHour[hour] hourly = append(hourly, map[string]interface{}{ "hour": hour, "1": sizes["1"], "2": sizes["2"], "3": sizes["3"], }) } // Top hops by frequency type hopEntry struct { hex string data map[string]interface{} } hopList := make([]hopEntry, 0, len(uniqueHops)) for hex, data := range uniqueHops { hopList = append(hopList, hopEntry{hex, data}) } sort.Slice(hopList, func(i, j int) bool { return hopList[i].data["count"].(int) > hopList[j].data["count"].(int) }) topHops := make([]map[string]interface{}, 0) for i, e := range hopList { if i >= 50 { break } topHops = append(topHops, map[string]interface{}{ "hex": e.hex, "size": e.data["size"], "count": e.data["count"], "name": e.data["name"], "pubkey": e.data["pubkey"], }) } // Multi-byte nodes multiByteNodes := make([]map[string]interface{}, 0) for name, data := range byNode { if data["hashSize"].(int) > 1 { multiByteNodes = append(multiByteNodes, map[string]interface{}{ "name": name, "hashSize": data["hashSize"], "packets": data["packets"], "lastSeen": data["lastSeen"], "pubkey": data["pubkey"], }) } } sort.Slice(multiByteNodes, func(i, j int) bool { return multiByteNodes[i]["packets"].(int) > multiByteNodes[j]["packets"].(int) }) return map[string]interface{}{ "total": total, "distribution": distribution, "hourly": hourly, "topHops": topHops, "multiByteNodes": multiByteNodes, } } // hashSizeNodeInfo holds per-node hash size tracking data. type hashSizeNodeInfo struct { HashSize int AllSizes map[int]bool Seq []int Inconsistent bool } // GetNodeHashSizeInfo scans advert packets to compute per-node hash size data. func (s *PacketStore) GetNodeHashSizeInfo() map[string]*hashSizeNodeInfo { s.mu.RLock() defer s.mu.RUnlock() info := make(map[string]*hashSizeNodeInfo) adverts := s.byPayloadType[4] for _, tx := range adverts { if tx.RawHex == "" || tx.DecodedJSON == "" { continue } if len(tx.RawHex) < 4 { continue } pathByte, err := strconv.ParseUint(tx.RawHex[2:4], 16, 8) if err != nil { continue } hs := int((pathByte>>6)&0x3) + 1 var d map[string]interface{} if json.Unmarshal([]byte(tx.DecodedJSON), &d) != nil { continue } pk := "" if v, ok := d["pubKey"].(string); ok { pk = v } else if v, ok := d["public_key"].(string); ok { pk = v } if pk == "" { continue } ni := info[pk] if ni == nil { ni = &hashSizeNodeInfo{AllSizes: make(map[int]bool)} info[pk] = ni } ni.HashSize = hs ni.AllSizes[hs] = true ni.Seq = append(ni.Seq, hs) } // Compute flip-flop (inconsistent) flag: need >= 3 observations, // >= 2 unique sizes, and >= 2 transitions in the sequence. for _, ni := range info { if len(ni.Seq) < 3 || len(ni.AllSizes) < 2 { continue } transitions := 0 for i := 1; i < len(ni.Seq); i++ { if ni.Seq[i] != ni.Seq[i-1] { transitions++ } } ni.Inconsistent = transitions >= 2 } return info } // EnrichNodeWithHashSize populates hash_size, hash_size_inconsistent, and // hash_sizes_seen on a node map using precomputed hash size info. func EnrichNodeWithHashSize(node map[string]interface{}, info *hashSizeNodeInfo) { if info == nil { return } node["hash_size"] = info.HashSize node["hash_size_inconsistent"] = info.Inconsistent if len(info.AllSizes) > 1 { sizes := make([]int, 0, len(info.AllSizes)) for s := range info.AllSizes { sizes = append(sizes, s) } sort.Ints(sizes) node["hash_sizes_seen"] = sizes } } // --- Bulk Health (in-memory) --- func (s *PacketStore) GetBulkHealth(limit int, region string) []map[string]interface{} { s.mu.RLock() defer s.mu.RUnlock() // Region filtering var regionNodeKeys map[string]bool if region != "" { regionObs := s.resolveRegionObservers(region) if regionObs != nil { regionalHashes := make(map[string]bool) for obsID := range regionObs { obsList := s.byObserver[obsID] for _, o := range obsList { tx := s.byTxID[o.TransmissionID] if tx != nil { regionalHashes[tx.Hash] = true } } } regionNodeKeys = make(map[string]bool) for pk, hashes := range s.nodeHashes { for h := range hashes { if regionalHashes[h] { regionNodeKeys[pk] = true break } } } } } // Get nodes from DB queryLimit := limit if regionNodeKeys != nil { queryLimit = 500 } rows, err := s.db.conn.Query("SELECT public_key, name, role, lat, lon FROM nodes ORDER BY last_seen DESC LIMIT ?", queryLimit) if err != nil { return []map[string]interface{}{} } defer rows.Close() type dbNode struct { pk, name, role string lat, lon interface{} } var nodes []dbNode for rows.Next() { var pk string var name, role sql.NullString var lat, lon sql.NullFloat64 rows.Scan(&pk, &name, &role, &lat, &lon) if regionNodeKeys != nil && !regionNodeKeys[pk] { continue } nodes = append(nodes, dbNode{ pk: pk, name: nullStrVal(name), role: nullStrVal(role), lat: nullFloat(lat), lon: nullFloat(lon), }) if regionNodeKeys == nil && len(nodes) >= limit { break } } if regionNodeKeys != nil && len(nodes) > limit { nodes = nodes[:limit] } todayStart := time.Now().UTC().Truncate(24 * time.Hour).Format(time.RFC3339) results := make([]map[string]interface{}, 0, len(nodes)) for _, n := range nodes { packets := s.byNode[n.pk] var packetsToday int var snrSum float64 var snrCount int var lastHeard string observerStats := map[string]*struct { name string snrSum, rssiSum float64 snrCount, rssiCount, count int }{} totalObservations := 0 for _, pkt := range packets { totalObservations += pkt.ObservationCount if totalObservations == 0 { totalObservations = 1 } if pkt.FirstSeen > todayStart { packetsToday++ } if pkt.SNR != nil { snrSum += *pkt.SNR snrCount++ } if lastHeard == "" || pkt.FirstSeen > lastHeard { lastHeard = pkt.FirstSeen } obsID := pkt.ObserverID if obsID != "" { obs := observerStats[obsID] if obs == nil { obs = &struct { name string snrSum, rssiSum float64 snrCount, rssiCount, count int }{name: pkt.ObserverName} observerStats[obsID] = obs } obs.count++ if pkt.SNR != nil { obs.snrSum += *pkt.SNR obs.snrCount++ } if pkt.RSSI != nil { obs.rssiSum += *pkt.RSSI obs.rssiCount++ } } } observerRows := make([]map[string]interface{}, 0) for id, o := range observerStats { var avgSnr, avgRssi interface{} if o.snrCount > 0 { avgSnr = o.snrSum / float64(o.snrCount) } if o.rssiCount > 0 { avgRssi = o.rssiSum / float64(o.rssiCount) } observerRows = append(observerRows, map[string]interface{}{ "observer_id": id, "observer_name": o.name, "avgSnr": avgSnr, "avgRssi": avgRssi, "packetCount": o.count, }) } sort.Slice(observerRows, func(i, j int) bool { return observerRows[i]["packetCount"].(int) > observerRows[j]["packetCount"].(int) }) var avgSnr interface{} if snrCount > 0 { avgSnr = snrSum / float64(snrCount) } var lhVal interface{} if lastHeard != "" { lhVal = lastHeard } results = append(results, map[string]interface{}{ "public_key": n.pk, "name": nilIfEmpty(n.name), "role": nilIfEmpty(n.role), "lat": n.lat, "lon": n.lon, "stats": map[string]interface{}{ "totalTransmissions": len(packets), "totalObservations": totalObservations, "totalPackets": len(packets), "packetsToday": packetsToday, "avgSnr": avgSnr, "lastHeard": lhVal, }, "observers": observerRows, }) } return results } // --- Subpaths Analytics --- func (s *PacketStore) GetAnalyticsSubpaths(region string, minLen, maxLen, limit int) map[string]interface{} { s.mu.RLock() defer s.mu.RUnlock() var regionObs map[string]bool if region != "" { regionObs = s.resolveRegionObservers(region) } _, pm := s.getCachedNodesAndPM() hopCache := make(map[string]*nodeInfo) resolveHop := func(hop string) string { if cached, ok := hopCache[hop]; ok { if cached != nil { return cached.Name } return hop } r := pm.resolve(hop) hopCache[hop] = r if r != nil { return r.Name } return hop } subpathCounts := map[string]*struct { count int raw string }{} totalPaths := 0 for _, tx := range s.packets { hops := txGetParsedPath(tx) if len(hops) < 2 { continue } if regionObs != nil { match := false for _, obs := range tx.Observations { if regionObs[obs.ObserverID] { match = true break } } if !match { continue } } totalPaths++ // Resolve hops to names named := make([]string, len(hops)) for i, h := range hops { named[i] = resolveHop(h) } for l := minLen; l <= maxLen && l <= len(named); l++ { for start := 0; start <= len(named)-l; start++ { sub := strings.Join(named[start:start+l], " → ") raw := strings.Join(hops[start:start+l], ",") entry := subpathCounts[sub] if entry == nil { entry = &struct { count int raw string }{raw: raw} subpathCounts[sub] = entry } entry.count++ } } } type subpathEntry struct { path string count int raw string } ranked := make([]subpathEntry, 0, len(subpathCounts)) for path, data := range subpathCounts { ranked = append(ranked, subpathEntry{path, data.count, data.raw}) } sort.Slice(ranked, func(i, j int) bool { return ranked[i].count > ranked[j].count }) if len(ranked) > limit { ranked = ranked[:limit] } subpaths := make([]map[string]interface{}, 0, len(ranked)) for _, e := range ranked { pct := 0.0 if totalPaths > 0 { pct = math.Round(float64(e.count)/float64(totalPaths)*1000) / 10 } subpaths = append(subpaths, map[string]interface{}{ "path": e.path, "rawHops": strings.Split(e.raw, ","), "count": e.count, "hops": len(strings.Split(e.path, " → ")), "pct": pct, }) } return map[string]interface{}{ "subpaths": subpaths, "totalPaths": totalPaths, } } // --- Subpath Detail --- func (s *PacketStore) GetSubpathDetail(rawHops []string) map[string]interface{} { s.mu.RLock() defer s.mu.RUnlock() _, pm := s.getCachedNodesAndPM() // Resolve the requested hops nodes := make([]map[string]interface{}, len(rawHops)) for i, hop := range rawHops { r := pm.resolve(hop) entry := map[string]interface{}{"hop": hop, "name": hop, "lat": nil, "lon": nil, "pubkey": nil} if r != nil { entry["name"] = r.Name entry["pubkey"] = r.PublicKey if r.HasGPS { entry["lat"] = r.Lat entry["lon"] = r.Lon } } nodes[i] = entry } hourBuckets := make([]int, 24) var snrSum, rssiSum float64 var snrCount, rssiCount int observers := map[string]int{} parentPaths := map[string]int{} var matchCount int var firstSeen, lastSeen interface{} for _, tx := range s.packets { hops := txGetParsedPath(tx) if len(hops) < len(rawHops) { continue } // Check if rawHops appears as contiguous subsequence found := false for i := 0; i <= len(hops)-len(rawHops); i++ { match := true for j := 0; j < len(rawHops); j++ { if !strings.EqualFold(hops[i+j], rawHops[j]) { match = false break } } if match { found = true break } } if !found { continue } matchCount++ ts := tx.FirstSeen if ts != "" { if firstSeen == nil || ts < firstSeen.(string) { firstSeen = ts } if lastSeen == nil || ts > lastSeen.(string) { lastSeen = ts } // Parse hour from timestamp for hourly distribution t, err := time.Parse(time.RFC3339, ts) if err != nil { t, err = time.Parse("2006-01-02 15:04:05", ts) } if err == nil { hourBuckets[t.Hour()]++ } } if tx.SNR != nil { snrSum += *tx.SNR snrCount++ } if tx.RSSI != nil { rssiSum += *tx.RSSI rssiCount++ } if tx.ObserverName != "" { observers[tx.ObserverName]++ } // Full parent path (resolved) resolved := make([]string, len(hops)) for i, h := range hops { r := pm.resolve(h) if r != nil { resolved[i] = r.Name } else { resolved[i] = h } } fullPath := strings.Join(resolved, " → ") parentPaths[fullPath]++ } var avgSnr, avgRssi interface{} if snrCount > 0 { avgSnr = snrSum / float64(snrCount) } if rssiCount > 0 { avgRssi = rssiSum / float64(rssiCount) } topParents := make([]map[string]interface{}, 0) for path, count := range parentPaths { topParents = append(topParents, map[string]interface{}{"path": path, "count": count}) } sort.Slice(topParents, func(i, j int) bool { return topParents[i]["count"].(int) > topParents[j]["count"].(int) }) if len(topParents) > 15 { topParents = topParents[:15] } topObs := make([]map[string]interface{}, 0) for name, count := range observers { topObs = append(topObs, map[string]interface{}{"name": name, "count": count}) } sort.Slice(topObs, func(i, j int) bool { return topObs[i]["count"].(int) > topObs[j]["count"].(int) }) if len(topObs) > 10 { topObs = topObs[:10] } return map[string]interface{}{ "hops": rawHops, "nodes": nodes, "totalMatches": matchCount, "firstSeen": firstSeen, "lastSeen": lastSeen, "signal": map[string]interface{}{"avgSnr": avgSnr, "avgRssi": avgRssi, "samples": snrCount}, "hourDistribution": hourBuckets, "parentPaths": topParents, "observers": topObs, } }