Skip to main content
Glama
test_telemetry_queue_worker.py1.87 kB
import types import threading import time import queue as q import core.telemetry as telemetry def test_telemetry_queue_backpressure_and_single_worker(monkeypatch, caplog): caplog.set_level("DEBUG") collector = telemetry.TelemetryCollector() # Force-enable telemetry regardless of env settings from conftest collector.config.enabled = True # Wake existing worker once so it observes the new queue on the next loop collector.record(telemetry.RecordType.TOOL_EXECUTION, {"i": -1}) # Replace queue with tiny one to trigger backpressure quickly small_q = q.Queue(maxsize=2) collector._queue = small_q # Give the worker a moment to switch queues time.sleep(0.02) # Make sends slow to build backlog and exercise worker def slow_send(self, rec): time.sleep(0.05) collector._send_telemetry = types.MethodType(slow_send, collector) # Fire many events quickly; record() should not block even when queue fills start = time.perf_counter() for i in range(50): collector.record(telemetry.RecordType.TOOL_EXECUTION, {"i": i}) elapsed_ms = (time.perf_counter() - start) * 1000.0 # Should be fast despite backpressure (non-blocking enqueue or drop) # Timeout relaxed to 200ms to handle thread scheduling variance in CI/local environments assert elapsed_ms < 200.0, f"Took {elapsed_ms:.1f}ms (expected <200ms)" # Allow worker to process some time.sleep(0.3) # Verify drops were logged (queue full backpressure) dropped_logs = [ m for m in caplog.messages if "Telemetry queue full; dropping" in m] assert len(dropped_logs) >= 1 # Ensure only one worker thread exists and is alive assert collector._worker.is_alive() worker_threads = [ t for t in threading.enumerate() if t is collector._worker] assert len(worker_threads) == 1

Latest Blog Posts

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/CoplayDev/unity-mcp'

If you have feedback or need assistance with the MCP directory API, please join our Discord server