# file: P:\mcp-dc\src\tiny_llm.py
# hypothesis_version: 6.142.4
[0.1, 0.2, 0.25, 0.3, 0.5, 0.7, 0.8, 0.9, 0.95, 1.0, 1.1, 1.5, 2.0, 100, 1000, 'address', 'assistant', 'broadcast', 'category', 'chat.completion', 'choices', 'completion_tokens', 'confidence', 'container', 'content', 'created', 'default', 'device', 'find', 'finish_reason', 'function_calls', 'host', 'hosts', 'id', 'include_certainty', 'index', 'instance', 'ip', 'ips', 'limit', 'locate', 'look', 'machine', 'message', 'model', 'name', 'network', 'object', 'parameters', 'prompt_tokens', 'reasoning', 'response_time', 'role', 'search', 'segment', 'server', 'stop', 'subnet', 'tiny-test-llm', 'tool_calls', 'tool_name', 'total_tokens', 'usage', 'user', 'virtual', 'vlan', 'vlans', 'vm', 'vms']