{"cache_status":"MISS (Response from database)","data":[{"model_name":"Deepseek","login_required":false,"response_time_info":"25~35 second","description":"Free, high‑capacity AI (164K context & output) with 2.9s response, 44 TPS throughput, and 77B+ tokens served — optimized routing for maximum uptime."},{"model_name":"DeepSeek_R1","login_required":false,"response_time_info":"20~30 second","description":"API with massive 164K context, 2.7s fast responses, 36 TPS throughput, and 37B+ tokens served — optimized for reliability and scale."},{"model_name":"DeepSeek_V3","login_required":false,"response_time_info":"15~25 second","description":"Ultra‑fast AI (1.1s latency) with 33K context, 16K output, and 210B+ tokens served — optimized routing for reliability and scale."},{"model_name":"Gemini","login_required":true,"response_time_info":"10~15 second","description":"Blazing-fast AI with huge context, low latency, and high throughput — designed for demanding, large-scale applications."},{"model_name":"Gemini_2.0","login_required":false,"response_time_info":"1~8 second","description":"Blazing‑fast AI with a huge 1.05M context, 8K output, 1.1s latency, and 3.9B+ tokens served — built for extreme‑scale workloads."},{"model_name":"Gemma_3","login_required":false,"response_time_info":"5~15 second","description":"High‑speed AI with 96K context, 8K output, 1.1s latency, and 410M+ tokens served — optimized for reliable, large‑scale tasks."},{"model_name":"Google_Gemma_3n","login_required":false,"response_time_info":"8~15 second","description":"Lightning‑fast AI with 8K context, 2K output, just 0.3s latency, and 20M+ tokens served — built for instant responses."},{"model_name":"Meta_Llama","login_required":false,"response_time_info":"15~20 second","description":"Large‑scale AI with 66K context & output, 1.1s latency, and 42M+ tokens served — powerful for complex, high‑volume tasks."},{"model_name":"NVIDIA_Llama","login_required":false,"response_time_info":"20~30 second","description":"Ultra‑large AI with 131K context & output — built for massive, complex workloads at scale."}]}