Groq: High-speed, low-latency LLM inference via custom… – AIDIRECTORY