### Chat completion test (vLLM local) # !! SSH access to tyrion.uv.es is needed !! POST http://127.0.0.1:8090/v1/chat/completions Authorization: Bearer hhOQ6QBqHKtOO9MKAUhIyU9auBkgIF40QJKa24jWJzdtxvdXMLi10xUAWMsdpFP0 Content-Type: application/json { "model": "/media/nas/peerobs_sync/shared/2025-ReviewSim/models/Qwen2.5-7B-Instruct-AWQ", "messages": [ { "role": "user", "content": "What's 1+1" } ], "max_tokens": 100, "temperature": 0.7 } ### POST http://tyrion.uv.es:8090/v1/chat/completions Authorization: Bearer hhOQ6QBqHKtOO9MKAUhIyU9auBkgIF40QJKa24jWJzdtxvdXMLi10xUAWMsdpFP0 Content-Type: application/json { "model": "/media/nas/peerobs_sync/shared/2025-ReviewSim/models/Qwen2.5-7B-Instruct-AWQ", "messages": [ { "role": "user", "content": "What's 1+1" } ], "max_tokens": 100, "temperature": 0.7 } ### ### Chat completion test (Groq OpenAI-compatible API) POST https://api.groq.com/openai/v1/chat/completions Authorization: Bearer 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 Content-Type: application/json { "model": "llama-3.3-70b-versatile", "messages": [ { "role": "user", "content": "What's 1+1" } ], "max_tokens": 100, "temperature": 0.7 }