llm.http 1.2 KB

1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556
  1. ### Chat completion test (vLLM local)
  2. # !! SSH access to tyrion.uv.es is needed !!
  3. POST http://127.0.0.1:8090/v1/chat/completions
  4. Authorization: Bearer hhOQ6QBqHKtOO9MKAUhIyU9auBkgIF40QJKa24jWJzdtxvdXMLi10xUAWMsdpFP0
  5. Content-Type: application/json
  6. {
  7. "model": "/media/nas/peerobs_sync/shared/2025-ReviewSim/models/Qwen2.5-7B-Instruct-AWQ",
  8. "messages": [
  9. {
  10. "role": "user",
  11. "content": "What's 1+1"
  12. }
  13. ],
  14. "max_tokens": 100,
  15. "temperature": 0.7
  16. }
  17. ###
  18. POST http://tyrion.uv.es:8090/v1/chat/completions
  19. Authorization: Bearer hhOQ6QBqHKtOO9MKAUhIyU9auBkgIF40QJKa24jWJzdtxvdXMLi10xUAWMsdpFP0
  20. Content-Type: application/json
  21. {
  22. "model": "/media/nas/peerobs_sync/shared/2025-ReviewSim/models/Qwen2.5-7B-Instruct-AWQ",
  23. "messages": [
  24. {
  25. "role": "user",
  26. "content": "What's 1+1"
  27. }
  28. ],
  29. "max_tokens": 100,
  30. "temperature": 0.7
  31. }
  32. ###
  33. ### Chat completion test (Groq OpenAI-compatible API)
  34. POST https://api.groq.com/openai/v1/chat/completions
  35. Authorization: Bearer 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
  36. Content-Type: application/json
  37. {
  38. "model": "llama-3.3-70b-versatile",
  39. "messages": [
  40. {
  41. "role": "user",
  42. "content": "What's 1+1"
  43. }
  44. ],
  45. "max_tokens": 100,
  46. "temperature": 0.7
  47. }