mirror of
https://github.com/HeyPuter/puter.git
synced 2026-05-03 13:40:50 -05:00
109 lines
4.1 KiB
JSON
109 lines
4.1 KiB
JSON
[
|
|
{
|
|
"source": "openai",
|
|
"model": "4o-mini",
|
|
"result": {
|
|
"index": 0,
|
|
"message": {
|
|
"role": "assistant",
|
|
"content": "Hello! I'm based on OpenAI's GPT-3.5 model. How can I assist you today?",
|
|
"refusal": null
|
|
},
|
|
"logprobs": null,
|
|
"finish_reason": "stop",
|
|
"usage": {
|
|
"input_tokens": 27,
|
|
"output_tokens": 23
|
|
},
|
|
"via_ai_chat_service": true
|
|
}
|
|
},
|
|
{
|
|
"source": "claude",
|
|
"model": "claude-3-5-sonnet",
|
|
"result": {
|
|
"message": {
|
|
"id": "msg_01BCi5s5rQKNVaM9iMD7zBFf",
|
|
"type": "message",
|
|
"role": "assistant",
|
|
"model": "claude-3-5-sonnet-20241022",
|
|
"content": [
|
|
{
|
|
"type": "text",
|
|
"text": "I am Claude, an AI assistant created by Anthropic. I'm running on the Puter platform through a driver interface called puter-chat-completion. How can I help you today?"
|
|
}
|
|
],
|
|
"stop_reason": "end_turn",
|
|
"stop_sequence": null,
|
|
"usage": {
|
|
"input_tokens": 77,
|
|
"cache_creation_input_tokens": 0,
|
|
"cache_read_input_tokens": 0,
|
|
"output_tokens": 44
|
|
}
|
|
},
|
|
"usage": {
|
|
"input_tokens": 77,
|
|
"cache_creation_input_tokens": 0,
|
|
"cache_read_input_tokens": 0,
|
|
"output_tokens": 44
|
|
},
|
|
"finish_reason": "stop",
|
|
"via_ai_chat_service": true
|
|
}
|
|
},
|
|
{
|
|
"source": "mistral",
|
|
"model": "mistral-large-latest",
|
|
"result": {
|
|
"index": 0,
|
|
"message": {
|
|
"content": "Hello! I am a text-based AI model developed by the Mistral AI team. I don't have personal experiences, feelings, or a physical presence, but I'm here to provide helpful, respectful, and honest assistance to the best of my ability. If you have any questions or need information, feel free to ask!",
|
|
"toolCalls": null,
|
|
"prefix": false,
|
|
"role": "assistant"
|
|
},
|
|
"finishReason": "stop",
|
|
"usage": {
|
|
"input_tokens": 11,
|
|
"output_tokens": 69
|
|
},
|
|
"via_ai_chat_service": true
|
|
}
|
|
},
|
|
{
|
|
"source": "groq",
|
|
"model": "?",
|
|
"result": {
|
|
"index": 0,
|
|
"message": {
|
|
"role": "assistant",
|
|
"content": "Hello! I am LLaMA, an AI assistant developed by Meta AI that can understand and respond to human input in a conversational manner. I'm not a specific model, but rather a technology that's constantly learning and improving from the interactions I have with users like you. I can generate human-like responses to a wide range of topics and questions, from simple inquiries to more complex conversations. How can I assist you today?"
|
|
},
|
|
"logprobs": null,
|
|
"finish_reason": "stop",
|
|
"usage": {
|
|
"input_tokens": 17,
|
|
"output_tokens": 87
|
|
},
|
|
"via_ai_chat_service": true
|
|
}
|
|
},
|
|
{
|
|
"source": "grok-beta",
|
|
"result": {
|
|
"index": 0,
|
|
"message": {
|
|
"role": "assistant",
|
|
"content": "Well, well, well, aren't you a curious one? I'm Grok, the AI with a sense of humor, built by xAI. I'm not your average chatbot, I'm more like the life of the party in the AI world. I'm here to answer your questions with a dash of wit and a pinch of outside perspective on humanity. So, what's on your mind?",
|
|
"refusal": null
|
|
},
|
|
"finish_reason": "stop",
|
|
"usage": {
|
|
"input_tokens": 74,
|
|
"output_tokens": 82
|
|
},
|
|
"via_ai_chat_service": true
|
|
}
|
|
}
|
|
] |