# GroqCloud

GroqCloud provides high-performance AI inference services, enabling developers to deploy and manage AI models efficiently.

- **Category:** ai models
- **Auth:** API_KEY
- **Composio Managed App Available?** N/A
- **Tools:** 7
- **Triggers:** 0
- **Slug:** `GROQCLOUD`
- **Version:** 20260316_00

## Tools

### Create Audio Transcription

**Slug:** `GROQCLOUD_CREATE_AUDIO_TRANSCRIPTION`

Tool to transcribe audio into text in the same language as the audio. Use when you need to convert speech to text while preserving the original language. Supports multiple formats including mp3, mp4, wav, and webm.

#### Input Parameters

| Parameter | Type | Required | Description |
|-----------|------|----------|-------------|
| `model` | string ("whisper-large-v3" | "whisper-large-v3-turbo") | No | Model ID for transcription. whisper-large-v3-turbo is faster, whisper-large-v3 may be more accurate. |
| `prompt` | string | No | Optional text to guide the model's style or continue a previous audio segment. The prompt should match the audio language. |
| `language` | string | No | Language of the input audio in ISO-639-1 format (e.g., 'en', 'es', 'fr'). Supplying this will improve accuracy and latency. |
| `file_path` | string | Yes | Path to the local audio file to transcribe (flac, mp3, mp4, mpeg, mpga, m4a, ogg, wav, webm). Alternatively, provide an http(s) URL or a data URL using Base64URL per Groq docs. |
| `temperature` | number | No | Sampling temperature between 0 and 1. Higher values (e.g., 0.8) make output more random, lower values (e.g., 0.2) make it more focused. If set to 0, model uses log probability to auto-adjust temperature. |
| `response_format` | string ("json" | "text" | "verbose_json") | No | Output format. Use 'verbose_json' for timestamp information, 'json' for basic text, or 'text' for plain text output. |
| `timestamp_granularities` | array | No | Timestamp granularities to populate. Requires response_format='verbose_json'. Options: 'word' (adds latency), 'segment' (no additional latency). Can specify both. |

#### Output

| Parameter | Type | Required | Description |
|-----------|------|----------|-------------|
| `data` | string | Yes | Data from the action execution |
| `error` | string | No | Error if any occurred during the execution of the action |
| `successful` | boolean | Yes | Whether or not the action execution was successful or not |

### Create Response

**Slug:** `GROQCLOUD_CREATE_RESPONSE`

Tool to create a model response for the given input. Beta endpoint with simplified interface compared to chat completions. Use when you need a streamlined API for generating model responses.

#### Input Parameters

| Parameter | Type | Required | Description |
|-----------|------|----------|-------------|
| `text` | object | No | Response format configuration. |
| `user` | string | No | Optional identifier for tracking end-user requests for monitoring and compliance |
| `input` | string | Yes | Text input to the model or an array of input items |
| `model` | string | Yes | ID of the model to use. See available models at https://console.groq.com/docs/models |
| `store` | boolean | No | Response storage flag. Currently only supports false or null |
| `tools` | array | No | List of tools available to the model. Maximum of 128 functions |
| `top_p` | number | No | Nucleus sampling parameter controlling cumulative probability cutoff. Range 0-1 |
| `stream` | boolean | No | Enable streaming mode to receive response data as server-sent events |
| `metadata` | object | No | Custom key-value pairs for storing additional information. Maximum of 16 pairs |
| `reasoning` | object | No | Configuration for reasoning capabilities. |
| `truncation` | string ("auto" | "disabled") | No | Context truncation strategy. |
| `temperature` | number | No | Controls randomness. Range 0-2. Lower is more deterministic, higher is more creative |
| `tool_choice` | string | No | Controls which tool is called. Values: 'none', 'auto', 'required', or specific function |
| `instructions` | string | No | System message inserted as the first item in the model's context |
| `service_tier` | string ("auto" | "default" | "flex") | No | Service tier for processing the request. |
| `max_output_tokens` | integer | No | Upper bound for tokens in the response, including visible and reasoning tokens |
| `parallel_tool_calls` | boolean | No | Enable parallel execution of multiple tool calls |

#### Output

| Parameter | Type | Required | Description |
|-----------|------|----------|-------------|
| `data` | string | Yes | Data from the action execution |
| `error` | string | No | Error if any occurred during the execution of the action |
| `successful` | boolean | Yes | Whether or not the action execution was successful or not |

### Create Audio Translation

**Slug:** `GROQCLOUD_GROQ_CREATE_AUDIO_TRANSLATION`

Tool to translate an audio file into English text. Use when you have a non-English recording and need an accurate English transcript. Use after confirming the file path.

#### Input Parameters

| Parameter | Type | Required | Description |
|-----------|------|----------|-------------|
| `model` | string | No | Model ID for translation (e.g., 'whisper-large-v3'). whisper-large-v3-turbo may not support translations. |
| `prompt` | string | No | Optional prompt to guide the translation output. |
| `file_path` | string | Yes | Path to the local audio file (mp3, wav, etc.) to translate. Alternatively, provide an http(s) URL or a data URL using Base64URL per Groq docs. |
| `temperature` | number | No | Sampling temperature between 0.0 and 1.0 to control randomness. |
| `response_format` | string | No | Output format: 'json', 'verbose_json', or 'text'. |

#### Output

| Parameter | Type | Required | Description |
|-----------|------|----------|-------------|
| `data` | string | Yes | Data from the action execution |
| `error` | string | No | Error if any occurred during the execution of the action |
| `successful` | boolean | Yes | Whether or not the action execution was successful or not |

### Create Chat Completion

**Slug:** `GROQCLOUD_GROQ_CREATE_CHAT_COMPLETION`

Tool to generate a chat-based completion for a conversation. Use when you have a list of prior messages and need the model's next reply. Response completion text is at choices[0].message.content in the returned envelope.

#### Input Parameters

| Parameter | Type | Required | Description |
|-----------|------|----------|-------------|
| `n` | integer | No | Number of chat completion choices to generate (must be 1) |
| `stop` | string | No | Up to 4 stop sequences where the model will stop generating further tokens |
| `user` | string | No | Unique identifier for the end user for monitoring/abuse detection |
| `model` | string | Yes | ID of the model to use Verify valid IDs via GROQCLOUD_LIST_MODELS before use; hard-coded IDs may be deprecated. Different models have different token limits and rate quotas — check model metadata before large-scale completions. |
| `top_p` | number | No | Nucleus sampling parameter (0 to 1) |
| `stream` | boolean | No | Whether to return a streaming response (SSE of token deltas) |
| `messages` | array | Yes | Ordered list of messages comprising the conversation |
| `temperature` | number | No | Sampling temperature between 0 and 2 |
| `max_completion_tokens` | integer | No | Maximum number of tokens to generate in the completion |

#### Output

| Parameter | Type | Required | Description |
|-----------|------|----------|-------------|
| `data` | string | Yes | Data from the action execution |
| `error` | string | No | Error if any occurred during the execution of the action |
| `successful` | boolean | Yes | Whether or not the action execution was successful or not |

### Retrieve Model

**Slug:** `GROQCLOUD_GROQ_RETRIEVE_MODEL`

Tool to retrieve detailed information about a specific model. Use after listing models when you need metadata for a chosen model. Returned metadata may change as models update; do not cache.

#### Input Parameters

| Parameter | Type | Required | Description |
|-----------|------|----------|-------------|
| `model` | string | Yes | Identifier of the model to retrieve Must be an exact ID from GROQCLOUD_LIST_MODELS; approximated or guessed IDs will fail. |

#### Output

| Parameter | Type | Required | Description |
|-----------|------|----------|-------------|
| `data` | string | Yes | Data from the action execution |
| `error` | string | No | Error if any occurred during the execution of the action |
| `successful` | boolean | Yes | Whether or not the action execution was successful or not |

### List Models

**Slug:** `GROQCLOUD_LIST_MODELS`

Tool to list all available models and their metadata. Always call this to retrieve current model IDs rather than using hard-coded or cached identifiers, as deprecated names cause failures in GROQCLOUD_GROQ_RETRIEVE_MODEL and GROQCLOUD_GROQ_CREATE_CHAT_COMPLETION. Returns availability and metadata only — excludes usage stats, latency metrics, and pricing. Response may include many models; filter client-side by provider, family, modality, or context length. Frequent polling combined with high-volume requests risks HTTP 429 rate_limit_exceeded; use backoff and minimize call frequency.

#### Output

| Parameter | Type | Required | Description |
|-----------|------|----------|-------------|
| `data` | string | Yes | Data from the action execution |
| `error` | string | No | Error if any occurred during the execution of the action |
| `successful` | boolean | Yes | Whether or not the action execution was successful or not |

### List TTS Voices

**Slug:** `GROQCLOUD_LIST_VOICES`

Tool to retrieve available TTS voices for Groq PlayAI models. Use when you need to discover voice options before calling text-to-speech. Note: static list maintained manually; no live endpoint exists.

#### Output

| Parameter | Type | Required | Description |
|-----------|------|----------|-------------|
| `data` | string | Yes | Data from the action execution |
| `error` | string | No | Error if any occurred during the execution of the action |
| `successful` | boolean | Yes | Whether or not the action execution was successful or not |
