From 63f47039b9edfbd46bca00f6aeeb99dea8dcfc66 Mon Sep 17 00:00:00 2001 From: Ping-Lin Chang Date: Mon, 30 Sep 2024 10:03:36 +0100 Subject: [PATCH] fix(docs): fix tasks.json inconsistent issue (#381) Because - there are many wording inconsistency issue in the `tasks.json` files. - duplicated `instillUIOrder` value causes the generated `README.mdx` inconsistent This commit - fix the issues --- ai/anthropic/v0/README.mdx | 4 ++-- ai/anthropic/v0/config/tasks.json | 4 ++-- ai/cohere/v0/README.mdx | 4 ++-- ai/cohere/v0/config/tasks.json | 4 ++-- ai/fireworksai/v0/README.mdx | 4 ++-- ai/fireworksai/v0/config/tasks.json | 4 ++-- ai/groq/v0/README.mdx | 4 ++-- ai/groq/v0/config/tasks.json | 4 ++-- ai/instill/v0/README.mdx | 16 ++++++++-------- ai/instill/v0/config/tasks.json | 8 ++++---- ai/mistralai/v0/README.mdx | 4 ++-- ai/mistralai/v0/config/tasks.json | 4 ++-- ai/ollama/v0/README.mdx | 4 ++-- ai/ollama/v0/config/tasks.json | 4 ++-- ai/openai/v0/README.mdx | 2 +- ai/openai/v0/config/tasks.json | 2 +- ai/universalai/v0/README.mdx | 14 +++++++------- ai/universalai/v0/config/tasks.json | 16 ++++++++-------- 18 files changed, 53 insertions(+), 53 deletions(-) diff --git a/ai/anthropic/v0/README.mdx b/ai/anthropic/v0/README.mdx index 4513f91d..b3446406 100644 --- a/ai/anthropic/v0/README.mdx +++ b/ai/anthropic/v0/README.mdx @@ -51,13 +51,13 @@ Anthropic's text generation models (often called generative pre-trained transfor | Task ID (required) | `task` | string | `TASK_TEXT_GENERATION_CHAT` | | Model Name (required) | `model-name` | string | The Anthropic model to be used | | Prompt (required) | `prompt` | string | The prompt text | -| System message | `system-message` | string | The system message helps set the behavior of the assistant. For example, you can modify the personality of the assistant or provide specific instructions about how it should behave throughout the conversation. By default, the model’s behavior is set using a generic message as "You are a helpful assistant." | +| System Message | `system-message` | string | The system message helps set the behavior of the assistant. For example, you can modify the personality of the assistant or provide specific instructions about how it should behave throughout the conversation. By default, the model’s behavior is set using a generic message as "You are a helpful assistant." | | Prompt Images | `prompt-images` | array[string] | The prompt images (Note: The prompt images will be injected in the order they are provided to the 'prompt' message. Anthropic doesn't support sending images via image-url, use this field instead) | | [Chat history](#text-generation-chat-chat-history) | `chat-history` | array[object] | Incorporate external chat history, specifically previous messages within the conversation. Please note that System Message will be ignored and will not have any effect when this field is populated. Each message should adhere to the format: : \{"role": "The message role, i.e. 'system', 'user' or 'assistant'", "content": "message content"\}. | | Seed | `seed` | integer | The seed (Note: Not supported by Anthropic Models) | | Temperature | `temperature` | number | The temperature for sampling | | Top K | `top-k` | integer | Top k for sampling | -| Max new tokens | `max-new-tokens` | integer | The maximum number of tokens for model to generate | +| Max New Tokens | `max-new-tokens` | integer | The maximum number of tokens for model to generate | diff --git a/ai/anthropic/v0/config/tasks.json b/ai/anthropic/v0/config/tasks.json index 52902b27..9e0def28 100644 --- a/ai/anthropic/v0/config/tasks.json +++ b/ai/anthropic/v0/config/tasks.json @@ -135,7 +135,7 @@ "value", "reference" ], - "title": "Max new tokens", + "title": "Max New Tokens", "type": "integer" }, "model-name": { @@ -227,7 +227,7 @@ "reference", "template" ], - "title": "System message", + "title": "System Message", "type": "string" }, "temperature": { diff --git a/ai/cohere/v0/README.mdx b/ai/cohere/v0/README.mdx index 7bac5f70..a634bbca 100644 --- a/ai/cohere/v0/README.mdx +++ b/ai/cohere/v0/README.mdx @@ -53,14 +53,14 @@ Cohere's text generation models (often called generative pre-trained transformer | Task ID (required) | `task` | string | `TASK_TEXT_GENERATION_CHAT` | | Model Name (required) | `model-name` | string | The Cohere command model to be used | | Prompt (required) | `prompt` | string | The prompt text | -| System message | `system-message` | string | The system message helps set the behavior of the assistant. For example, you can modify the personality of the assistant or provide specific instructions about how it should behave throughout the conversation. By default, the model’s behavior is using a generic message as "You are a helpful assistant." | +| System Message | `system-message` | string | The system message helps set the behavior of the assistant. For example, you can modify the personality of the assistant or provide specific instructions about how it should behave throughout the conversation. By default, the model’s behavior is using a generic message as "You are a helpful assistant." | | Documents | `documents` | array[string] | The documents to be used for the model, for optimal performance, the length of each document should be less than 300 words. | | Prompt Images | `prompt-images` | array[string] | The prompt images (Note: As for 2024-06-24 Cohere models are not multimodal, so images will be ignored.) | | [Chat history](#text-generation-chat-chat-history) | `chat-history` | array[object] | Incorporate external chat history, specifically previous messages within the conversation. Each message should adhere to the format: : \{"role": "The message role, i.e. 'USER' or 'CHATBOT'", "content": "message content"\}. | | Seed | `seed` | integer | The seed (default=42) | | Temperature | `temperature` | number | The temperature for sampling (default=0.7) | | Top K | `top-k` | integer | Top k for sampling (default=10) | -| Max new tokens | `max-new-tokens` | integer | The maximum number of tokens for model to generate (default=50) | +| Max New Tokens | `max-new-tokens` | integer | The maximum number of tokens for model to generate (default=50) | diff --git a/ai/cohere/v0/config/tasks.json b/ai/cohere/v0/config/tasks.json index 91d5f6a1..8224c8d1 100644 --- a/ai/cohere/v0/config/tasks.json +++ b/ai/cohere/v0/config/tasks.json @@ -254,7 +254,7 @@ "value", "reference" ], - "title": "Max new tokens", + "title": "Max New Tokens", "type": "integer" }, "model-name": { @@ -349,7 +349,7 @@ "reference", "template" ], - "title": "System message", + "title": "System Message", "type": "string" }, "temperature": { diff --git a/ai/fireworksai/v0/README.mdx b/ai/fireworksai/v0/README.mdx index 4d0509eb..51b9680a 100644 --- a/ai/fireworksai/v0/README.mdx +++ b/ai/fireworksai/v0/README.mdx @@ -52,13 +52,13 @@ Fireworks AI's text generation models (often called generative pre-trained trans | Task ID (required) | `task` | string | `TASK_TEXT_GENERATION_CHAT` | | Model Name (required) | `model` | string | The OSS model to be used | | Prompt (required) | `prompt` | string | The prompt text | -| System message | `system-message` | string | The system message helps set the behavior of the assistant. For example, you can modify the personality of the assistant or provide specific instructions about how it should behave throughout the conversation. By default, the model’s behavior is set using a generic message as "You are a helpful assistant." | +| System Message | `system-message` | string | The system message helps set the behavior of the assistant. For example, you can modify the personality of the assistant or provide specific instructions about how it should behave throughout the conversation. By default, the model’s behavior is set using a generic message as "You are a helpful assistant." | | Prompt Images | `prompt-images` | array[string] | The prompt images (Note: According to Fireworks AI documentation on 2024-07-24, the total number of images included in a single API request should not exceed 30, and all the images should be smaller than 5MB in size) | | [Chat history](#text-generation-chat-chat-history) | `chat-history` | array[object] | Incorporate external chat history, specifically previous messages within the conversation. Please note that System Message will be ignored and will not have any effect when this field is populated. Each message should adhere to the format: : \{"role": "The message role, i.e. 'system', 'user' or 'assistant'", "content": "message content"\} | | Seed | `seed` | integer | The seed | | Temperature | `temperature` | number | The temperature for sampling | | Top K | `top-k` | integer | Integer to define the top tokens considered within the sample operation to create new text | -| Max new tokens | `max-new-tokens` | integer | The maximum number of tokens for model to generate | +| Max New Tokens | `max-new-tokens` | integer | The maximum number of tokens for model to generate | | Top P | `top-p` | number | Float to define the tokens that are within the sample operation of text generation. Add tokens in the sample for more probable to least probable until the sum of the probabilities is greater than top-p (default=0.5) | diff --git a/ai/fireworksai/v0/config/tasks.json b/ai/fireworksai/v0/config/tasks.json index 50143972..7ba080d2 100644 --- a/ai/fireworksai/v0/config/tasks.json +++ b/ai/fireworksai/v0/config/tasks.json @@ -153,7 +153,7 @@ "value", "reference" ], - "title": "Max new tokens", + "title": "Max New Tokens", "type": "integer" }, "model": { @@ -266,7 +266,7 @@ "reference", "template" ], - "title": "System message", + "title": "System Message", "type": "string" }, "temperature": { diff --git a/ai/groq/v0/README.mdx b/ai/groq/v0/README.mdx index 57c27d4d..f2e3e182 100644 --- a/ai/groq/v0/README.mdx +++ b/ai/groq/v0/README.mdx @@ -51,13 +51,13 @@ Groq serves open source text generation models (often called generative pre-trai | Task ID (required) | `task` | string | `TASK_TEXT_GENERATION_CHAT` | | Model (required) | `model` | string | The OSS model to be used | | Prompt (required) | `prompt` | string | The prompt text | -| System message | `system-message` | string | The system message helps set the behavior of the assistant. For example, you can modify the personality of the assistant or provide specific instructions about how it should behave throughout the conversation. By default, the model’s behavior is set using a generic message as "You are a helpful assistant." | +| System Message | `system-message` | string | The system message helps set the behavior of the assistant. For example, you can modify the personality of the assistant or provide specific instructions about how it should behave throughout the conversation. By default, the model’s behavior is set using a generic message as "You are a helpful assistant." | | Prompt Images | `prompt-images` | array[string] | The prompt images (Note: Only a subset of OSS models support image inputs) | | [Chat history](#text-generation-chat-chat-history) | `chat-history` | array[object] | Incorporate external chat history, specifically previous messages within the conversation. Please note that System Message will be ignored and will not have any effect when this field is populated. Each message should adhere to the format: : \{"role": "The message role, i.e. 'system', 'user' or 'assistant'", "content": "message content"\} | | Seed | `seed` | integer | The seed | | Temperature | `temperature` | number | The temperature for sampling | | Top K | `top-k` | integer | Integer to define the top tokens considered within the sample operation to create new text | -| Max new tokens | `max-new-tokens` | integer | The maximum number of tokens for model to generate | +| Max New Tokens | `max-new-tokens` | integer | The maximum number of tokens for model to generate | | Top P | `top-p` | number | Float to define the tokens that are within the sample operation of text generation. Add tokens in the sample for more probable to least probable until the sum of the probabilities is greater than top-p (default=0.5) | | User | `user` | string | The user name passed to GroqPlatform | diff --git a/ai/groq/v0/config/tasks.json b/ai/groq/v0/config/tasks.json index cc259d34..3b3b01e4 100644 --- a/ai/groq/v0/config/tasks.json +++ b/ai/groq/v0/config/tasks.json @@ -135,7 +135,7 @@ "value", "reference" ], - "title": "Max new tokens", + "title": "Max New Tokens", "type": "integer" }, "model": { @@ -237,7 +237,7 @@ "reference", "template" ], - "title": "System message", + "title": "System Message", "type": "string" }, "temperature": { diff --git a/ai/instill/v0/README.mdx b/ai/instill/v0/README.mdx index 758133ca..7c865976 100644 --- a/ai/instill/v0/README.mdx +++ b/ai/instill/v0/README.mdx @@ -327,10 +327,10 @@ Generate texts from input text prompts. | Task ID (required) | `task` | string | `TASK_TEXT_GENERATION` | | Model Name (required) | `model-name` | string | The Instill Model model to be used. | | Prompt (required) | `prompt` | string | The prompt text | -| System message | `system-message` | string | The system message helps set the behavior of the assistant. For example, you can modify the personality of the assistant or provide specific instructions about how it should behave throughout the conversation. By default, the model’s behavior is using a generic message as "You are a helpful assistant." | +| System Message | `system-message` | string | The system message helps set the behavior of the assistant. For example, you can modify the personality of the assistant or provide specific instructions about how it should behave throughout the conversation. By default, the model’s behavior is using a generic message as "You are a helpful assistant." | | Seed | `seed` | integer | The seed | | Temperature | `temperature` | number | The temperature for sampling | -| Max new tokens | `max-new-tokens` | integer | The maximum number of tokens for model to generate | +| Max New Tokens | `max-new-tokens` | integer | The maximum number of tokens for model to generate | @@ -356,12 +356,12 @@ Generate texts from input text prompts and chat history. | Task ID (required) | `task` | string | `TASK_TEXT_GENERATION_CHAT` | | Model Name (required) | `model-name` | string | The Instill Model model to be used. | | Prompt (required) | `prompt` | string | The prompt text | -| System message | `system-message` | string | The system message helps set the behavior of the assistant. For example, you can modify the personality of the assistant or provide specific instructions about how it should behave throughout the conversation. By default, the model’s behavior is using a generic message as "You are a helpful assistant." | +| System Message | `system-message` | string | The system message helps set the behavior of the assistant. For example, you can modify the personality of the assistant or provide specific instructions about how it should behave throughout the conversation. By default, the model’s behavior is using a generic message as "You are a helpful assistant." | | Prompt Images | `prompt-images` | array[string] | The prompt images | | [Chat history](#text-generation-chat-chat-history) | `chat-history` | array[object] | Incorporate external chat history, specifically previous messages within the conversation. Please note that System Message will be ignored and will not have any effect when this field is populated. Each message should adhere to the format: : \{"role": "The message role, i.e. 'system', 'user' or 'assistant'", "content": "message content"\}. | | Seed | `seed` | integer | The seed | | Temperature | `temperature` | number | The temperature for sampling | -| Max new tokens | `max-new-tokens` | integer | The maximum number of tokens for model to generate | +| Max New Tokens | `max-new-tokens` | integer | The maximum number of tokens for model to generate | @@ -452,12 +452,12 @@ Answer questions based on a prompt and an image. | Task ID (required) | `task` | string | `TASK_VISUAL_QUESTION_ANSWERING` | | Model Name (required) | `model-name` | string | The Instill Model model to be used. | | Prompt (required) | `prompt` | string | The prompt text | -| System message | `system-message` | string | The system message helps set the behavior of the assistant. For example, you can modify the personality of the assistant or provide specific instructions about how it should behave throughout the conversation. By default, the model’s behavior is using a generic message as "You are a helpful assistant." | +| System Message | `system-message` | string | The system message helps set the behavior of the assistant. For example, you can modify the personality of the assistant or provide specific instructions about how it should behave throughout the conversation. By default, the model’s behavior is using a generic message as "You are a helpful assistant." | | Prompt Images | `prompt-images` | array[string] | The prompt images | | [Chat history](#visual-question-answering-chat-history) | `chat-history` | array[object] | Incorporate external chat history, specifically previous messages within the conversation. Please note that System Message will be ignored and will not have any effect when this field is populated. Each message should adhere to the format: : \{"role": "The message role, i.e. 'system', 'user' or 'assistant'", "content": "message content"\}. | | Seed | `seed` | integer | The seed | | Temperature | `temperature` | number | The temperature for sampling | -| Max new tokens | `max-new-tokens` | integer | The maximum number of tokens for model to generate | +| Max New Tokens | `max-new-tokens` | integer | The maximum number of tokens for model to generate | @@ -519,12 +519,12 @@ Generate texts from input text prompts and chat history. | Task ID (required) | `task` | string | `TASK_CHAT` | | Model Name (required) | `model-name` | string | The Instill Model model to be used. | | Prompt (required) | `prompt` | string | The prompt text | -| System message | `system-message` | string | The system message helps set the behavior of the assistant. For example, you can modify the personality of the assistant or provide specific instructions about how it should behave throughout the conversation. By default, the model’s behavior is using a generic message as "You are a helpful assistant." | +| System Message | `system-message` | string | The system message helps set the behavior of the assistant. For example, you can modify the personality of the assistant or provide specific instructions about how it should behave throughout the conversation. By default, the model’s behavior is using a generic message as "You are a helpful assistant." | | Prompt Images | `prompt-images` | array[string] | The prompt images | | [Chat history](#chat-chat-history) | `chat-history` | array[object] | Incorporate external chat history, specifically previous messages within the conversation. Please note that System Message will be ignored and will not have any effect when this field is populated. Each message should adhere to the format: : \{"role": "The message role, i.e. 'system', 'user' or 'assistant'", "content": "message content"\}. | | Seed | `seed` | integer | The seed | | Temperature | `temperature` | number | The temperature for sampling | -| Max new tokens | `max-new-tokens` | integer | The maximum number of tokens for model to generate | +| Max New Tokens | `max-new-tokens` | integer | The maximum number of tokens for model to generate | diff --git a/ai/instill/v0/config/tasks.json b/ai/instill/v0/config/tasks.json index 78539dfd..18cc0c2e 100644 --- a/ai/instill/v0/config/tasks.json +++ b/ai/instill/v0/config/tasks.json @@ -171,7 +171,7 @@ "value", "reference" ], - "title": "Max new tokens", + "title": "Max New Tokens", "type": "integer" }, "model-name": { @@ -230,7 +230,7 @@ "reference", "template" ], - "title": "System message", + "title": "System Message", "type": "string" }, "temperature": { @@ -315,7 +315,7 @@ "value", "reference" ], - "title": "Max new tokens", + "title": "Max New Tokens", "type": "integer" }, "model-name": { @@ -389,7 +389,7 @@ "reference", "template" ], - "title": "System message", + "title": "System Message", "type": "string" }, "temperature": { diff --git a/ai/mistralai/v0/README.mdx b/ai/mistralai/v0/README.mdx index 81bd60ec..1b13b69a 100644 --- a/ai/mistralai/v0/README.mdx +++ b/ai/mistralai/v0/README.mdx @@ -52,13 +52,13 @@ Mistral AI's text generation models (often called generative pre-trained transfo | Task ID (required) | `task` | string | `TASK_TEXT_GENERATION_CHAT` | | Model Name (required) | `model-name` | string | The Mistral model to be used | | Prompt (required) | `prompt` | string | The prompt text | -| System message | `system-message` | string | The system message helps set the behavior of the assistant. For example, you can modify the personality of the assistant or provide specific instructions about how it should behave throughout the conversation. By default, the model’s behavior is set using a generic message as "You are a helpful assistant." | +| System Message | `system-message` | string | The system message helps set the behavior of the assistant. For example, you can modify the personality of the assistant or provide specific instructions about how it should behave throughout the conversation. By default, the model’s behavior is set using a generic message as "You are a helpful assistant." | | Prompt Images | `prompt-images` | array[string] | The prompt images (Note: The Mistral models are not trained to process images, thus images will be omitted) | | [Chat history](#text-generation-chat-chat-history) | `chat-history` | array[object] | Incorporate external chat history, specifically previous messages within the conversation. Please note that System Message will be ignored and will not have any effect when this field is populated. Each message should adhere to the format: : \{"role": "The message role, i.e. 'system', 'user' or 'assistant'", "content": "message content"\} | | Seed | `seed` | integer | The seed | | Temperature | `temperature` | number | The temperature for sampling | | Top K | `top-k` | integer | Integer to define the top tokens considered within the sample operation to create new text (Note: The Mistral models does not support top-k sampling) | -| Max new tokens | `max-new-tokens` | integer | The maximum number of tokens for model to generate | +| Max New Tokens | `max-new-tokens` | integer | The maximum number of tokens for model to generate | | Top P | `top-p` | number | Float to define the tokens that are within the sample operation of text generation. Add tokens in the sample for more probable to least probable until the sum of the probabilities is greater than top-p (default=0.5) | | Safe | `safe` | boolean | Safe generation mode | diff --git a/ai/mistralai/v0/config/tasks.json b/ai/mistralai/v0/config/tasks.json index 52846090..9cbb0c5e 100644 --- a/ai/mistralai/v0/config/tasks.json +++ b/ai/mistralai/v0/config/tasks.json @@ -153,7 +153,7 @@ "value", "reference" ], - "title": "Max new tokens", + "title": "Max New Tokens", "type": "integer" }, "model-name": { @@ -249,7 +249,7 @@ "reference", "template" ], - "title": "System message", + "title": "System Message", "type": "string" }, "temperature": { diff --git a/ai/ollama/v0/README.mdx b/ai/ollama/v0/README.mdx index ca16724a..fa96aa33 100644 --- a/ai/ollama/v0/README.mdx +++ b/ai/ollama/v0/README.mdx @@ -53,13 +53,13 @@ Open-source large language models (OSS LLMs) are artificial intelligence models | Task ID (required) | `task` | string | `TASK_TEXT_GENERATION_CHAT` | | Model Name (required) | `model` | string | The OSS model to be used, check [here](https://ollama.com/library) for list of models available | | Prompt (required) | `prompt` | string | The prompt text | -| System message | `system-message` | string | The system message helps set the behavior of the assistant. For example, you can modify the personality of the assistant or provide specific instructions about how it should behave throughout the conversation. By default, the model’s behavior is set using a generic message as "You are a helpful assistant." | +| System Message | `system-message` | string | The system message helps set the behavior of the assistant. For example, you can modify the personality of the assistant or provide specific instructions about how it should behave throughout the conversation. By default, the model’s behavior is set using a generic message as "You are a helpful assistant." | | Prompt Images | `prompt-images` | array[string] | The prompt images | | [Chat history](#text-generation-chat-chat-history) | `chat-history` | array[object] | Incorporate external chat history, specifically previous messages within the conversation. Please note that System Message will be ignored and will not have any effect when this field is populated. Each message should adhere to the format: : \{"role": "The message role, i.e. 'system', 'user' or 'assistant'", "content": "message content"\}. | | Seed | `seed` | integer | The seed | | Temperature | `temperature` | number | The temperature for sampling | | Top K | `top-k` | integer | Top k for sampling | -| Max new tokens | `max-new-tokens` | integer | The maximum number of tokens for model to generate | +| Max New Tokens | `max-new-tokens` | integer | The maximum number of tokens for model to generate | diff --git a/ai/ollama/v0/config/tasks.json b/ai/ollama/v0/config/tasks.json index c3a6d2ed..d991015a 100644 --- a/ai/ollama/v0/config/tasks.json +++ b/ai/ollama/v0/config/tasks.json @@ -109,7 +109,7 @@ "value", "reference" ], - "title": "Max new tokens", + "title": "Max New Tokens", "type": "integer" }, "model": { @@ -184,7 +184,7 @@ "reference", "template" ], - "title": "System message", + "title": "System Message", "type": "string" }, "temperature": { diff --git a/ai/openai/v0/README.mdx b/ai/openai/v0/README.mdx index d0042666..4a969e3d 100644 --- a/ai/openai/v0/README.mdx +++ b/ai/openai/v0/README.mdx @@ -56,7 +56,7 @@ OpenAI's text generation models (often called generative pre-trained transformer | Task ID (required) | `task` | string | `TASK_TEXT_GENERATION` | | Model (required) | `model` | string | ID of the model to use. | | Prompt (required) | `prompt` | string | The prompt text | -| System message | `system-message` | string | The system message helps set the behavior of the assistant. For example, you can modify the personality of the assistant or provide specific instructions about how it should behave throughout the conversation. By default, the model’s behavior is using a generic message as "You are a helpful assistant." | +| System Message | `system-message` | string | The system message helps set the behavior of the assistant. For example, you can modify the personality of the assistant or provide specific instructions about how it should behave throughout the conversation. By default, the model’s behavior is using a generic message as "You are a helpful assistant." | | Image | `images` | array[string] | The images | | [Chat history](#text-generation-chat-history) | `chat-history` | array[object] | Incorporate external chat history, specifically previous messages within the conversation. Please note that System Message will be ignored and will not have any effect when this field is populated. Each message should adhere to the format \{"role": "The message role, i.e. 'system', 'user' or 'assistant'", "content": "message content"\}. | | Temperature | `temperature` | number | What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic. We generally recommend altering this or `top-p` but not both. | diff --git a/ai/openai/v0/config/tasks.json b/ai/openai/v0/config/tasks.json index a421902a..69cb78a3 100644 --- a/ai/openai/v0/config/tasks.json +++ b/ai/openai/v0/config/tasks.json @@ -506,7 +506,7 @@ "reference", "template" ], - "title": "System message", + "title": "System Message", "type": "string" }, "temperature": { diff --git a/ai/universalai/v0/README.mdx b/ai/universalai/v0/README.mdx index 061000b6..e8e6aaf1 100644 --- a/ai/universalai/v0/README.mdx +++ b/ai/universalai/v0/README.mdx @@ -51,7 +51,7 @@ Generate response base on conversation input | Input | ID | Type | Description | | :--- | :--- | :--- | :--- | | Task ID (required) | `task` | string | `TASK_CHAT` | -| [Chat data](#chat-chat-data) (required) | `data` | object | Input data | +| [Chat Data](#chat-chat-data) (required) | `data` | object | Input data | | [Input Parameter](#chat-input-parameter) | `parameter` | object | Input parameter | @@ -90,8 +90,8 @@ Input parameter | Field | Field ID | Type | Note | | :--- | :--- | :--- | :--- | -| Max new tokens | `max-tokens` | integer | The maximum number of tokens for model to generate | -| Number of choices | `n` | integer | How many chat completion choices to generate for each input message. | +| Max New Tokens | `max-tokens` | integer | The maximum number of tokens for model to generate | +| Number of Choices | `n` | integer | How many chat completion choices to generate for each input message. | | Seed | `seed` | integer | The seed, default is 0 | | Stream | `stream` | boolean | If set, partial message deltas will be sent. Tokens will be sent as data-only server-sent events as they become available. | | Temperature | `temperature` | number | The temperature for sampling | @@ -128,7 +128,7 @@ Input parameter | Field | Field ID | Type | Note | | :--- | :--- | :--- | :--- | | Created | `created` | integer | The Unix timestamp (in seconds) of when the chat completion was created. | -| Finish reason | `finish-reason` | string | The reason the model stopped generating tokens. | +| Finish Reason | `finish-reason` | string | The reason the model stopped generating tokens. | | Index | `index` | integer | The index of the choice in the list of choices. | | [Message](#chat-message) | `message` | object | A chat message generated by the model. | @@ -158,8 +158,8 @@ Input parameter | Field | Field ID | Type | Note | | :--- | :--- | :--- | :--- | -| Completion tokens | `completion-tokens` | integer | Number of tokens in the generated response. | -| Prompt tokens | `prompt-tokens` | integer | Number of tokens in the prompt. | -| Total tokens | `total-tokens` | integer | Total number of tokens used in the request (prompt + completion). | +| Completion Tokens | `completion-tokens` | integer | Number of tokens in the generated response. | +| Prompt Tokens | `prompt-tokens` | integer | Number of tokens in the prompt. | +| Total Tokens | `total-tokens` | integer | Total number of tokens used in the request (prompt + completion). | diff --git a/ai/universalai/v0/config/tasks.json b/ai/universalai/v0/config/tasks.json index bea99d04..365fcc87 100644 --- a/ai/universalai/v0/config/tasks.json +++ b/ai/universalai/v0/config/tasks.json @@ -10,7 +10,7 @@ "type": "object", "properties": { "data": { - "title": "Chat data", + "title": "Chat Data", "description": "Input data", "instillShortDescription": "Input data", "type": "object", @@ -219,7 +219,7 @@ "type": "object", "properties": { "max-tokens": { - "title": "Max new tokens", + "title": "Max New Tokens", "type": "integer", "description": "The maximum number of tokens for model to generate", "instillShortDescription": "The maximum number of tokens for model to generate", @@ -241,7 +241,7 @@ "instillUIOrder": 1 }, "n": { - "title": "Number of choices", + "title": "Number of Choices", "type": "integer", "description": "How many chat completion choices to generate for each input message.", "instillShortDescription": "How many chat completion choices to generate for each input message.", @@ -286,7 +286,7 @@ } }, "required": [], - "instillUIOrder": 0, + "instillUIOrder": 1, "title": "Input Parameter" } }, @@ -316,7 +316,7 @@ "type": "object", "properties": { "finish-reason": { - "title": "Finish reason", + "title": "Finish Reason", "type": "string", "description": "The reason the model stopped generating tokens.", "instillShortDescription": "The reason the model stopped generating tokens.", @@ -393,7 +393,7 @@ "type": "object", "properties": { "completion-tokens": { - "title": "Completion tokens", + "title": "Completion Tokens", "type": "integer", "description": "Number of tokens in the generated response.", "instillShortDescription": "Number of tokens in the generated response.", @@ -401,7 +401,7 @@ "instillUIOrder": 0 }, "prompt-tokens": { - "title": "Prompt tokens", + "title": "Prompt Tokens", "type": "integer", "description": "Number of tokens in the prompt.", "instillShortDescription": "Number of tokens in the prompt.", @@ -409,7 +409,7 @@ "instillUIOrder": 1 }, "total-tokens": { - "title": "Total tokens", + "title": "Total Tokens", "type": "integer", "description": "Total number of tokens used in the request (prompt + completion).", "instillShortDescription": "Total number of tokens used in the request (prompt + completion).",