From 98d63cfb5b3c948ced9087b76465ce1e9efa887e Mon Sep 17 00:00:00 2001 From: "Chang, Hui-Tang" Date: Tue, 7 Nov 2023 13:16:38 +0800 Subject: [PATCH] chore(component): update `instillShortDescription` and `instillUIOrder` in components (#40) Because - we will use `instillShortDescription` to display short description on Console. This commit - update `instillShortDescription` and `instillUIOrder` in components --- go.mod | 2 +- go.sum | 4 +- pkg/huggingface/config/tasks.json | 28 ++ pkg/instill/config/instill.json | 410 ------------------------------ pkg/instill/config/tasks.json | 22 +- pkg/instill/main.go | 8 +- pkg/numbers/config/tasks.json | 5 +- pkg/openai/config/openai.json | 10 +- pkg/openai/config/tasks.json | 11 + pkg/stabilityai/config/tasks.json | 28 +- 10 files changed, 81 insertions(+), 447 deletions(-) delete mode 100644 pkg/instill/config/instill.json diff --git a/go.mod b/go.mod index 729a61a..d150ff9 100644 --- a/go.mod +++ b/go.mod @@ -11,7 +11,7 @@ require ( github.com/ghodss/yaml v1.0.0 github.com/gofrs/uuid v4.4.0+incompatible github.com/h2non/filetype v1.1.3 - github.com/instill-ai/component v0.6.1-alpha.0.20231106153938-6032b6ce48ed + github.com/instill-ai/component v0.6.1-alpha.0.20231107045726-dc5371eae05c github.com/instill-ai/protogen-go v0.3.3-alpha.0.20231019202606-71607ddcd93f github.com/santhosh-tekuri/jsonschema/v5 v5.3.0 github.com/stretchr/testify v1.8.4 diff --git a/go.sum b/go.sum index 2b60727..df7148a 100644 --- a/go.sum +++ b/go.sum @@ -124,8 +124,8 @@ github.com/grpc-ecosystem/grpc-gateway/v2 v2.7.2 h1:I/pwhnUln5wbMnTyRbzswA0/JxpK github.com/grpc-ecosystem/grpc-gateway/v2 v2.7.2/go.mod h1:lsuH8kb4GlMdSlI4alNIBBSAt5CHJtg3i+0WuN9J5YM= github.com/h2non/filetype v1.1.3 h1:FKkx9QbD7HR/zjK1Ia5XiBsq9zdLi5Kf3zGyFTAFkGg= github.com/h2non/filetype v1.1.3/go.mod h1:319b3zT68BvV+WRj7cwy856M2ehB3HqNOt6sy1HndBY= -github.com/instill-ai/component v0.6.1-alpha.0.20231106153938-6032b6ce48ed h1:cZfbReRzNeNdMEIFNs7oW+6ewxH87GtHGlEkLdDAEqU= -github.com/instill-ai/component v0.6.1-alpha.0.20231106153938-6032b6ce48ed/go.mod h1:aQfeKtOn74cRQJOxvua1mQGpecasDaD+l46C60ILgdA= +github.com/instill-ai/component v0.6.1-alpha.0.20231107045726-dc5371eae05c h1:vMAp210Zvvpxopftwi7WBHWmDmM4pcb8oXd/a4wx2g4= +github.com/instill-ai/component v0.6.1-alpha.0.20231107045726-dc5371eae05c/go.mod h1:aQfeKtOn74cRQJOxvua1mQGpecasDaD+l46C60ILgdA= github.com/instill-ai/protogen-go v0.3.3-alpha.0.20231019202606-71607ddcd93f h1:hweU93u6qsg8GH/YSogOfa+wOZEnkilGsijcy1xKX7E= github.com/instill-ai/protogen-go v0.3.3-alpha.0.20231019202606-71607ddcd93f/go.mod h1:q/YL5TZXD9nvmJ7Rih4gY3/B2HT2+GiFdxeZp9D+yE4= github.com/kisielk/errcheck v1.5.0/go.mod h1:pFxgyoBC7bSaBwPgfKdkLd5X25qrDl4LWUI2bnpBCr8= diff --git a/pkg/huggingface/config/tasks.json b/pkg/huggingface/config/tasks.json index 43e4284..c840d03 100644 --- a/pkg/huggingface/config/tasks.json +++ b/pkg/huggingface/config/tasks.json @@ -1,6 +1,7 @@ { "$defs": { "model": { + "description": "The Hugging Face model to be used", "instillAcceptFormats": [ "string", "text/*" @@ -49,6 +50,7 @@ "type": "object" }, "string_input": { + "description": "String input", "instillAcceptFormats": [ "string", "text/*" @@ -68,6 +70,7 @@ "instillUIOrder": 0, "properties": { "audio": { + "description": "The audio file", "instillAcceptFormats": [ "audio/*" ], @@ -158,6 +161,7 @@ "array:string", "array:text/*" ], + "instillShortDescription": "A list of strings corresponding to the earlier replies from the user.", "instillUIOrder": 1, "instillUpstreamTypes": [ "value", @@ -221,6 +225,7 @@ "number", "integer" ], + "instillShortDescription": "The amount of time in seconds that the query should take maximum.", "instillUIOrder": 1, "instillUpstreamTypes": [ "value", @@ -267,6 +272,7 @@ "number", "integer" ], + "instillShortDescription": "The temperature of the sampling operation.", "instillUIOrder": 4, "instillUpstreamTypes": [ "value", @@ -296,6 +302,7 @@ "number", "integer" ], + "instillShortDescription": "Float to define the tokens that are within the sample operation of text generation.", "instillUIOrder": 6, "instillUpstreamTypes": [ "value", @@ -445,6 +452,7 @@ "instillUIOrder": 0, "properties": { "image": { + "description": "The image file", "instillAcceptFormats": [ "image/*" ], @@ -510,6 +518,7 @@ "instillUIOrder": 0, "properties": { "image": { + "description": "The image file", "instillAcceptFormats": [ "image/*" ], @@ -583,6 +592,7 @@ "instillUIOrder": 0, "properties": { "image": { + "description": "The image file", "instillAcceptFormats": [ "image/*" ], @@ -626,6 +636,7 @@ "instillUIOrder": 0, "properties": { "image": { + "description": "The image file", "instillAcceptFormats": [ "image/*" ], @@ -733,6 +744,7 @@ "instillUIOrder": 1, "properties": { "context": { + "description": "The context for answering the question.", "instillAcceptFormats": [ "string", "text/*" @@ -747,6 +759,7 @@ "type": "string" }, "question": { + "description": "The question", "instillAcceptFormats": [ "string", "text/*" @@ -852,6 +865,7 @@ "string", "text/*" ], + "instillShortDescription": "The string that you wish to compare the other strings with.", "instillUIOrder": 1, "instillUpstreamTypes": [ "value", @@ -909,6 +923,7 @@ "instillUIOrder": 0, "properties": { "audio": { + "description": "The audio file", "instillAcceptFormats": [ "audio/*" ], @@ -1031,6 +1046,7 @@ "number", "integer" ], + "instillShortDescription": "The temperature of the sampling operation.", "instillUIOrder": 4, "instillUpstreamTypes": [ "value", @@ -1060,6 +1076,7 @@ "number", "integer" ], + "instillShortDescription": "Float to define the tokens that are within the sample operation of text generation.", "instillUIOrder": 6, "instillUpstreamTypes": [ "value", @@ -1302,6 +1319,7 @@ "instillAcceptFormats": [ "integer" ], + "instillShortDescription": "The amount of new tokens to be generated.", "instillUIOrder": 1, "instillUpstreamTypes": [ "value", @@ -1316,6 +1334,7 @@ "number", "integer" ], + "instillShortDescription": "The amount of time in seconds that the query should take maximum.", "instillUIOrder": 2, "instillUpstreamTypes": [ "value", @@ -1370,6 +1389,7 @@ "number", "integer" ], + "instillShortDescription": "The temperature of the sampling operation.", "instillUIOrder": 6, "instillUpstreamTypes": [ "value", @@ -1397,6 +1417,7 @@ "number", "integer" ], + "instillShortDescription": "Float to define the tokens that are within the sample operation of text generation.", "instillUIOrder": 8, "instillUpstreamTypes": [ "value", @@ -1455,6 +1476,7 @@ "instillUIOrder": 2, "properties": { "guidance_scale": { + "description": "Guidance scale", "instillAcceptFormats": [ "number", "integer" @@ -1468,6 +1490,7 @@ "type": "number" }, "height": { + "description": "Image Height", "instillAcceptFormats": [ "integer" ], @@ -1480,6 +1503,7 @@ "type": "integer" }, "negative_prompt": { + "description": "Negative prompt for generating the image", "instillAcceptFormats": [ "string", "text/*" @@ -1494,6 +1518,7 @@ "type": "string" }, "num_inference_steps": { + "description": "Number of inference steps", "instillAcceptFormats": [ "integer" ], @@ -1506,6 +1531,7 @@ "type": "integer" }, "width": { + "description": "Image width", "instillAcceptFormats": [ "integer" ], @@ -1571,6 +1597,7 @@ "string", "text/*" ], + "instillShortDescription": "There are several aggregation strategies: none, simple, first, average, and max.", "instillUIOrder": 0, "instillUpstreamTypes": [ "value", @@ -1711,6 +1738,7 @@ "properties": { "candidate_labels": { "description": "a list of strings that are potential classes for inputs. (max 10 candidate_labels, for more, simply run multiple requests, results are going to be misleading if using too many candidate_labels anyway. If you want to keep the exact same, you can simply run multi_label=True and do the scaling on your end. )", + "instillShortDescription": "a list of strings that are potential classes for inputs.", "instillUIOrder": 0, "instillUpstreamTypes": [ "value", diff --git a/pkg/instill/config/instill.json b/pkg/instill/config/instill.json deleted file mode 100644 index cfae58f..0000000 --- a/pkg/instill/config/instill.json +++ /dev/null @@ -1,410 +0,0 @@ -{ - "TASK_CLASSIFICATION": { - "info": { - "title": "Instill-Model Connector Openapi", - "version": "1.0.0" - }, - "openapi": "3.0.0", - "paths": { - "/execute": { - "post": { - "requestBody": { - "content": { - "application/json": { - "schema": { - "properties": { - "inputs": { - "items": { - "$ref": "data.json#/TASK_CLASSIFICATION/input" - }, - "type": "array" - } - }, - "type": "object" - } - } - }, - "required": true - }, - "responses": { - "200": { - "content": { - "application/json": { - "schema": { - "properties": { - "outputs": { - "items": { - "$ref": "data.json#/TASK_CLASSIFICATION/output" - }, - "type": "array" - } - }, - "type": "object" - } - } - }, - "description": "" - } - } - } - } - } - }, - "TASK_DETECTION": { - "info": { - "title": "Instill-Model Connector Openapi", - "version": "1.0.0" - }, - "openapi": "3.0.0", - "paths": { - "/execute": { - "post": { - "requestBody": { - "content": { - "application/json": { - "schema": { - "properties": { - "inputs": { - "items": { - "$ref": "data.json#/TASK_DETECTION/input" - }, - "type": "array" - } - }, - "type": "object" - } - } - }, - "required": true - }, - "responses": { - "200": { - "content": { - "application/json": { - "schema": { - "properties": { - "outputs": { - "items": { - "$ref": "data.json#/TASK_DETECTION/output" - }, - "type": "array" - } - }, - "type": "object" - } - } - }, - "description": "" - } - } - } - } - } - }, - "TASK_INSTANCE_SEGMENTATION": { - "info": { - "title": "Instill-Model Connector Openapi", - "version": "1.0.0" - }, - "openapi": "3.0.0", - "paths": { - "/execute": { - "post": { - "requestBody": { - "content": { - "application/json": { - "schema": { - "properties": { - "inputs": { - "items": { - "$ref": "data.json#/TASK_INSTANCE_SEGMENTATION/input" - }, - "type": "array" - } - }, - "type": "object" - } - } - }, - "required": true - }, - "responses": { - "200": { - "content": { - "application/json": { - "schema": { - "properties": { - "outputs": { - "items": { - "$ref": "data.json#/TASK_INSTANCE_SEGMENTATION/output" - }, - "type": "array" - } - }, - "type": "object" - } - } - }, - "description": "" - } - } - } - } - } - }, - "TASK_KEYPOINT": { - "info": { - "title": "Instill-Model Connector Openapi", - "version": "1.0.0" - }, - "openapi": "3.0.0", - "paths": { - "/execute": { - "post": { - "requestBody": { - "content": { - "application/json": { - "schema": { - "properties": { - "inputs": { - "items": { - "$ref": "data.json#/TASK_KEYPOINT/input" - }, - "type": "array" - } - }, - "type": "object" - } - } - }, - "required": true - }, - "responses": { - "200": { - "content": { - "application/json": { - "schema": { - "properties": { - "outputs": { - "items": { - "$ref": "data.json#/TASK_KEYPOINT/output" - }, - "type": "array" - } - }, - "type": "object" - } - } - }, - "description": "" - } - } - } - } - } - }, - "TASK_OCR": { - "info": { - "title": "Instill-Model Connector Openapi", - "version": "1.0.0" - }, - "openapi": "3.0.0", - "paths": { - "/execute": { - "post": { - "requestBody": { - "content": { - "application/json": { - "schema": { - "properties": { - "inputs": { - "items": { - "$ref": "data.json#/TASK_OCR/input" - }, - "type": "array" - } - }, - "type": "object" - } - } - }, - "required": true - }, - "responses": { - "200": { - "content": { - "application/json": { - "schema": { - "properties": { - "outputs": { - "items": { - "$ref": "data.json#/TASK_OCR/output" - }, - "type": "array" - } - }, - "type": "object" - } - } - }, - "description": "" - } - } - } - } - } - }, - "TASK_SEMANTIC_SEGMENTATION": { - "info": { - "title": "Instill-Model Connector Openapi", - "version": "1.0.0" - }, - "openapi": "3.0.0", - "paths": { - "/execute": { - "post": { - "requestBody": { - "content": { - "application/json": { - "schema": { - "properties": { - "inputs": { - "items": { - "$ref": "data.json#/TASK_SEMANTIC_SEGMENTATION/input" - }, - "type": "array" - } - }, - "type": "object" - } - } - }, - "required": true - }, - "responses": { - "200": { - "content": { - "application/json": { - "schema": { - "properties": { - "outputs": { - "items": { - "$ref": "data.json#/TASK_SEMANTIC_SEGMENTATION/output" - }, - "type": "array" - } - }, - "type": "object" - } - } - }, - "description": "" - } - } - } - } - } - }, - "TASK_TEXT_GENERATION": { - "info": { - "title": "Instill-Model Connector Openapi", - "version": "1.0.0" - }, - "openapi": "3.0.0", - "paths": { - "/execute": { - "post": { - "requestBody": { - "content": { - "application/json": { - "schema": { - "properties": { - "inputs": { - "items": { - "$ref": "data.json#/TASK_TEXT_GENERATION/input" - }, - "type": "array" - } - }, - "type": "object" - } - } - }, - "required": true - }, - "responses": { - "200": { - "content": { - "application/json": { - "schema": { - "properties": { - "outputs": { - "items": { - "$ref": "data.json#/TASK_TEXT_GENERATION/output" - }, - "type": "array" - } - }, - "type": "object" - } - } - }, - "description": "" - } - } - } - } - } - }, - "TASK_TEXT_TO_IMAGE": { - "info": { - "title": "Instill-Model Connector Openapi", - "version": "1.0.0" - }, - "openapi": "3.0.0", - "paths": { - "/execute": { - "post": { - "requestBody": { - "content": { - "application/json": { - "schema": { - "properties": { - "inputs": { - "items": { - "$ref": "data.json#/TASK_TEXT_TO_IMAGE/input" - }, - "type": "array" - } - }, - "type": "object" - } - } - }, - "required": true - }, - "responses": { - "200": { - "content": { - "application/json": { - "schema": { - "properties": { - "outputs": { - "items": { - "$ref": "data.json#/TASK_TEXT_TO_IMAGE/output" - }, - "type": "array" - } - }, - "type": "object" - } - } - }, - "description": "" - } - } - } - } - } - } -} diff --git a/pkg/instill/config/tasks.json b/pkg/instill/config/tasks.json index 04b6979..2b4b3e9 100644 --- a/pkg/instill/config/tasks.json +++ b/pkg/instill/config/tasks.json @@ -150,7 +150,7 @@ "instillUIOrder": 0, "properties": { "bad_words_list": { - "description": "Bad words list", + "description": "The words not to avoid being generated by the model", "instillAcceptFormats": [ "string", "text/*" @@ -195,7 +195,7 @@ "type": "string" }, "output_len": { - "description": "Output length", + "description": "The maximum number of tokens for model to generate", "instillAcceptFormats": [ "integer" ], @@ -208,7 +208,7 @@ "type": "integer" }, "prompt": { - "description": "Prompt", + "description": "The prompt text", "instillAcceptFormats": [ "string", "text/*" @@ -223,7 +223,7 @@ "type": "string" }, "seed": { - "description": "Seed", + "description": "The seed", "instillAcceptFormats": [ "integer" ], @@ -236,7 +236,7 @@ "type": "integer" }, "stop_words_list": { - "description": "Stop words list", + "description": "The trigger words to stop generation", "instillAcceptFormats": [ "string", "text/*" @@ -251,7 +251,7 @@ "type": "string" }, "top_k": { - "description": "Top K", + "description": "Top k for sampling", "instillAcceptFormats": [ "integer" ], @@ -305,7 +305,7 @@ "instillUIOrder": 0, "properties": { "cfg_scale": { - "description": "CFG Scale", + "description": "The guidance scale, default is 7.5", "instillAcceptFormats": [ "number", "integer" @@ -349,7 +349,7 @@ "type": "string" }, "prompt": { - "description": "Prompt", + "description": "The prompt text", "instillAcceptFormats": [ "string", "text/*" @@ -364,7 +364,7 @@ "type": "string" }, "samples": { - "description": "Samples", + "description": "The number of generated samples, default is 1", "instillAcceptFormats": [ "integer" ], @@ -377,7 +377,7 @@ "type": "integer" }, "seed": { - "description": "Seed", + "description": "The seed, default is 0", "instillAcceptFormats": [ "integer" ], @@ -390,7 +390,7 @@ "type": "integer" }, "steps": { - "description": "Steps", + "description": "The steps, default is 5", "instillAcceptFormats": [ "integer" ], diff --git a/pkg/instill/main.go b/pkg/instill/main.go index 4b3516f..749d734 100644 --- a/pkg/instill/main.go +++ b/pkg/instill/main.go @@ -31,10 +31,8 @@ var ( definitionsJSON []byte //go:embed config/tasks.json tasksJSON []byte - //go:embed config/instill.json - instillJSON []byte - once sync.Once - connector base.IConnector + once sync.Once + connector base.IConnector ) type Connector struct { @@ -64,7 +62,7 @@ func Init(logger *zap.Logger) base.IConnector { Component: base.Component{Logger: logger}, }, } - err := connector.LoadConnectorDefinitions(definitionsJSON, tasksJSON, map[string][]byte{"instill.json": instillJSON}) + err := connector.LoadConnectorDefinitions(definitionsJSON, tasksJSON, nil) if err != nil { logger.Fatal(err.Error()) } diff --git a/pkg/numbers/config/tasks.json b/pkg/numbers/config/tasks.json index cab932d..0df2fdb 100644 --- a/pkg/numbers/config/tasks.json +++ b/pkg/numbers/config/tasks.json @@ -55,7 +55,7 @@ }, "digital_source_type": { "default": "trainedAlgorithmicMedia", - "description": "Specify the type of the source. More details see https://docs.numbersprotocol.io/introduction/numbers-protocol/defining-web3-assets/assettree/digitalsourcetype", + "description": "Specify the type of the source. More details here", "enum": [ "trainedAlgorithmicMedia", "trainedAlgorithmicData", @@ -149,7 +149,7 @@ }, "mining_preference": { "default": "notAllowed", - "description": "Designates the selection made by the asset creators or licensed owners to decide if the asset is suitable for inclusion in a data mining or AI/ML training workflow. More details see https://docs.numbersprotocol.io/introduction/numbers-protocol/defining-web3-assets/assettree/miningpreference", + "description": "Designates the selection made by the asset creators or licensed owners to decide if the asset is suitable for inclusion in a data mining or AI/ML training workflow. More details here", "enum": [ "dataMining", "aiInference", @@ -163,6 +163,7 @@ "string", "text/*" ], + "instillShortDescription": "Designates the selection made by the asset creators or licensed owners. More details here", "instillUIOrder": 5, "instillUpstreamTypes": [ "value", diff --git a/pkg/openai/config/openai.json b/pkg/openai/config/openai.json index 1dd9e32..fa24a49 100644 --- a/pkg/openai/config/openai.json +++ b/pkg/openai/config/openai.json @@ -197,7 +197,7 @@ }, "max_tokens": { "default": "inf", - "description": "The maximum number of [tokens](/tokenizer) to generate in the chat completion.\n\nThe total length of input tokens and generated tokens is limited by the model's context length. [Example Python code](https://github.com/openai/openai-cookbook/blob/main/examples/How_to_count_tokens_with_tiktoken.ipynb) for counting tokens.\n", + "description": "The maximum number of tokens to generate in the chat completion.\n\nThe total length of input tokens and generated tokens is limited by the model's context length. Example Python code for counting tokens.\n", "type": "integer" }, "messages": { @@ -209,7 +209,7 @@ "type": "array" }, "model": { - "description": "ID of the model to use. See the [model endpoint compatibility](https://platform.openai.com/docs/models/model-endpoint-compatibility) table for details on which models work with the Chat API.", + "description": "ID of the model to use. See the model endpoint compatibility table for details on which models work with the Chat API.", "enum": [ "gpt-4", "gpt-4-0314", @@ -1377,7 +1377,7 @@ "x-oaiTypeLabel": "file" }, "language": { - "description": "The language of the input audio. Supplying the input language in [ISO-639-1](https://en.wikipedia.org/wiki/List_of_ISO_639-1_codes) format will improve accuracy and latency.\n", + "description": "The language of the input audio. Supplying the input language in ISO-639-1 format will improve accuracy and latency.\n", "type": "string" }, "model": { @@ -1390,7 +1390,7 @@ "x-oaiTypeLabel": "string" }, "prompt": { - "description": "An optional text to guide the model's style or continue a previous audio segment. The [prompt](https://platform.openai.com/docs/guides/speech-to-text/prompting) should match the audio language.\n", + "description": "An optional text to guide the model's style or continue a previous audio segment. The prompt should match the audio language.\n", "type": "string" }, "response_format": { @@ -1407,7 +1407,7 @@ }, "temperature": { "default": 0, - "description": "The sampling temperature, between 0 and 1. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic. If set to 0, the model will use [log probability](https://en.wikipedia.org/wiki/Log_probability) to automatically increase the temperature until certain thresholds are hit.\n", + "description": "The sampling temperature, between 0 and 1. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic. If set to 0, the model will use log probability to automatically increase the temperature until certain thresholds are hit.\n", "type": "number" } }, diff --git a/pkg/openai/config/tasks.json b/pkg/openai/config/tasks.json index c25ac7d..db568ae 100644 --- a/pkg/openai/config/tasks.json +++ b/pkg/openai/config/tasks.json @@ -20,6 +20,7 @@ "string", "text/*" ], + "instillShortDescription": "The language of the input audio.", "instillUIOrder": 3, "instillUpstreamTypes": [ "value", @@ -34,6 +35,7 @@ "string", "text/*" ], + "instillShortDescription": "ID of the model to use", "instillUIOrder": 0, "instillUpstreamTypes": [ "value", @@ -48,6 +50,7 @@ "string", "text/*" ], + "instillShortDescription": "An optional text to guide the model's style or continue a previous audio segment.", "instillUIOrder": 2, "instillUpstreamTypes": [ "value", @@ -62,6 +65,7 @@ "number", "integer" ], + "instillShortDescription": "The sampling temperature, between 0 and 1.", "instillUIOrder": 4, "instillUpstreamTypes": [ "value", @@ -104,6 +108,7 @@ "string", "text/*" ], + "instillShortDescription": "ID of the model to use", "instillUIOrder": 0, "instillUpstreamTypes": [ "value", @@ -113,6 +118,7 @@ "title": "Model" }, "text": { + "description": "The text", "instillAcceptFormats": [ "string", "text/*" @@ -159,6 +165,7 @@ "instillAcceptFormats": [ "integer" ], + "instillShortDescription": "The maximum number of tokens to generate in the chat completion.", "instillUIOrder": 5, "instillUpstreamTypes": [ "value", @@ -172,6 +179,7 @@ "string", "text/*" ], + "instillShortDescription": "ID of the model to use", "instillUIOrder": 0, "instillUpstreamTypes": [ "value", @@ -193,6 +201,7 @@ "title": "N" }, "prompt": { + "description": "The prompt text", "instillAcceptFormats": [ "string", "text/*" @@ -213,6 +222,7 @@ "string", "text/*" ], + "instillShortDescription": "The system message helps set the behavior of the assistant", "instillUIOrder": 2, "instillUpstreamTypes": [ "value", @@ -229,6 +239,7 @@ "number", "integer" ], + "instillShortDescription": "What sampling temperature to use, between 0 and 2.", "instillUIOrder": 3, "instillUpstreamTypes": [ "value", diff --git a/pkg/stabilityai/config/tasks.json b/pkg/stabilityai/config/tasks.json index 79bd670..1bd4fda 100644 --- a/pkg/stabilityai/config/tasks.json +++ b/pkg/stabilityai/config/tasks.json @@ -56,7 +56,7 @@ "string", "text/*" ], - "instillUIOrder": 4, + "instillUIOrder": 0, "instillUpstreamTypes": [ "value", "reference", @@ -71,6 +71,7 @@ "number", "integer" ], + "instillShortDescription": "How much influence the `init_image` has on the diffusion process.", "instillUIOrder": 5, "instillUpstreamTypes": [ "value", @@ -109,7 +110,7 @@ "array:string", "array:text/*" ], - "instillUIOrder": 0, + "instillUIOrder": 1, "instillUpstreamTypes": [ "reference" ], @@ -126,6 +127,7 @@ "string", "text/*" ], + "instillShortDescription": "Which sampler to use for the diffusion process", "instillUIOrder": 8, "instillUpstreamTypes": [ "value", @@ -165,7 +167,8 @@ "number", "integer" ], - "instillUIOrder": 11, + "instillShortDescription": "Skips a proportion of the end of the diffusion steps", + "instillUIOrder": 12, "instillUpstreamTypes": [ "value", "reference" @@ -178,7 +181,8 @@ "number", "integer" ], - "instillUIOrder": 12, + "instillShortDescription": "Skips a proportion of the start of the diffusion steps", + "instillUIOrder": 11, "instillUpstreamTypes": [ "value", "reference" @@ -217,7 +221,7 @@ "array:number", "array:integer" ], - "instillUIOrder": 1, + "instillUIOrder": 2, "instillUpstreamTypes": [ "reference" ], @@ -254,7 +258,7 @@ "number", "integer" ], - "instillUIOrder": 2, + "instillUIOrder": 3, "instillUpstreamTypes": [ "value", "reference" @@ -268,7 +272,7 @@ "string", "text/*" ], - "instillUIOrder": 3, + "instillUIOrder": 4, "instillUpstreamTypes": [ "value", "reference", @@ -296,7 +300,7 @@ "string", "text/*" ], - "instillUIOrder": 4, + "instillUIOrder": 0, "instillUpstreamTypes": [ "value", "reference", @@ -307,6 +311,7 @@ }, "height": { "$ref": "stabilityai.json#/components/schemas/DiffuseImageHeight", + "description": "The image height", "instillAcceptFormats": [ "integer" ], @@ -323,7 +328,7 @@ "array:string", "array:text/*" ], - "instillUIOrder": 0, + "instillUIOrder": 1, "instillUpstreamTypes": [ "reference" ], @@ -405,7 +410,7 @@ "array:number", "array:integer" ], - "instillUIOrder": 1, + "instillUIOrder": 2, "instillUpstreamTypes": [ "reference" ], @@ -418,10 +423,11 @@ }, "width": { "$ref": "stabilityai.json#/components/schemas/DiffuseImageWidth", + "description": "The image width", "instillAcceptFormats": [ "integer" ], - "instillUIOrder": 11, + "instillUIOrder": 5, "instillUpstreamTypes": [ "value", "reference"