Skip to content

Commit 686da8e

Browse files
Merge pull request #4 from ritual-net/sr/feat/update_services_2.0.0
update services to v2.0.0
2 parents a358cae + 7fbcf7a commit 686da8e

File tree

12 files changed

+43
-39
lines changed

12 files changed

+43
-39
lines changed

node/anvil/1.3.0/config.json

+1-1
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
{
22
"config": {
3-
"forward_stats": false,
3+
"forward_stats": true,
44
"manage_containers": true,
55
"startup_wait": 1.0,
66
"chain": {

node/other/1.3.0/config.json

+1-1
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
{
22
"config": {
3-
"forward_stats": false,
3+
"forward_stats": true,
44
"manage_containers": true,
55
"startup_wait": 1.0,
66
"chain": {
File renamed without changes.

services/css-inference/1.0.0/recipe.json services/css-inference/2.0.0/recipe.json

+2-2
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
{
22
"config": {
3-
"id": "css-inference-1.0.0",
4-
"image": "ritualnetwork/css_inference_service:1.0.0",
3+
"id": "css-inference-2.0.0",
4+
"image": "ritualnetwork/css_inference_service:2.0.0",
55
"description": "Closed Source Inference Service",
66
"command": "--bind=0.0.0.0:3000 --workers=${NUM_WORKERS}",
77
"env": {},

services/hf-client-inference/1.0.0/recipe.json services/hf-client-inference/2.0.0/recipe.json

+6-6
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
{
22
"config": {
3-
"id": "hf-client-inference-1.0.0",
4-
"image": "ritualnetwork/hf_inference_client_service:1.0.0",
3+
"id": "hf-client-inference-2.0.0",
4+
"image": "ritualnetwork/hf_inference_client_service:2.0.0",
55
"description": "HuggingFace Client Inference Service",
66
"command": "--bind=0.0.0.0:3000 --workers=${NUM_WORKERS}",
77
"env": {},
@@ -16,11 +16,11 @@
1616
},
1717
"inputs": [
1818
{
19-
"id": "HF_TOKEN",
20-
"path": "env.HF_TOKEN",
21-
"description": "The Hugging Face API token.",
19+
"id": "HF_INF_TOKEN",
20+
"path": "env.HF_INF_TOKEN",
21+
"description": "The Hugging Face API token. Used for increased rate limits and access to private models.",
2222
"type": "string",
23-
"required": true
23+
"required": false
2424
},
2525
{
2626
"id": "NUM_WORKERS",

services/onnx-inference/1.0.0/recipe.json services/onnx-inference/2.0.0/recipe.json

+8-8
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
{
22
"config": {
3-
"id": "onnx-inference-1.0.0",
4-
"image": "ritualnetwork/onnx_inference_service:1.0.0",
3+
"id": "onnx-inference-2.0.0",
4+
"image": "ritualnetwork/onnx_inference_service:2.0.0",
55
"description": "ONNX Inference Service",
66
"command": "--bind=0.0.0.0:3000 --workers=${NUM_WORKERS}",
77
"env": {},
@@ -16,16 +16,16 @@
1616
},
1717
"inputs": [
1818
{
19-
"id": "MODEL_SOURCE",
20-
"path": "env.MODEL_SOURCE",
21-
"description": "The source of the model to load. Used for preloading a model.",
19+
"id": "ONNX_DEFAULT_MODEL_ID",
20+
"path": "env.ONNX_DEFAULT_MODEL_ID",
21+
"description": "The default model to preload on startup.",
2222
"type": "string",
2323
"required": false
2424
},
2525
{
26-
"id": "LOAD_ARGS",
27-
"path": "env.LOAD_ARGS",
28-
"description": "Arguments to load model with, as a stringified JSON object. Used for preloading a model.",
26+
"id": "ONNX_CACHE_DIR",
27+
"path": "env.ONNX_CACHE_DIR",
28+
"description": "The directory to cache ONNX models in.",
2929
"type": "string",
3030
"required": false
3131
},

services/tgi-client-inference/1.0.0/README.md services/tgi-client-inference/2.0.0/README.md

+2-1
Original file line numberDiff line numberDiff line change
@@ -2,6 +2,7 @@
22

33
### Description
44

5-
This service enables ML inference requests via HuggingFace's [Text Generation Inference (TGI)](https://huggingface.co/docs/text-generation-inference/en/index) API, which serves Large Language Models (LLMs).
5+
This service connects to a specified [Text Generation Inference (TGI)](https://huggingface.co/docs/text-generation-inference/en/index) instance and serves Large Language Model (LLMs) inference requests
6+
via the TGI REST interface.
67

78
For configuration and usage details, check out the [TGI Client Inference Service](https://infernet-services.docs.ritual.net/reference/tgi_client_inference_service) documentation.

services/tgi-client-inference/1.0.0/recipe.json services/tgi-client-inference/2.0.0/recipe.json

+13-10
Original file line numberDiff line numberDiff line change
@@ -1,11 +1,14 @@
11
{
22
"config": {
3-
"id": "tgi-client-inference-1.0.0",
4-
"image": "ritualnetwork/tgi_client_inference_service:1.0.0",
3+
"id": "tgi-client-inference-2.0.0",
4+
"image": "ritualnetwork/tgi_client_inference_service:2.0.0",
55
"description": "TGI Client Inference Service",
66
"command": "--bind=0.0.0.0:3000 --workers=${NUM_WORKERS}",
77
"env": {
8-
"TGI_INF_WORKFLOW_POSITIONAL_ARGS": "[\"${TGI_SERVICE_URL}\", ${CONNECTION_TIMEOUT}, {\"Authorization\": \"Bearer ${HF_TOKEN}\"}]"
8+
"TGI_INF_WORKFLOW_POSITIONAL_ARGS": [
9+
"${SERVICE_URL}",
10+
"${CONNECTION_TIMEOUT}"
11+
]
912
},
1013
"external": true,
1114
"gpu": false,
@@ -18,16 +21,16 @@
1821
},
1922
"inputs": [
2023
{
21-
"id": "TGI_SERVICE_URL",
22-
"path": "env.TGI_INF_WORKFLOW_POSITIONAL_ARGS#TGI_SERVICE_URL",
23-
"description": "The URL of the TGI service to connect to.",
24+
"id": "TGI_INF_TOKEN",
25+
"path": "env.TGI_INF_TOKEN",
26+
"description": "The Hugging Face API token. Used for increased rate limits and access to private models.",
2427
"type": "string",
25-
"required": true
28+
"required": false
2629
},
2730
{
28-
"id": "HF_TOKEN",
29-
"path": "env.TGI_INF_WORKFLOW_POSITIONAL_ARGS#HF_TOKEN",
30-
"description": "The Hugging Face API token.",
31+
"id": "SERVICE_URL",
32+
"path": "env.TGI_INF_WORKFLOW_POSITIONAL_ARGS#SERVICE_URL",
33+
"description": "The URL of the TGI service to connect to.",
3134
"type": "string",
3235
"required": true
3336
},

services/torch-inference/1.0.0/recipe.json services/torch-inference/2.0.0/recipe.json

+10-10
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
{
22
"config": {
3-
"id": "torch-inference-1.0.0",
4-
"image": "ritualnetwork/torch_inference_service:1.0.0",
3+
"id": "torch-inference-2.0.0",
4+
"image": "ritualnetwork/torch_inference_service:2.0.0",
55
"description": "PyTorch Inference Service",
66
"command": "--bind=0.0.0.0:3000 --workers=${NUM_WORKERS}",
77
"env": {},
@@ -16,22 +16,22 @@
1616
},
1717
"inputs": [
1818
{
19-
"id": "MODEL_SOURCE",
20-
"path": "env.MODEL_SOURCE",
21-
"description": "The source of the model to load. Used for preloading a model.",
19+
"id": "TORCH_DEFAULT_MODEL_ID",
20+
"path": "env.TORCH_DEFAULT_MODEL_ID",
21+
"description": "The default model to preload on startup.",
2222
"type": "string",
2323
"required": false
2424
},
2525
{
26-
"id": "LOAD_ARGS",
27-
"path": "env.LOAD_ARGS",
28-
"description": "Arguments to load model with, as a stringified JSON object. Used for preloading a model.",
26+
"id": "TORCH_CACHE_DIR",
27+
"path": "env.TORCH_CACHE_DIR",
28+
"description": "The directory to cache Torch models in.",
2929
"type": "string",
3030
"required": false
3131
},
3232
{
33-
"id": "USE_JIT",
34-
"path": "env.USE_JIT",
33+
"id": "TORCH_USE_JIT",
34+
"path": "env.TORCH_USE_JIT",
3535
"description": "Whether to use JIT for model loading.",
3636
"type": "string",
3737
"required": false,

0 commit comments

Comments
 (0)