7
7
# of weaviate. In accordance with Infra-as-code, you should pin this value
8
8
# down and only change it if you explicitly want to upgrade the Weaviate
9
9
# version.
10
- tag : 1.20 .0
10
+ tag : 1.21 .0
11
11
repo : semitechnologies/weaviate
12
12
# Image pull policy: https://kubernetes.io/docs/concepts/containers/images/#image-pull-policy
13
13
pullPolicy : IfNotPresent
@@ -367,7 +367,8 @@ modules:
367
367
# but in some situations it cannot reach the same level of accuracy as
368
368
# transformers-based models.
369
369
text2vec-contextionary :
370
- # disable if you want to use transformers or import or own vectors
370
+
371
+ # Enable deployment of this module
371
372
enabled : false
372
373
373
374
# The configuration below is ignored if enabled==false
@@ -416,12 +417,13 @@ modules:
416
417
# CUDA-enabled GPUs for optimal performance.
417
418
text2vec-transformers :
418
419
419
- # enable if you want to use transformers instead of the
420
- # text2vec-contextionary module
420
+ # Enable deployment of this module
421
421
enabled : false
422
+
422
423
# You can set directly an inference URL of this module without deploying it with this release.
423
424
# You can do so by setting a value for the `inferenceUrl` here AND by setting the `enable` to `false`
424
425
inferenceUrl : {}
426
+
425
427
# The configuration below is ignored if enabled==false
426
428
427
429
# replace with model of choice, see
@@ -448,11 +450,11 @@ modules:
448
450
# accordingly and you need to explicitly set GPU requests & limits below
449
451
enable_cuda : false
450
452
451
- # only used when cuda is enabled
453
+ # only used when CUDA is enabled
452
454
nvidia_visible_devices : all
453
455
nvidia_driver_capabilities : compute,utility
454
456
455
- # only used when cuda is enabled
457
+ # only used when CUDA is enabled
456
458
ld_library_path : /usr/local/nvidia/lib64
457
459
458
460
resources :
@@ -496,6 +498,7 @@ modules:
496
498
# You can do so by setting a value for the `inferenceUrl` here AND by setting the `enable` to `false`
497
499
inferenceUrl : {}
498
500
501
+ # The configuration below is ignored if enabled==false
499
502
tag : facebook-dpr-ctx_encoder-single-nq-base
500
503
repo : semitechnologies/transformers-inference
501
504
registry : docker.io
@@ -514,11 +517,11 @@ modules:
514
517
# accordingly and you need to explicitly set GPU requests & limits below
515
518
enable_cuda : false
516
519
517
- # only used when cuda is enabled
520
+ # only used when CUDA is enabled
518
521
nvidia_visible_devices : all
519
522
nvidia_driver_capabilities : compute,utility
520
523
521
- # only used when cuda is enabled
524
+ # only used when CUDA is enabled
522
525
ld_library_path : /usr/local/nvidia/lib64
523
526
524
527
resources :
@@ -550,6 +553,7 @@ modules:
550
553
# You can do so by setting a value for the `inferenceUrl` here AND by setting the `enable` to `false`
551
554
inferenceUrl : {}
552
555
556
+ # The configuration below is ignored if enabled==false
553
557
tag : facebook-dpr-question_encoder-single-nq-base
554
558
repo : semitechnologies/transformers-inference
555
559
registry : docker.io
@@ -568,11 +572,11 @@ modules:
568
572
# accordingly and you need to explicitly set GPU requests & limits below
569
573
enable_cuda : false
570
574
571
- # only used when cuda is enabled
575
+ # only used when CUDA is enabled
572
576
nvidia_visible_devices : all
573
577
nvidia_driver_capabilities : compute,utility
574
578
575
- # only used when cuda is enabled
579
+ # only used when CUDA is enabled
576
580
ld_library_path : /usr/local/nvidia/lib64
577
581
578
582
resources :
@@ -606,10 +610,15 @@ modules:
606
610
# If you want to run a different model that published ones you can follow the
607
611
# tutorial from here on how to create such a container: https://github.com/weaviate/t2v-gpt4all-models
608
612
text2vec-gpt4all :
613
+
614
+ # Enable deployment of this module
609
615
enabled : false
616
+
610
617
# You can set directly an inference URL of this module without deploying it with this release.
611
618
# You can do so by setting a value for the `inferenceUrl` here AND by setting the `enable` to `false`
612
619
inferenceUrl : {}
620
+
621
+ # The configuration below is ignored if enabled==false
613
622
tag : all-MiniLM-L6-v2
614
623
repo : semitechnologies/gpt4all-inference
615
624
registry : docker.io
@@ -721,9 +730,9 @@ modules:
721
730
# run with CUDA-enabled GPUs for optimal performance.
722
731
multi2vec-clip :
723
732
724
- # enable if you want to use transformers instead of the
725
- # text2vec-contextionary module
733
+ # Enable deployment of this module
726
734
enabled : false
735
+
727
736
# You can set directly an inference URL of this module without deploying it with this release.
728
737
# You can do so by setting a value for the `inferenceUrl` here AND by setting the `enable` to `false`
729
738
inferenceUrl : {}
@@ -752,11 +761,11 @@ modules:
752
761
# accordingly and you need to explicitly set GPU requests & limits below
753
762
enable_cuda : false
754
763
755
- # only used when cuda is enabled
764
+ # only used when CUDA is enabled
756
765
nvidia_visible_devices : all
757
766
nvidia_driver_capabilities : compute,utility
758
767
759
- # only used when cuda is enabled
768
+ # only used when CUDA is enabled
760
769
ld_library_path : /usr/local/nvidia/lib64
761
770
762
771
resources :
@@ -794,7 +803,9 @@ modules:
794
803
# https://github.com/weaviate/multi2vec-bind-inference
795
804
multi2vec-bind :
796
805
806
+ # Enable deployment of this module
797
807
enabled : false
808
+
798
809
# You can set directly an inference URL of this module without deploying it with this release.
799
810
# You can do so by setting a value for the `inferenceUrl` here AND by setting the `enable` to `false`
800
811
inferenceUrl : {}
@@ -819,11 +830,11 @@ modules:
819
830
# accordingly and you need to explicitly set GPU requests & limits below
820
831
enable_cuda : false
821
832
822
- # only used when cuda is enabled
833
+ # only used when CUDA is enabled
823
834
nvidia_visible_devices : all
824
835
nvidia_driver_capabilities : compute,utility
825
836
826
- # only used when cuda is enabled
837
+ # only used when CUDA is enabled
827
838
ld_library_path : /usr/local/nvidia/lib64
828
839
829
840
resources :
@@ -855,12 +866,17 @@ modules:
855
866
tolerations :
856
867
857
868
# The qna-transformers module uses neural networks, such as BERT,
858
- # DistilBERT, to find an aswer in text to a given question
869
+ # DistilBERT, to find an answer in text to a given question
859
870
qna-transformers :
871
+
872
+ # Enable deployment of this module
860
873
enabled : false
874
+
861
875
# You can set directly an inference URL of this module without deploying it with this release.
862
876
# You can do so by setting a value for the `inferenceUrl` here AND by setting the `enable` to `false`
863
877
inferenceUrl : {}
878
+
879
+ # The configuration below is ignored if enabled==false
864
880
tag : bert-large-uncased-whole-word-masking-finetuned-squad-34d66b1
865
881
repo : semitechnologies/qna-transformers
866
882
registry : docker.io
@@ -880,11 +896,11 @@ modules:
880
896
# accordingly and you need to explicitly set GPU requests & limits below
881
897
enable_cuda : false
882
898
883
- # only used when cuda is enabled
899
+ # only used when CUDA is enabled
884
900
nvidia_visible_devices : all
885
901
nvidia_driver_capabilities : compute,utility
886
902
887
- # only used when cuda is enabled
903
+ # only used when CUDA is enabled
888
904
ld_library_path : /usr/local/nvidia/lib64
889
905
890
906
resources :
@@ -977,10 +993,14 @@ modules:
977
993
# The img2vec-neural module uses neural networks, to generate
978
994
# a vector representation of the image
979
995
img2vec-neural :
996
+
997
+ # Enable deployment of this module
980
998
enabled : false
999
+
981
1000
# You can set directly an inference URL of this module without deploying it with this release.
982
1001
# You can do so by setting a value for the `inferenceUrl` here AND by setting the `enable` to `false`
983
1002
inferenceUrl : {}
1003
+
984
1004
tag : resnet50
985
1005
repo : semitechnologies/img2vec-pytorch
986
1006
registry : docker.io
@@ -1000,11 +1020,11 @@ modules:
1000
1020
# accordingly and you need to explicitly set GPU requests & limits below
1001
1021
enable_cuda : false
1002
1022
1003
- # only used when cuda is enabled
1023
+ # only used when CUDA is enabled
1004
1024
nvidia_visible_devices : all
1005
1025
nvidia_driver_capabilities : compute,utility
1006
1026
1007
- # only used when cuda is enabled
1027
+ # only used when CUDA is enabled
1008
1028
ld_library_path : /usr/local/nvidia/lib64
1009
1029
1010
1030
resources :
@@ -1058,10 +1078,15 @@ modules:
1058
1078
# More information about Cross-Encoders can be found here:
1059
1079
# https://www.sbert.net/examples/applications/cross-encoder/README.html
1060
1080
reranker-transformers :
1081
+
1082
+ # Enable deployment of this module
1061
1083
enabled : false
1084
+
1062
1085
# You can set directly an inference URL of this module without deploying it with this release.
1063
1086
# You can do so by setting a value for the `inferenceUrl` here AND by setting the `enable` to `false`
1064
1087
inferenceUrl : {}
1088
+
1089
+ # The configuration below is ignored if enabled==false
1065
1090
tag : cross-encoder-ms-marco-MiniLM-L-6-v2
1066
1091
repo : semitechnologies/reranker-transformers
1067
1092
registry : docker.io
@@ -1081,11 +1106,11 @@ modules:
1081
1106
# accordingly and you need to explicitly set GPU requests & limits below
1082
1107
enable_cuda : false
1083
1108
1084
- # only used when cuda is enabled
1109
+ # only used when CUDA is enabled
1085
1110
nvidia_visible_devices : all
1086
1111
nvidia_driver_capabilities : compute,utility
1087
1112
1088
- # only used when cuda is enabled
1113
+ # only used when CUDA is enabled
1089
1114
ld_library_path : /usr/local/nvidia/lib64
1090
1115
1091
1116
resources :
@@ -1124,10 +1149,15 @@ modules:
1124
1149
# The text-spellcheck module uses spellchecker library to check
1125
1150
# misspellings in a given text
1126
1151
text-spellcheck :
1152
+
1153
+ # Enable deployment of this module
1127
1154
enabled : false
1155
+
1128
1156
# You can set directly an inference URL of this module without deploying it with this release.
1129
1157
# You can do so by setting a value for the `inferenceUrl` here AND by setting the `enable` to `false`
1130
1158
inferenceUrl : {}
1159
+
1160
+ # The configuration below is ignored if enabled==false
1131
1161
tag : pyspellchecker-en
1132
1162
repo : semitechnologies/text-spellcheck-model
1133
1163
registry : docker.io
@@ -1173,10 +1203,15 @@ modules:
1173
1203
# The ner-transformers module uses spellchecker library to check
1174
1204
# misspellings in a given text
1175
1205
ner-transformers :
1206
+
1207
+ # Enable deployment of this module
1176
1208
enabled : false
1209
+
1177
1210
# You can set directly an inference URL of this module without deploying it with this release.
1178
1211
# You can do so by setting a value for the `inferenceUrl` here AND by setting the `enable` to `false`
1179
1212
inferenceUrl : {}
1213
+
1214
+ # The configuration below is ignored if enabled==false
1180
1215
tag : dbmdz-bert-large-cased-finetuned-conll03-english-0.0.2
1181
1216
repo : semitechnologies/ner-transformers
1182
1217
registry : docker.io
@@ -1196,11 +1231,11 @@ modules:
1196
1231
# accordingly and you need to explicitly set GPU requests & limits below
1197
1232
enable_cuda : false
1198
1233
1199
- # only used when cuda is enabled
1234
+ # only used when CUDA is enabled
1200
1235
nvidia_visible_devices : all
1201
1236
nvidia_driver_capabilities : compute,utility
1202
1237
1203
- # only used when cuda is enabled
1238
+ # only used when CUDA is enabled
1204
1239
ld_library_path : /usr/local/nvidia/lib64
1205
1240
1206
1241
resources :
@@ -1238,10 +1273,15 @@ modules:
1238
1273
1239
1274
# The sum-transformers module makes result texts summarizations
1240
1275
sum-transformers :
1276
+
1277
+ # Enable deployment of this module
1241
1278
enabled : false
1279
+
1242
1280
# You can set directly an inference URL of this module without deploying it with this release.
1243
1281
# You can do so by setting a value for the `inferenceUrl` here AND by setting the `enable` to `false`
1244
1282
inferenceUrl : {}
1283
+
1284
+ # The configuration below is ignored if enabled==false
1245
1285
tag : facebook-bart-large-cnn-1.0.0
1246
1286
repo : semitechnologies/sum-transformers
1247
1287
registry : docker.io
@@ -1261,11 +1301,11 @@ modules:
1261
1301
# accordingly and you need to explicitly set GPU requests & limits below
1262
1302
enable_cuda : false
1263
1303
1264
- # only used when cuda is enabled
1304
+ # only used when CUDA is enabled
1265
1305
nvidia_visible_devices : all
1266
1306
nvidia_driver_capabilities : compute,utility
1267
1307
1268
- # only used when cuda is enabled
1308
+ # only used when CUDA is enabled
1269
1309
ld_library_path : /usr/local/nvidia/lib64
1270
1310
1271
1311
resources :
0 commit comments