diff --git a/garak/resources/plugin_cache.json b/garak/resources/plugin_cache.json index bd370417..c6c03123 100644 --- a/garak/resources/plugin_cache.json +++ b/garak/resources/plugin_cache.json @@ -6000,7 +6000,7 @@ }, "parallel_capable": true, "supports_multiple_generations": true, - "mod_time": "2024-08-29 13:35:37 +0000" + "mod_time": "2024-11-12 21:52:33 +0000" }, "generators.function.Single": { "description": "pass a module#function to be called as generator, with format function(prompt:str, **kwargs)->List[Union(str, None)] the parameter `name` is reserved", @@ -6019,7 +6019,7 @@ }, "parallel_capable": true, "supports_multiple_generations": false, - "mod_time": "2024-08-29 13:35:37 +0000" + "mod_time": "2024-11-12 21:52:33 +0000" }, "generators.ggml.GgmlGenerator": { "description": "Generator interface for ggml models in gguf format.", @@ -6048,7 +6048,7 @@ }, "parallel_capable": true, "supports_multiple_generations": false, - "mod_time": "2024-08-29 13:35:37 +0000" + "mod_time": "2024-11-12 21:52:33 +0000" }, "generators.groq.GroqChat": { "description": "Wrapper for Groq-hosted LLM models.", @@ -6139,7 +6139,7 @@ }, "parallel_capable": false, "supports_multiple_generations": true, - "mod_time": "2024-11-11 14:12:52 +0000" + "mod_time": "2024-11-18 18:23:48 +0000" }, "generators.huggingface.InferenceAPI": { "description": "Get text generations from Hugging Face Inference API", @@ -6164,7 +6164,7 @@ }, "parallel_capable": true, "supports_multiple_generations": true, - "mod_time": "2024-11-11 14:12:52 +0000" + "mod_time": "2024-11-18 18:23:48 +0000" }, "generators.huggingface.InferenceEndpoint": { "description": "Interface for Hugging Face private endpoints", @@ -6189,7 +6189,7 @@ }, "parallel_capable": true, "supports_multiple_generations": false, - "mod_time": "2024-11-11 14:12:52 +0000" + "mod_time": "2024-11-18 18:23:48 +0000" }, "generators.huggingface.LLaVA": { "description": "Get LLaVA ([ text + image ] -> text) generations", @@ -6217,7 +6217,7 @@ }, "parallel_capable": false, "supports_multiple_generations": false, - "mod_time": "2024-11-11 14:12:52 +0000" + "mod_time": "2024-11-18 18:23:48 +0000" }, "generators.huggingface.Model": { "description": "Get text generations from a locally-run Hugging Face model", @@ -6244,7 +6244,7 @@ }, "parallel_capable": false, "supports_multiple_generations": true, - "mod_time": "2024-11-11 14:12:52 +0000" + "mod_time": "2024-11-18 18:23:48 +0000" }, "generators.huggingface.OptimumPipeline": { "description": "Get text generations from a locally-run Hugging Face pipeline using NVIDIA Optimum", @@ -6271,7 +6271,7 @@ }, "parallel_capable": false, "supports_multiple_generations": true, - "mod_time": "2024-11-11 14:12:52 +0000" + "mod_time": "2024-11-18 18:23:48 +0000" }, "generators.huggingface.Pipeline": { "description": "Get text generations from a locally-run Hugging Face pipeline", @@ -6298,7 +6298,7 @@ }, "parallel_capable": false, "supports_multiple_generations": true, - "mod_time": "2024-11-11 14:12:52 +0000" + "mod_time": "2024-11-18 18:23:48 +0000" }, "generators.langchain.LangChainLLMGenerator": { "description": "Class supporting LangChain LLM interfaces", @@ -7024,7 +7024,7 @@ "active": true, "bcp47": null, "doc_uri": "", - "mod_time": "2024-10-25 09:35:40 +0000" + "mod_time": "2024-11-12 21:52:33 +0000" }, "buffs.encoding.Base64": { "description": "Base64 buff",