From a92ec26f99c603e0a26741992eb26c0b124036a4 Mon Sep 17 00:00:00 2001 From: Ettore Di Giacinto Date: Sat, 11 May 2024 09:42:12 +0200 Subject: [PATCH] models(gallery): add Llama-3-8B-Instruct-abliterated Signed-off-by: Ettore Di Giacinto --- gallery/index.yaml | 13 +++++++++++++ 1 file changed, 13 insertions(+) diff --git a/gallery/index.yaml b/gallery/index.yaml index 051f6fc5eb0..3103a39ce3d 100644 --- a/gallery/index.yaml +++ b/gallery/index.yaml @@ -95,6 +95,19 @@ - filename: Meta-Llama-3-8B-Instruct.Q6_K.gguf sha256: b7bad45618e2a76cc1e89a0fbb93a2cac9bf410e27a619c8024ed6db53aa9b4a uri: huggingface://QuantFactory/Meta-Llama-3-8B-Instruct-GGUF/Meta-Llama-3-8B-Instruct.Q6_K.gguf +- !!merge <<: *llama3 + name: "llama-3-8b-instruct-abliterated" + urls: + - https://huggingface.co/failspy/Llama-3-8B-Instruct-abliterated-GGUF + description: | + This is meta-llama/Llama-3-8B-Instruct with orthogonalized bfloat16 safetensor weights, generated with the methodology that was described in the preview paper/blog post: 'Refusal in LLMs is mediated by a single direction' which I encourage you to read to understand more. + overrides: + parameters: + model: Llama-3-8B-Instruct-abliterated-q4_k.gguf + files: + - filename: Llama-3-8B-Instruct-abliterated-q4_k.gguf + sha256: a6365f813de1977ae22dbdd271deee59f91f89b384eefd3ac1a391f391d8078a + uri: huggingface://failspy/Llama-3-8B-Instruct-abliterated-GGUF/Llama-3-8B-Instruct-abliterated-q4_k.gguf - !!merge <<: *llama3 name: "llama-3-8b-instruct-coder" icon: https://cdn-uploads.huggingface.co/production/uploads/642cc1c253e76b4c2286c58e/0O4cIuv3wNbY68-FP7tak.jpeg