Skip to content

Script for Quantizing_with_LlamaCpp AI models for Slackware64-current

Notifications You must be signed in to change notification settings

rizitis/Quantizing_with_LlamaCpp

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

29 Commits
 
 
 
 
 
 

Repository files navigation

EDIT 24-11-2024

Sackware-current now use python 3.12.7, but i did not upgrade my system so far.
So I have not test yet script for new python version.

Script is based on GratisStudio HowTo for windows. It is tested on Slackware64 current systems without issues. If you found a bug please open an issue.

Reqirements to run script:

python3.11--> {numpy,sentencepiece,gguf}

GPT4All or ollama (if you use ollama gguf models can work whith zed IDE)
git lfs
   ========= OPTIONAL:=============
     Vulkan SDK (AMD GPU Support)
     Cuda toolkit (Nvidia GPU Support)
     ********************************

Normally all other needs should be by default in your distro, if not..when script fail read what is missing and install from your distro package manager.

USAGE

  1. Make script executable if not...
chmod +x quantizing_ai_models.sh
  1. Find the LL model you want from https://huggingface.co
    Copy ONLY the provided url for git clone, example:
    copy url
  2. Now exexute script in terminal following by model url example:
./quantizing_ai_models.sh https://huggingface.co/Cadenza-Labs/dolphin-llama3-8B-sleeper-agent-standard-l
  1. Just answer questions when needed and wait for results...

  2. If you have success 👊 you can now load your model.gguf using gpt4all app or ollama.

NOTE:

Warning: Running this script as root is not recommended.

Supported models

mistral llama llama3 ggml

About

Script for Quantizing_with_LlamaCpp AI models for Slackware64-current

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages