EnlistedGhost commited on
Commit
58b90dd
·
verified ·
1 Parent(s): c7802e7

Upload Modelfile-Pixtral-12B-Q6_K.md

Browse files
Files changed (1) hide show
  1. Modelfile-Pixtral-12B-Q6_K.md +71 -0
Modelfile-Pixtral-12B-Q6_K.md ADDED
@@ -0,0 +1,71 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Pixtral-12B-GGUF Modelfile (Q2_K)
2
+ # ---------------------------------
3
+ #
4
+ # Tested with: Ollama v0.11.X-->v0.12.6(latest)
5
+ # Quantization: Q6_K (Quant created by = mradermacher)
6
+ # Quality: Very-Good: Reccomende for users with -
7
+ # more RAM than Q4_K_M requires.
8
+ # ----------------------------------------------------
9
+ #
10
+ # Vision Notes:
11
+ # Some users may need to set the context value -or- "num_ctx"
12
+ # value to ~9K-->19K.
13
+ # Personally tested with: num_ctx=9982 and num_ctx=19982
14
+ # -----------------------------------------------------------
15
+ #
16
+ # Created by:
17
+ # EnlistedGhost (aka Jon Zaretsky)
18
+ # Original GGUF by: https://huggingface.co/mradermacher
19
+ # Original GGUF type: Static Quantize (non-iMatrix)
20
+ # ----------------------------------------------------------
21
+ # | Warning! - iMatrix Quantize seems to suffer in regards |
22
+ # | to vision quality, but is still made available |
23
+ # ----------------------------------------------------------
24
+ #
25
+ # Goal:
26
+ # To provide the FIRST actually functional and usable
27
+ # GGUF model version of the Mistral Pixtral-12B for
28
+ # direct-usage with Ollama!
29
+ # Currently, there are NO USABLE OR WORKING versions
30
+ # of this model...
31
+ # ---------------------------------------------------
32
+ #
33
+ # Big/Giant/Huge Thank You:
34
+ # (ggml-org, bartowski, and the Ollama team)
35
+ # ggml-org: Working mmproj-pixtral vision projector!
36
+ # Bartowki: Working I-Matrix Quants that can be paired with ggml-org vision projector!
37
+ # Ollama team: Because without them, this wouldn't be possible in the first place!
38
+ # ------------------------------------------------------------------------------------
39
+ #
40
+ # Import our GGUF quant files:
41
+ # (Assuming: Linux Operating System)
42
+ # (Assuming: downloaded files are stored in "Downloads" directory/folder)
43
+ FROM ~/Downloads/mmproj-pixtral-12b-f16.gguf
44
+ FROM ~/Downloads/Pixtral-12B-Q8_0.gguf
45
+ # ------------------------------------------------------------------------
46
+ #
47
+ # Set Default System-Message/Prompt:
48
+ SYSTEM """
49
+ #
50
+ # !!!-WARNING-!!!
51
+ # (Do not modify for: "recommended" configuration and behavior)
52
+ #
53
+ # !!!-OPTIONAL-!!!
54
+ # Pixtral-12B by default does NOT include a system-prompt, however, you can choose to input one within this section of the Ollama-Modelfile. Please be aware that you can possibly damage the linking between the Pixtral-->VisionProjector within the system-prompt field; BE CAREFUL!
55
+ """
56
+ # -------------------------------------------------------------------
57
+ #
58
+ # Define model-chat template (Thank you to: @rick-github for this mic-drop)
59
+ # Link to @rick-github post: https://github.com/ollama/ollama/issues/6748#issuecomment-3368146231
60
+ TEMPLATE [INST] {{ if .System }}{{ .System }} {{ end }}{{ .Prompt }} [/INST]
61
+ #
62
+ # Below are stop params (required for proper "assistant-->user" multi-turn)
63
+ PARAMETER stop [INST]
64
+ PARAMETER stop [/INST]
65
+ #
66
+ # Enjoy Pixtral-12B-GGUF for the ppl!
67
+ # Erm, or at least for Ollama users...
68
+ # <3 (^.^) <3
69
+ #
70
+ # Notice: Please, read the "Instructions.md" at HuggingFace or Ollama-Website
71
+ # for a how-to usage and guide on using this modelfile!