Update README.md
Browse files
    	
        README.md
    CHANGED
    
    | 
         @@ -35,7 +35,8 @@ The following example starts at the root of D drive and quantizes mistral's Mixt 
     | 
|
| 35 | 
         | 
| 36 | 
         
             
            ## Download llama.cpp
         
     | 
| 37 | 
         
             
            Assuming you want CUDA for your NVIDIA RTX GPU(s) use the links below, or grab latest compiled executables from https://github.com/ggerganov/llama.cpp/releases
         
     | 
| 38 | 
         
            -
             
     | 
| 
         | 
|
| 39 | 
         
             
            * https://github.com/ggerganov/llama.cpp/releases/download/b2253/cudart-llama-bin-win-cu12.2.0-x64.zip
         
     | 
| 40 | 
         
             
            * https://github.com/ggerganov/llama.cpp/releases/download/b2253/llama-b2253-bin-win-cublas-cu12.2.0-x64.zip
         
     | 
| 41 | 
         | 
| 
         | 
|
| 35 | 
         | 
| 36 | 
         
             
            ## Download llama.cpp
         
     | 
| 37 | 
         
             
            Assuming you want CUDA for your NVIDIA RTX GPU(s) use the links below, or grab latest compiled executables from https://github.com/ggerganov/llama.cpp/releases
         
     | 
| 38 | 
         
            +
             
     | 
| 39 | 
         
            +
            ### Latest version as of Feb 24, 2024:
         
     | 
| 40 | 
         
             
            * https://github.com/ggerganov/llama.cpp/releases/download/b2253/cudart-llama-bin-win-cu12.2.0-x64.zip
         
     | 
| 41 | 
         
             
            * https://github.com/ggerganov/llama.cpp/releases/download/b2253/llama-b2253-bin-win-cublas-cu12.2.0-x64.zip
         
     | 
| 42 | 
         |