Model,Quantization,GPU(s),CPU(s),RAM (GB),Context Length,Total Tokens,Duration (sec),Tokens/sec,Time to Retrieve (sec),Start Type,GPU Max Power (W),Server Name,Test Link,Comment llama3.1:8b-instruct,q4_0,"1x GTX 1050Ti 4GB","1x Intel i7",32,4000,243,37.38,6.50,105.62,Cold,-,pasha-lt,-,- llama3.1:8b-instruct,q4_0,"1x GTX 1050Ti 4GB","1x Intel i7",32,4000,248,38.20,6.49,38.48,Warm,-,pasha-lt,-,- llama3.1:8b-instruct,q4_0,"1x RTX 3050 8GB","2x Xeon 2960v2",128,4000,234,6.03,38.83,9.27,Cold,-,gpu01,-,- llama3.1:8b-instruct,q4_0,"1x RTX 3050 8GB","2x Xeon 2960v2",128,4000,303,7.82,38.73,7.97,Warm,-,gpu01,-,- llama3.1:70b-instruct,q4_0,"1x RTX 3050 8GB","2x Xeon 2960v2",128,4000,500,608.09,0.82,628.27,Cold,-,gpu01,-,- llama3.1:70b-instruct,q4_0,"1x RTX 3050 8GB","2x Xeon 2960v2",128,4000,500,608.47,0.82,609.85,Warm,-,gpu01,-,- llama3.1:8b-instruct,q4_0,"1x RTX 4090 24GB","1x Ryzen 9 5950x",128,4000,282,1.95,144.66,44.73,Cold,-,gpu02,-,- llama3.1:8b-instruct,q4_0,"1x RTX 4090 24GB","1x Ryzen 9 5950x",128,4000,316,2.19,144.62,2.32,Warm,-,gpu02,-,- llama3.1:70b-instruct,q4_0,"1x RTX 4090 24GB","1x Ryzen 9 5950x",128,4000,500,308.62,1.62,412.98,Cold,-,gpu02,-,- llama3.1:70b-instruct,q4_0,"1x RTX 4090 24GB","1x Ryzen 9 5950x",128,4000,500,297.64,1.68,305.18,Cold,-,gpu02,-,- llama3.1:70b-instruct,q4_0,"1x RTX 4090 24GB","1x Ryzen 9 5950x",128,4000,500,297.75,1.68,298.44,Warm,-,gpu02,-,- llama3.1:8b-instruct,fp16,"1x Tesla P40 24GB",1x Xeon E5-2680 v4,128,4000,229,14.26,16.06,38.33,-,135,nn01,https://t.me/evilfreelancer_chat/73576,- llama3.1:8b-instruct,q4_0,"1x Tesla P40 24GB",1x Xeon E5-2680 v4,128,4000,345,8.39,41.14,8.54,-,135,nn01,https://t.me/evilfreelancer_chat/73622,- llama3.1:8b-instruct,q4_0,"1x Tesla P40 24GB",1x Xeon E5-2680 v4,128,4000,225,5.07,44.39,7.74,-,250,nn01,https://t.me/evilfreelancer_chat/73668,- llama3.1:8b-instruct,q4_0,1x RTX 4060 16GB,1x Intel i7,16,4000,471,10.15,46.40,15.33,-,-,sn01,https://t.me/evilfreelancer_chat/73684,- aya:35b-23,q4_K_M,"1x Tesla P40 24GB",1x Xeon E5-2680 v4,128,200,300,30.32,9.90,39.38,-,-,nn01,https://t.me/evilfreelancer_chat/73749,- aya:35b-23,q4_K_M,3x RTX 4060 16GB,2х Xeon E5-2690 v1,128,4000,327,28.23,11.59,38.81,Cold,-,sn02,https://t.me/evilfreelancer_chat/73748,- aya:35b-23,q4_K_M,3x RTX 4060 16GB,2х Xeon E5-2690 v1,128,4000,324,27.93,11.60,28.22,Warm,-,sn02,https://t.me/evilfreelancer_chat/73748,- aya:35b-23,q4_K_M,3x RTX 4060 16GB,2х Xeon E5-2690 v1,128,200,307,26.39,11.63,45.27,-,-,sn02,https://t.me/evilfreelancer_chat/73762,- aya:35b-23,q4_K_M,"1x RTX 4090 24GB","1x Ryzen 9 5950x",128,200,325,8.46,38.44,13.19,Cold,-,gpu02,https://t.me/evilfreelancer_chat/73763,- llama3.1:70b-instruct,q4_0,1x RTX 8000 48GB,1x i9-13900KF,-,-,-,-,13.1,-,-,-,mn1,https://t.me/evilfreelancer_chat/74682,- llama3.1:8b-instruct,q4_0,1x RTX 8000 48GB,1x i9-13900KF,-,-,-,-,90,-,-,-,mn1,https://t.me/evilfreelancer_chat/74682,- mistral-small:22b-instruct-2409,q4_0,1x RTX 4060 16GB,1x Intel i7,16,200,500,24.28,20.59,29.88,Cold,-,sn01,-,- mistral-small:22b-instruct-2409,q4_0,1x RTX 4060 16GB,1x Intel i7,16,200,500,24.22,20.63,24.42,Warm,-,sn01,-,- mistral-small:22b-instruct-2409,q4_0,1x RTX 4090 24GB,1x Ryzen 9 5950x,128,200,500,8.24,60.63,10.16,Cold,-,gpu02,https://t.me/evilfreelancer_chat/75209,- mistral-small:22b-instruct-2409,q4_0,1x RTX 4090 24GB,1x Ryzen 9 5950x,128,200,500,8.24,60.6,8.36,Warm,-,gpu02,https://t.me/evilfreelancer_chat/75209,- mixtral:8x7b-instruct-v0.1,q4_0,3x RTX 4060 16GB,2х Xeon E5-2690 v1,128,200,367,11.77,31.16,119.34,Cold,-,sn02,https://t.me/evilfreelancer_chat/75228,- mixtral:8x7b-instruct-v0.1,q4_0,3x RTX 4060 16GB,2х Xeon E5-2690 v1,128,200,471,15.09,31.19,15.28,Warm,-,sn02,https://t.me/evilfreelancer_chat/75228,- mixtral:8x7b-instruct-v0.1,q4_0,1x RTX 4090 24GB,1x Ryzen 9 5950x,128,200,313,15.2,20.58,20.11,Cold,-,gpu02,-,- mixtral:8x7b-instruct-v0.1,q4_0,1x RTX 4090 24GB,1x Ryzen 9 5950x,128,200,500,24.42,20.47,24.76,Warm,-,gpu02,-,- llama3.1:70b-instruct,q8_0,2x RTX 8000 48GB,1x i9-13900KF,-,-,-,-,7,-,-,-,mn1,https://t.me/evilfreelancer_chat/75552,- mistral-small:22b-instruct-2409,q4_0,1x RTX 8000 48GB,1x i9-13900KF,-,-,-,-,37,-,-,-,mn1,https://t.me/evilfreelancer_chat/75557,- llama3.1:70b-instruct,q8_0,4x RTX 4090 24GB,-,-,-,-,-,11,-,-,-,mn2,https://t.me/evilfreelancer_chat/75724,- llama3.1:8b-instruct,fp16,4x RTX 4090 24GB,-,-,-,-,-,51,-,-,-,mn2,https://t.me/evilfreelancer_chat/75741,- llama3.1:8b-instruct,q4_0,-,1x Xeon E5-2680 v4,-,-,238,38.3,6.21,38.59,Warm,-,nn01,https://t.me/evilfreelancer_chat/75771,- llama3.1:8b-instruct,q4_0,"1x RTX 3090 Ti 24GB 1x RTX 3090 24GB",1x Threadripper 1920X,-,4000,275,2.55,107.8,5.21,-,-,tig1,https://t.me/evilfreelancer_chat/75772,- llama3.1:8b-instruct,fp16,"1x RTX 3090 Ti 24GB 1x RTX 3090 24GB",1x Threadripper 1920X,-,4000,229,5.05,45.31,10.17,-,-,tig1,https://t.me/evilfreelancer_chat/75772,- llama3.1:70b-instruct,q4_0,"1x RTX 3090 Ti 24GB 1x RTX 3090 24GB",1x Threadripper 1920X,-,4000,496,26.25,18.89,37.34,-,-,tig1,https://t.me/evilfreelancer_chat/75772,- aya:35b-23,q4_K_M,"1x RTX 3090 Ti 24GB 1x RTX 3090 24GB",1x Threadripper 1920X,-,4000,368,11.34,32.44,20.14,-,-,tig1,https://t.me/evilfreelancer_chat/75772,- llama3.1:8b-instruct,q4_0,"1x RTX 3090 Ti 24GB 1x RTX 3090 24GB",1x Threadripper 1920X,-,200,260,2.35,110.41,4.82,-,-,tig1,https://t.me/evilfreelancer_chat/75779,- llama3.1:8b-instruct,fp16,"1x RTX 3090 Ti 24GB 1x RTX 3090 24GB",1x Threadripper 1920X,-,200,242,5.3,45.64,10.39,-,-,tig1,https://t.me/evilfreelancer_chat/75779,- llama3.1:70b-instruct,q4_0,"1x RTX 3090 Ti 24GB 1x RTX 3090 24GB",1x Threadripper 1920X,-,200,354,18.65,18.97,31.69,-,-,tig1,https://t.me/evilfreelancer_chat/75779,- aya:35b-23,q4_K_M,"1x RTX 3090 Ti 24GB 1x RTX 3090 24GB",1x Threadripper 1920X,-,200,321,9.54,33.63,18.58,-,-,tig1,https://t.me/evilfreelancer_chat/75779,- llama3.1:8b-instruct,q4_0,1x RX 7900 XTX 24GB,-,-,200,500,5.47,91.30,11.36,-,-,fp1,https://t.me/evilfreelancer_chat/76808,- llama3.1:8b-instruct,q4_0,5x GTX 1060 6GB,1x i5-4570,8,200,244,9.4,25.94,19.38,-,-,sh1,https://t.me/evilfreelancer_chat/81926,- llama3.1:8b-instruct,q4_0,1x Intel Arc A770 16GB,-,-,200,372,7.35992,50.544,23.2908,-,-,fp2,https://t.me/c/1564516454/155938,"Windows, драйвер 6795 ipex-llm (sycl)" llama3.1:8b-instruct,q4_0,1x Intel Arc A770 16GB,-,-,4000,500,9.94434,50.2799,24.0089,-,-,fp2,https://t.me/c/1564516454/155938,"Windows, драйвер 6795 ipex-llm (sycl)" llama3.1:8b-instruct,q4_0,1x Intel Arc A770 16GB,-,-,200,372,6.44538,57.7158,17.9413,-,-,fp2,https://t.me/c/1564516454/155939,"ubuntu 25.04 (6.14) ipex-llm (sycl) bigdl-core-cpp 2.7.0b20250519 ipex-llm 2.3.0b20250519" llama3.1:8b-instruct,q4_0,1x Intel Arc A770 16GB,-,-,4000,500,8.77687,56.9679,20.0268,-,-,fp2,https://t.me/c/1564516454/155939,"ubuntu 25.04 (6.14) ipex-llm (sycl) bigdl-core-cpp 2.7.0b20250519 ipex-llm 2.3.0b20250519" llama3.1:8b-instruct,q4_K_M,"1x RTX 5060Ti 16GB",-,-,200,23,0.407316,56.4672,3.6409,-,-,yt1,https://t.me/c/1564516454/161465,Без Flash Attention llama3.1:8b-instruct,q4_0,"1x RTX 5060Ti 16GB",-,-,200,251,4.03426,62.2171,7.81242,-,-,yt1,https://t.me/c/1564516454/161465,Без Flash Attention llama3.1:8b-instruct,q8_0,"1x RTX 5060Ti 16GB",-,-,200,244,5.9729,40.8512,10.5656,-,-,yt1,https://t.me/c/1564516454/161465,Без Flash Attention llama3.1:8b-instruct,q4_K_M,"1x RTX 5060Ti 16GB",-,-,200,379,6.28207,60.3305,9.85483,-,-,yt1,https://t.me/c/1564516454/161469,Flash Attention включено llama3.1:8b-instruct,q4_0,"1x RTX 5060Ti 16GB",-,-,200,407,6.05915,67.1711,10.0823,-,-,yt1,https://t.me/c/1564516454/161469,Flash Attention включено llama3.1:8b-instruct,q8_0,"1x RTX 5060Ti 16GB",-,-,200,242,5.83086,41.5033,10.7458,-,-,yt1,https://t.me/c/1564516454/161469,Flash Attention включено llama3.1:8b-instruct,q4_0,"1x Intel Arc A770 16GB",-,-,200,372,5.74311,64.7732,17.0155,-,-,fp2,https://t.me/c/1564516454/161486,- llama3.1:8b-instruct,q4_0,"1x Intel Arc A770 16GB",-,-,4000,500,7.89941,63.2959,19.4269,-,-,fp2,https://t.me/c/1564516454/161486,- llama3.1:8b-instruct,q4_K_M,"1x Intel Arc A770 16GB",-,-,200,500,8.15532,61.3097,17.7278,-,-,fp2,https://t.me/c/1564516454/161486,- llama3.1:8b-instruct,q4_K_M,"1x Intel Arc A770 16GB",-,-,4000,348,5.67623,61.3083,15.4789,-,-,fp2,https://t.me/c/1564516454/161486,- llama3.1:70b-instruct,q4_0,3x Intel Arc A770 16GB,-,-,200,500,54.4203,9.18775,144.996,-,-,fp2,https://t.me/c/1564516454/161510,"load_tensors: SYCL0 model buffer size = 12394.69 MiB load_tensors: SYCL1 model buffer size = 12394.69 MiB load_tensors: SYCL2 model buffer size = 12757.61 MiB" llama3.1:8b-instruct,q8_0,"1x Intel Arc A770 16GB",-,-,200,242,5.9856,40.4304,21.3258,-,-,fp2,https://t.me/c/1564516454/161511,- llama3.1:8b-instruct,q8_0,"1x Intel Arc A770 16GB",-,-,4000,229,5.66146,40.4489,21.2253,-,-,fp2,https://t.me/c/1564516454/161511,-