Anybody running an Intel arc 750 or 770? #1923
Replies: 5 comments 18 replies
-
This is gaming oriented, but it seems like it's about half the speed of a 3090: https://gpu.userbenchmark.com/Compare/Nvidia-RTX-3090-vs-Intel-Arc-A770/4081vsm1850973 It also has 33% less memory. However, it costs about half as much. Assuming the OpenCL performance is in line with the gaming performance, it could possibly make sense to get two of them and use stuff like GGML GPU splitting feature. However, the cards have 250 watt TDP so that's a huge amount of power. The cheapest one I found was $339, used 3090s are $700-800. I didn't find anything much about OpenCL usage, but it seems like the card is pretty well supported on Linux. |
Beta Was this translation helpful? Give feedback.
-
I checked out the the specs on these recently. Yes, price for performance is very good. Though am I right to assume they are using 50-100% more electricity than an equaivalent nvidia card? I'm SUPER impressed with the performance versus energy of my MacOS M2. The Amazon Inferentia etc ARM chips are also super energy efficient. |
Beta Was this translation helpful? Give feedback.
-
I have an A380 (ASRock Challenger) and tried llama.cpp. 7B (vicuna-1.0, Q4_0) produced a terrible result of 240ms/tok. (ArchLinux, E5-2670 v3 with DDR4-2133 32GB) |
Beta Was this translation helpful? Give feedback.
-
Run LLama-2 13B, very fast, Locally on Low Cost Intel's ARC GPU , iGPU and on CPU:- https://youtu.be/FRWy7rzOsRs |
Beta Was this translation helpful? Give feedback.
-
here is video how to run using sycl its father than CLBLAST :-
https://www.youtube.com/watch?v=Q7t4CmziaqA
…On Mon, 19 Aug 2024 at 07:25, Neo Zhang Jianyu ***@***.***> wrote:
@delock <https://github.com/delock>
You could try with llama.cpp for SYCL backend, refer to
https://github.com/ggerganov/llama.cpp/blob/master/docs/backend/SYCL.md
It's faster than CLBlast on Intel GPU.
—
Reply to this email directly, view it on GitHub
<#1923 (reply in thread)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/AFKFRVK4HWROL7SE6OXRLQLZSFGAXAVCNFSM6AAAAABL64FVKCVHI2DSMVQWIX3LMV43URDJONRXK43TNFXW4Q3PNVWWK3TUHMYTAMZXG4ZDEOA>
.
You are receiving this because you were mentioned.Message ID:
***@***.***>
|
Beta Was this translation helpful? Give feedback.
-
If anyone is running these cards for their vram capacity , what is your experience like? How many iterations/ms are you getting through opencl offloading? Does it work with UI's like oobabooga and is it worth getting one?
Beta Was this translation helpful? Give feedback.
All reactions