Skip to main content
1 vote
0 answers
34 views

I have created a .NET app that uses Microsoft.ML.OnnxRuntime.Gpu for interference. Now I'm trying to integrate it with Azure Kubernetes. We have made the setup with Tesla T4 GPU and we confirmed it's ...
ervin's user avatar
  • 555
-3 votes
0 answers
64 views

I am currently writing a driver for the Intel ARC GPU series (specifically I use the A750 for testing purposes) for my own operating system. I am already able to execute compute kernels that use ...
Joel Marker's user avatar
-2 votes
0 answers
66 views

I am making a particle simulator in python, and noticed that my collision detection is ruining the performance the most,I am not even sure if this is a thing but is it possible to tell the GPU to do ...
Dario's user avatar
  • 1
0 votes
0 answers
46 views

This is a bit of a slog so bare with me. I'm currently writing a 3D S(moothed) P(article) H(ydrodynamics) simulation in Unity with a parallel HLSL backend. It's a Lagrangian method of fluid simulation,...
Ben Williams's user avatar
Tooling
0 votes
0 replies
24 views

I am running Flux 1 dev text to image model through ComfyUI in Kaggle. Everything works but I noticed that Kaggle offers a second GPU inside the notebook. If I try to run two instances of the ComfyUI ...
Bram Fran's user avatar
  • 113
-4 votes
0 answers
31 views

Im trying to use tensorflow with gpu on my windows device, i have python 3.13 venv. Is newer version of tensorflow support gou acceleration on windows. Ive read that it stopped in tensorflow version 2....
Med Yassine Ghaoui's user avatar
Tooling
0 votes
0 replies
85 views

I'm exploring options for running large language models locally on my workstation and would appreciate guidance on suitable models given my hardware constraints. Hardware specifications: CPU: Intel ...
GbreH's user avatar
  • 13
1 vote
1 answer
72 views

Is @ray.remote def run_experiment(...): (...) if __name__ == '__main__': ray.init() exp_config = sys.argv[1] params_tuples, num_cpus, num_gpus = load_exp_config(exp_config) ray.get(...
Blupon's user avatar
  • 1,091
0 votes
0 answers
52 views

I am invoking a compute shader, writing to it, then reading it to then write to disk. According to renderdoc the image is properly generated. Additionally, when compiled in debug mode I get the right ...
Makogan's user avatar
  • 9,991
2 votes
0 answers
115 views

Context I'm implementing a Julia set fractal renderer using CubeCL (a Rust GPU compute framework). I want to achieve "infinite zoom" similar to deep Mandelbrot zoom videos, which requires ...
Marco Fanelli's user avatar
3 votes
0 answers
112 views

I have encountered a particular problem while executing a function from the transformers library of huggingface on an Intel GPU wheel of torch. Since I am doing something I normally shouldn't be ...
Logarithmnepnep's user avatar
1 vote
0 answers
65 views

For testing purposes I need a tool that will occupy some amount of VRAM, leaving a reduced available VRAM to the rest of the applications. I implemented a version that somewhat works using D3D12 API, ...
Virgileo's user avatar
0 votes
0 answers
64 views

I have a machine-translation model. In this model, I calculate a vector for a given sentence and I take this vector, aggregate with each generated output of RNN and put it into RNN again for ...
cuneyttyler's user avatar
  • 1,395
0 votes
0 answers
90 views

I am using an i686 system, with the compiler mingw G++. I can run code that creates a GPU device and attaches it to a window fine on that machine. However, when I attempt to run it on my i686 windows ...
Jasper Stocks's user avatar
0 votes
1 answer
99 views

I am trying to implement producer consumer problem in GPU-CPU. Required for some other project. GPU requests some data via Unified memory to CPU. CPU copies that data to a specific location in global ...
Chinmaya Bhat K K's user avatar

15 30 50 per page
1
2 3 4 5
610