81 questions
0 votes
0 answers
42 views
Marketing Mix MCMC Model has zero acceptance probability
I am trying to build a Marketing Mix Model using the MCMC implementation from Jax and Black Jax. However when running the model, the sample acceptance probability stays zero, and so there are no ...
1 vote
1 answer
207 views
JAX Point Cloud Processing: Slow index_points_3d operation causing extreme XLA fusion loops in backpropagation
I'm trying to use JAX for implementing point cloud processing. However, I found that training becomes extremely slow due to my implementation of the following index_points_3d operation, which performs ...
0 votes
1 answer
133 views
tensorFlow GPU slower than CPU, ptxas warning
I was trying to make a model testing series I ran into some horrible compilation behavior. The model runs about 5x faster on my CPU (Xeon W-2225, 4 core) than on my RTX 3060 (12GB) OR on my T1000 (4GB)...
1 vote
1 answer
90 views
Jax / Flax potential tracing issue
I'm currently using Flax for neural network implementations. My model takes two inputs: x and θ. It first processes x through an LSTM, then concatenates the LSTM's output with θ — or more precisely, ...
0 votes
0 answers
48 views
Signing xla macros using certificate stored in Azure key vault with HSM
We have been fetching the signing certificate from the Azure Key Vault and adding it to the local Windows store, which allows for the signing of macros in the .xla application through Excel. var ...
0 votes
1 answer
88 views
Does TensorFlow or XLA provide a python API to read and parse the dumped MHLO mlir module?
I turned on XLA when running TensorFLow, and in order to further optimize the fused kernels, I added export XLA_FLAGS="--xla_dump_to=/tmp/xla_dump", and got the dumped IRs, including lmhlo....
1 vote
1 answer
144 views
How to compile tensorflow serving (tensorflow/xla) to have llvm/mlir as shared objects rather than statically included in the binary?
I am trying to compile the tensorflow serving project and I would like to have llvm/mlir compiled as a shared objects. The project is tensorflow serving -> tensorflow -> xla and compiles to a ...
1 vote
0 answers
219 views
Hard to understand the semantics of the stablehlo.scatter
Trying to understand the semantics of https://github.com/openxla/stablehlo/blob/main/docs/spec.md#scatter. Many of attributes do not have any explanation or definition. Can someone please explain how ...
2 votes
1 answer
438 views
looking for a tool to calculate FLOPs of XLA-HLO computational graph
I'm looking for a tool to calculate the FLOPs when given the computational graph of XLA-HLO. Is someone know some HLO cost models or analytical models for print the FLOPs of operator node for ...
2 votes
0 answers
461 views
No registered 'RaggedTensorToTensor' OpKernel for XLA_GPU_JIT devices
In short, I have the problem of getting the following error when running a keras_cv/retina_net based object-detection model: "No registered 'RaggedTensorToTensor' OpKernel for XLA_GPU_JIT devices ...
4 votes
0 answers
3k views
Is there a way to suppress STDERR message from tensorflow and XLA
When I run my python script, I had the messages below: WARNING: All log messages before absl::InitializeLog() is called are written to STDERR I0000 00:00:1701341037.989729 1542352 device_compiler.h:...
0 votes
1 answer
297 views
Is it okay to use python operators for tensorflow tensors?
TL;DR Is (a and b) equivalent to tf.logical_and(a, b) in terms of optimization and performance? (a and b are tensorflow tensors) Details: I use python with tensorflow. My first priority is to make the ...
2 votes
1 answer
2k views
Why does tensorflow.function (without jit_compile) speed up forward passes of a Keras model?
XLA can be enabled using model = tf.function(model, jit_compile=True). Some model types are faster that way, some are slower. So far, so good. But why can model = tf.function(model, jit_compile=None) ...
0 votes
0 answers
70 views
Passing user defined variables to xlam file
How can I pass user-defined variables to xlam files? I have written a macro and saved on an xlam file. I reference the xlam in an xlsm file and I call the macro from the xlsm. I want to pass a user-...
0 votes
1 answer
1k views
Enable multiprocessing on pytorch XLA for TPU VM
I'm fairly new to this and have little to no experience. I had a notebook running PyTorch that I wanted to run a Google Cloud TPU VM. Machine specs: - Ubuntu - TPU v2-8 - pt-2.0 I should have 8 cores....