Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Boolean algebra and bit counting on tensor elements #2641

Open
0x7CFE opened this issue Dec 23, 2024 · 0 comments
Open

Boolean algebra and bit counting on tensor elements #2641

0x7CFE opened this issue Dec 23, 2024 · 0 comments

Comments

@0x7CFE
Copy link

0x7CFE commented Dec 23, 2024

Feature description

I would like to have a way to perform boolean algebra operations on tensor elements, as well as bit counting operations, like count_ones.

Feature motivation

In my discrete AI research I am dealing with billions of binary embedding values that are represented by u128s or [u64; N]s. For certain operations I am required to do boolean algebra on them. One especially tricky operation is discrete correlation which requires bit counting of tensor elements.

The equivalent scalar code looks like this:

pub fn correlation<T>(x: BitVector<T>, y: BitVector<T>) -> f32
where
    BitVector<T>: BitOps,
{
    let nx = x.count_ones();
    let ny = y.count_ones();
    let s = (x & y).count_ones();
    let product = nx * ny;

    if product == 0 {
        0.
    } else {
        s as f32 / (product as f32).sqrt()
    }
}

Even with rudimentary support for integer computation in consumer GPUs, I was able to achieve ~30x speedup when compared to CPU. Currently everything is implemented in custom shaders using Vulkan and working fine. But the code is definitely not composable and PITA to support. So I am looking for a better solution.

(Optional) Suggest a Solution

So, technically it is possible to implement and was already proven to be efficient. But I have no idea whether Burn or any of its backends would be suitable to do that sanely.

I assume it can be done by writing custom wgpu kernel, but that would probably not be generic enough. For my quite niche case (to say the least) this is probably the way to go, but I wanted to ask for your opinion first. Maybe someone else would also look for something as crazy.

Thanks in advance!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant