Skip to content

llama

Summary

Source: https://github.com/alpaka-group/llama

License: GNU Lesser General Public License v3.0

Path: /software/llama/

Documentation: https://alpaka-group.github.io/llama/

Low-Level Abstraction of Memory Access. LLAMA is a cross-platform C++17 template header-only library for the abstraction of memory access patterns. It distinguishes between the view of the algorithm on the memory and the real layout in the background. This enables performance portability for multicore, manycore and gpu applications with the very same code.

Using llama

There is no need for any initialization. You just need to know the location of headers:

module load maxwell
module avail llama

--------------- /software/etc/modulefiles -------------
llama/0.2 llama/0.4

module load maxwell llama

     this is just a dummy module not actually setting up anything

     CMAKE: /software/llama/0.4/lib64/cmake/llama/
     INCLUDES: /software/llama/0.4/include/