llama¶
Summary
Source: https://github.com/alpaka-group/llama
License: GNU Lesser General Public License v3.0
Path: /software/llama/
Documentation: https://alpaka-group.github.io/llama/
Low-Level Abstraction of Memory Access. LLAMA is a cross-platform C++17 template header-only library for the abstraction of memory access patterns. It distinguishes between the view of the algorithm on the memory and the real layout in the background. This enables performance portability for multicore, manycore and gpu applications with the very same code.
Using llama¶
There is no need for any initialization. You just need to know the location of headers:
module load maxwell
module avail llama
--------------- /software/etc/modulefiles -------------
llama/0.2 llama/0.4
module load maxwell llama
this is just a dummy module not actually setting up anything
CMAKE: /software/llama/0.4/lib64/cmake/llama/
INCLUDES: /software/llama/0.4/include/