GithubHelp home page GithubHelp logo

omarnagy91 / attentiongrid Goto Github PK

View Code? Open in Web Editor NEW

This project forked from kyegomez/attentiongrid

0.0 0.0 0.0 1.84 MB

A network of attention mechanisms at your fingertips. Unleash the potential of attention mechanisms for diverse AI applications. AttentionGrid is all you need!

Home Page: https://apac.ai/Agora

License: Apache License 2.0

Python 100.00%

attentiongrid's Introduction

Agora

Agora Banner

AttentionGrid is brought to you by Agora, we're an all-new open source multi-modal AI research organization devoted to advancing Humanity.

Join us Here to contribute to this project or recieve support!

AttentionGrid: Unleashing Attention Power in AI Models ๐Ÿš€

AttentionGrid Image

AttentionGrid is a cutting-edge framework designed to democratize the incorporation of advanced attention mechanisms into AI models. Powered by the latest developments in attention-based transformer models, AttentionGrid opens up the world of attention mechanisms to machine learning practitioners, researchers, and enthusiasts alike.

Getting Started: Installation ๐Ÿš€

To blast off with AttentionGrid, install the package using pip:

pip install AttentionGrid

Implementing an attention mechanism or a transformer model with AttentionGrid is as easy as:

from AttentionGrid import BlockwiseParallelJax
import jax.numpy as jnp

# Initialize the class
bpjax = BlockwiseParallelJax(
    q_chunk_size=64, 
    k_chunk_size=64, 
    hidden_size=768, 
    num_heads=12, 
    rotary_dim=32, 
    intermediate_size=3072
)

# Suppose we have hidden_states, attention_mask, and position_ids as input data
hidden_states = jnp.random.rand(1, 100, 768)
attention_mask = jnp.random.rand(1, 1, 100, 100)
position_ids = jnp.arange(100).reshape(1, 100)

# You can now apply the attention mechanism to your input data
output = bpjax.forward(hidden_states, attention_mask, position_ids)

Spread the Word ๐Ÿ“ฃ

We encourage you to share AttentionGrid with your community! Here are quick share links for several social media platforms:

Thank you for supporting AttentionGrid and contributing to the democratization of AI! Together, we can push the boundaries of what's possible.

Vision ๐Ÿ‘๏ธ

In the vast landscape of AI, attention mechanisms have revolutionized our ability to create powerful models that can discern the subtleties in data, focusing on important aspects and improving overall performance. Our vision with AttentionGrid is to bridge the gap between these state-of-the-art mechanisms and their practical applications, providing a tool that makes these techniques accessible and easy to implement in diverse AI applications.

Architecture ๐Ÿ—๏ธ

AttentionGrid is designed with an intuitive and flexible architecture, partitioned into four primary components:

  1. Core ๐Ÿ’ก: This is the bedrock of our framework, housing abstract classes that layout the basic structure for attention mechanisms and transformer models.

  2. Attentions ๐Ÿง : The directory dedicated to various attention mechanisms. Each attention mechanism is implemented based on the blueprint provided in the Core.

  3. Transformers ๐Ÿค–: This is where transformer models come to life, each sculpted following the design defined in the Core.

  4. Utils ๐Ÿ› ๏ธ: A toolbox filled with helper classes for essential tasks like model loading, data preprocessing, and more.

  5. Examples ๐ŸŽฏ: Demystifying the implementation with hands-on examples and usage scenarios.

Key Features โœจ

  • Modular Structure: Mix and match different attention mechanisms with a variety of transformer models.

  • User Friendly: Clear documentation and examples to help you get started quickly.

  • Open Source: Open to contributions, AttentionGrid thrives on collective knowledge and shared progress.

For more detailed examples, please refer to the 'examples' folder in our repository.

Contribution ๐Ÿค

We openly invite contributions to AttentionGrid! Whether you have a new feature suggestion, bug report, or want to add to our code, please feel free to open an issue or submit a pull request.

License ๐Ÿ“œ

AttentionGrid is proudly open-source software, licensed under the APACHE License.

Why AttentionGrid? ๐ŸŽฏ

Attention mechanisms have transformed AI, enabling machines to 'focus' on significant parts of input data. With AttentionGrid, we aim to democratize access to these powerful tools. We believe that the future of AI lies in the power of attention, and through AttentionGrid, we hope to accelerate this journey. Explore our repository, join our cause, and let's navigate this exciting landscape together!

"The details are not the details. They make the design." - Charles Eames

Roadmap

  • Integrate Flash Attention, and variants

  • Integrate landmark attention

  • Integrate blockwise parallel attention

  • Integrate dynamic sparse flash attention

  • Integrate cross attention from imagebind

  • Integrate COLT-5 Attention

  • Integrate multi-query attention

  • Integrate wrappers from lucid rains x_transformers, decoder, attention, encoder, transformer wrapper

Documentation

Dynamic Sparse Attention

Agora's dynamic_sparse_attention function allows the flexibility of choosing between the hash-sparse implementation and the qk-sparse implementation. This function's objective is to dynamically direct the sparse attention mechanism based on the selected sparsity_mode.

The function parameters are as follows:

  • q: Query tensor of shape (BATCH, N_CTX_Q, H, D_HEAD)
  • k: Key tensor of shape (BATCH, N_CTX_KV, H, D_HEAD)
  • v: Value tensor of shape (BATCH, N_CTX_KV, H, D_HEAD)
  • q_idx & k_idx: Represent either the bucket index if sparsity_mode is 'hash' or whether to keep a given head if sparsity_mode is 'qk'. The tensor shapes are (BATCH, N_CTX_Q, H) and (BATCH, N_CTX_KV, H) respectively.
  • sm_scale: Normalization constant, 1/sqrt(D_HEAD) unless specified.
  • sparsity_mode: 'hash' to select the hash-sparse implementation and 'qk' for the qk-sparse implementation.

The sm_scale is calculated by default if not provided, and if an unknown sparsity_mode is given, it throws a KeyError.

The function then checks the sparsity_mode and based on its value, it calls either hash_sparse_attention or qk_sparse_attention.

Compact Function

The compact function builds a compact representation of the input tensor x using the information from keep_tensor.

The function parameters are:

  • x: Input tensor to compact, with shape (BATCH, N_CTX, H, D_HEAD).
  • keep_tensor: Float tensor of shape (BATCH, N_CTX, H) containing a 1 when the head is kept and 0 otherwise.

The function first calculates the indices_per_head which computes the number of non-killed elements per head. It sorts the keep_tensor in a descending order while preserving the order of equal elements (stable=True). It then gathers the elements of x based on the index tensor. The result is a compact representation of x along with the index tensor and the tensor representing the number of non-killed elements per head.

Pad Index Function

The pad_index function pads the index tensor to comply with the kernel. It takes the following parameters:

  • index: Original index tensor given by compact, with shape (BATCH, buffer_size, H). For each batch and timestep, it represents the head index it's originating from.
  • indices_per_head: For each head, contains how many indices have not been dropped.

It creates a copy of the index tensor and creates a mask based on the size of indices_per_head. Then it modifies the indices in the copy that correspond to True in the mask to be equal to pad_idx.

QK Sparse Attention Function

The qk_sparse_attention function is part of the dynamic sparse attention mechanism. It is used when sparsity_mode is set to 'qk'. This function implements the qk-sparse attention mechanism and requires that the q_keep and k_keep parameters are of type float.

It first builds compact representations of the query, key, and value tensors using the compact function. It then pads the index tensors using the pad_index function. The tensors are then transposed for compatibility with the kernel. Finally, the function calls the qk_sparse_attention_kernel function and scatters the resulting tensor back into the original dimension space.

Hash Sparse Attention Function

The hash_sparse_attention function is part of the dynamic sparse attention mechanism. It is used when sparsity_mode is set to 'hash'. This function implements the hash-sparse attention mechanism.

The function takes the same input parameters as qk_sparse_attention. However, instead of q_keep and k_keep parameters, the hash_sparse_attention function requires q_bucket_idx and k_bucket_idx which represent bucket indices for queries and keys respectively.

The hash_sparse_attention function first sorts the query, key, and value tensors based on the bucket indices using the sort_bucketed_attention function. Then it builds compact representations of the sorted query, key, and value tensors using the compact function. It then pads the index tensors using the pad_index function.

The tensors are then transposed for compatibility with the kernel. The function then calls the hash_sparse_attention_kernel function and scatters the resulting tensor back into the original dimension space.

Sort Bucketed Attention Function

The sort_bucketed_attention function is a helper function used in hash_sparse_attention. It sorts the input tensors based on the given bucket indices.

The function parameters are:

  • qkv: Query, Key, Value tensors of shape (BATCH, N_CTX, H, D_HEAD)
  • qkv_bucket_idx: Bucket indices for queries, keys, and values of shape (BATCH, N_CTX, H)

The function first sorts the qkv_bucket_idx tensor and gets the sorted indices. Then it sorts the qkv tensors using the sorted indices. It also expands qkv_bucket_idx to be the same shape as qkv for compatibility.

QK Sparse Attention Kernel Function

The qk_sparse_attention_kernel function is a kernel function used in qk_sparse_attention. It calculates the weighted sum of values based on the softmax of the query and key product.

The function parameters are:

  • q: Query tensor of shape (BATCH, N_CTX_Q, H, D_HEAD)
  • k: Key tensor of shape (BATCH, N_CTX_KV, H, D_HEAD)
  • v: Value tensor of shape (BATCH, N_CTX_KV, H, D_HEAD)
  • sm_scale: Normalization constant, 1/sqrt(D_HEAD) unless specified.

Hash Sparse Attention Kernel Function

The hash_sparse_attention_kernel function is a kernel function used in hash_sparse_attention. It works similarly to qk_sparse_attention_kernel but handles bucketing for hash-sparse attention.

The function parameters are the same as those of qk_sparse_attention_kernel. However, q, k, and v have been sorted and compacted based on the bucket indices.

The kernel computes the product of the query and key, scales it by sm_scale, applies softmax to get the weights, and then calculates the weighted sum of the values.

Please note that this is a general interpretation of the documentation, and understanding and modifying these functions in practice may require in-depth knowledge of sparse attention mechanisms and deep learning principles.

Blockwise Parallel Documentation

blockwise_compute_attn Function:

The blockwise_compute_attn function is an important part of the BlockwiseParallelJax class and is used to compute the attention mechanism of the model in a blockwise manner.

Parameters:

  • query, key, value: These parameters are the main inputs for the attention computation, representing queries, keys, and values, respectively.
  • bias: Optional parameter used to add a bias to the attention scores before softmax.
  • deterministic: A boolean flag used to decide whether or not to apply dropout.
  • dropout_rng: The random number generator for dropout.
  • attn_pdrop: The probability of dropout for attention.
  • causal_mask: A boolean flag for whether or not to use a causal attention mask.
  • query_chunk_size, key_chunk_size: The size of each query and key chunk, respectively.
  • dtype: The data type of the computation. It's default is jnp.float32.
  • policy: This parameter defines the policy for gradient checkpointing.
  • precision: This parameter is used to set the level of precision for the computation. The default value is lax.Precision.HIGHEST.
  • prevent_cse: A boolean flag used to prevent common subexpression elimination.

blockwise_compute_ffn Function:

The blockwise_compute_ffn function is used to compute the feed-forward network of the model in a blockwise manner.

Parameters:

  • cell: The cell in the network to which the function is applied.
  • inputs: The input data for the feed-forward network.
  • chunk_size: The size of each chunk for the blockwise computation.
  • deterministic: A boolean flag used to decide whether or not to apply dropout.
  • policy: This parameter defines the policy for gradient checkpointing.
  • prevent_cse: A boolean flag used to prevent common subexpression elimination.

Blockwise_LM_Head Class:

The Blockwise_LM_Head class is a module that applies a linear transformation followed by a softmax function to produce a distribution over the vocabulary for each position in the input.

  • vocab_size: The size of the vocabulary, which is also the size of the output dimension of the linear transformation.
  • chunk_size: The size of each chunk for the blockwise computation.
  • policy: This parameter defines the policy for gradient checkpointing.
  • dtype: The data type of the computation. It's default is jnp.float32.
  • prevent_cse: A boolean flag used to prevent common subexpression elimination.

blockwise_cross_entropy Function:

The blockwise_cross_entropy function calculates the cross-entropy loss for the model's predictions in a blockwise manner.

Parameters:

  • logits: The model's output predictions.
  • tokens: The true labels.
  • valid: A mask that specifies the valid positions in the input.
  • chunk_size: The size of each chunk for the blockwise computation.
  • policy: This parameter defines the policy for gradient checkpointing.
  • prevent_cse: A boolean flag used to prevent common subexpression elimination.

BlockwiseParallelJax Class:

BlockwiseParallelJax(q_chunk_size, k_chunk_size, hidden_size, num_heads, rotary_dim, intermediate_size, layer_norm_epsilon=1e-5, activation_function="gelu", attn_pdrop=0.0, resid_pdrop=0.0, max_position_embeddings=1024, dtype=jnp.float32, causal=True, policy='nothing_saveable', prevent_cse=False, float32_logits=False)

Parameters

  • q_chunk_size : Integer. Chunk size for the query in self-attention.
  • k_chunk_size : Integer. Chunk size for the key in self-attention.
  • hidden_size : Integer. Dimensionality of the hidden layer in the transformer.
  • num_heads : Integer. Number of attention heads in the self-attention mechanism.
  • rotary_dim : Integer or None. Number of dimensions to use for rotary positional encoding.
  • intermediate_size : Integer. Size of the intermediate layer in the feed-forward network.
  • layer_norm_epsilon : Float. Small constant to prevent division by zero in layer normalization. Default is 1e-5.
  • activation_function : String. Activation function to use in the feed-forward network. Default is 'gelu'.
  • attn_pdrop : Float. Dropout probability for the attention mechanism. Default is 0.0.
  • resid_pdrop : Float. Dropout probability for the residual connections. Default is 0.0.
  • max_position_embeddings : Integer. Maximum number of position embeddings to use. Default is 1024.
  • dtype : jnp.dtype. Data type to use for computation. Default is jnp.float32.
  • causal : Boolean. Whether to use causal (auto-regressive) mode or not. Default is True.
  • policy : String. Policy for checkpointing gradients. Default is 'nothing_saveable'.
  • prevent_cse : Boolean. Whether to prevent common subexpression elimination (CSE). Default is False.
  • float32_logits : Boolean. Whether to use float32 for logits computation. Default is False.

Methods

The main method of the BlockwiseParallelJax class is the forward method, which performs the forward pass of the transformer block.

forward(hidden_states, attention_mask, position_ids, deterministic=True, init_cache=False)
  • hidden_states : jnp.ndarray. The input tensor to the transformer block. It should have shape (batch_size, sequence_length, hidden_size).
  • attention_mask : jnp.ndarray. The attention mask for the self-attention mechanism. It should have shape (batch_size, 1, 1, sequence_length).
  • position_ids : jnp.ndarray. The position ids for positional encoding. It should have shape (1, sequence_length).
  • deterministic : Boolean. Whether to use deterministic mode (no dropout) or not. Default is True.
  • init_cache : Boolean. Whether to initialize the cache for fast decoding. Default is False.

This method returns the output tensor of the transformer block, which has the same shape as hidden_states.

Example Usage

The following example demonstrates how to use the BlockwiseParallelJax class.

# Initialize

from jax import random
import jax.numpy as jnp
from agora_ai_transformer.models import BlockwiseParallelJax

# Initialize transformer block
block = BlockwiseParallelJax(
    q_chunk_size=64,
    k_chunk_size=64,
    hidden_size=768,
    num_heads=12,
    rotary_dim=64,
    intermediate_size=3072,
)

# Create a batch of input tensors
key = random.PRNGKey(0)
batch_size = 8
sequence_length = 128
hidden_states = random.normal(key, (batch_size, sequence_length, block.hidden_size))

# Create attention mask
attention_mask = jnp.ones((batch_size, 1, 1, sequence_length))

# Create position ids
position_ids = jnp.arange(sequence_length)[None, :]

# Forward pass
output = block.forward(hidden_states, attention_mask, position_ids)

print(output.shape)  # prints: (8, 128, 768)

Landmark Attention

Class FusedLandmarkAttention

This is a PyTorch Function class that encapsulates the forward and backward functions of the Fused Landmark Attention mechanism.

forward(ctx, q, k, v, n_prefix_q, sm_scale, block_size)

This function performs the forward pass of the Fused Landmark Attention.

Parameters:
  • ctx: An object to which we can save variables for use in the backward pass. Provided by PyTorch's autograd system.
  • q: The queries tensor. It is assumed to be contiguous, and its shape should be (batch, nheads, seqlen_q, d).
  • k: The keys tensor. It is assumed to be contiguous, and its shape should match q's shape, i.e., (batch, nheads, seqlen_k, d).
  • v: The values tensor. It is assumed to be contiguous, and its shape should match q's and k's shapes, i.e., (batch, nheads, seqlen_k, d).
  • n_prefix_q: The number of prefixes in the queries.
  • sm_scale: The scaling factor used in the softmax operation.
  • block_size: The block size for performing block-wise operations.
Returns:
  • o: The output tensor from the forward pass of the Fused Landmark Attention mechanism.

backward(ctx, do)

This function performs the backward pass of the Fused Landmark Attention, i.e., it calculates the gradients.

Parameters:
  • ctx: An object from which we can retrieve variables saved in the forward pass. Provided by PyTorch's autograd system.
  • do: The gradient of the loss with respect to the output of the forward function.
Returns:
  • A tuple containing the gradients of the loss with respect to the inputs to the forward function, in the same order as they were provided in. If a certain input does not require gradient, its corresponding gradient will be None.

Function fused_landmark_attention(q, k, v, is_mem, sm_scale=None, block_size=64)

This function is a convenient wrapper for the FusedLandmarkAttention class.

Parameters:
  • q: The queries tensor.
  • k: The keys tensor.
  • v: The values tensor.
  • is_mem: A boolean tensor indicating whether each key-value pair should be treated as memory. It should have the same length as the sequence length of the keys.
  • sm_scale: The scaling factor used in the softmax operation. If None, it will be set to 1.0 / sqrt(d).
  • block_size: The block size for performing block-wise operations.
Returns:
  • The output tensor from the forward pass of the Fused Landmark Attention mechanism.

Example:

Here is a basic example of how to use fused_landmark_attention function.

import torch
from your_module import fused_landmark_attention

# Initialize some tensors
batch = 8
nheads = 12
seqlen = 128
d = 64
q = torch.randn(batch, nheads, seqlen, d)
k = torch.randn(batch, nheads, seqlen, d)
v = torch.randn(batch, nheads, seqlen, d)
is_mem = torch.zeros(seqlen, dtype=torch.bool)

# Call the function
output = fused_landmark_attention(q, k, v, is_mem)

print(output.shape)  # prints: (8, 12, 128, 64)

This example first initializes some tensors to serve as the queries, keys

, and values. Then it calls the fused_landmark_attention function and prints the shape of the output tensor.

Torchscale

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.