Skip to content

Commit 3d59d19

Browse files
authored
Update sparsity_config.py
1 parent 73dfe6b commit 3d59d19

1 file changed

Lines changed: 1 addition & 2 deletions

File tree

deepspeed/ops/sparse_attention/sparsity_config.py

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -681,10 +681,9 @@ def __init__(self,
681681
num_heads: required: an integer determining number of attention heads of the layer.
682682
block: optional: an integer determining the block size. Current implementation of sparse self-attention is based on blocked sparse matrices. In which this parameter defines size of such blocks, `Block X Block`.
683683
num_sliding_window_blocks: optional: an integer determining the number of blocks in sliding local attention window.
684-
num_global_blocks: optional: an integer determining how many consecutive blocks, starting from index 0, are considered as global attention. Global block tokens will be attended by all other block tokens and will attend to all other block tokens as well.
685684
"""
686685

687-
super().__init__(num_heads, block, different_layout_per_head)
686+
super().__init__(num_heads, block)
688687
self.num_sliding_window_blocks = num_sliding_window_blocks
689688

690689
def set_sliding_window_layout(self, h, layout):

0 commit comments

Comments
 (0)