Merge pull request #75 from Xrenya/patch-1
Update docstring of image_encoder.py
This commit is contained in:
commit
ca981bf377
@ -144,7 +144,7 @@ class Block(nn.Module):
|
||||
rel_pos_zero_init (bool): If True, zero initialize relative positional parameters.
|
||||
window_size (int): Window size for window attention blocks. If it equals 0, then
|
||||
use global attention.
|
||||
input_size (int or None): Input resolution for calculating the relative positional
|
||||
input_size (tuple(int, int) or None): Input resolution for calculating the relative positional
|
||||
parameter size.
|
||||
"""
|
||||
super().__init__()
|
||||
@ -201,7 +201,7 @@ class Attention(nn.Module):
|
||||
qkv_bias (bool): If True, add a learnable bias to query, key, value.
|
||||
rel_pos (bool): If True, add relative positional embeddings to the attention map.
|
||||
rel_pos_zero_init (bool): If True, zero initialize relative positional parameters.
|
||||
input_size (int or None): Input resolution for calculating the relative positional
|
||||
input_size (tuple(int, int) or None): Input resolution for calculating the relative positional
|
||||
parameter size.
|
||||
"""
|
||||
super().__init__()
|
||||
|
Loading…
x
Reference in New Issue
Block a user