Spaces:
Runtime error
Runtime error
Commit
β’
02bb056
1
Parent(s):
fa38555
Upload 24 files
Browse files- tsr/.DS_Store +0 -0
- tsr/__pycache__/system.cpython-310.pyc +0 -0
- tsr/__pycache__/utils.cpython-310.pyc +0 -0
- tsr/models/.DS_Store +0 -0
- tsr/models/__pycache__/camera.cpython-310.pyc +0 -0
- tsr/models/__pycache__/isosurface.cpython-310.pyc +0 -0
- tsr/models/__pycache__/nerf_renderer.cpython-310.pyc +0 -0
- tsr/models/__pycache__/network_utils.cpython-310.pyc +0 -0
- tsr/models/isosurface.py +48 -0
- tsr/models/nerf_renderer.py +180 -0
- tsr/models/network_utils.py +124 -0
- tsr/models/tokenizers/__pycache__/dino.cpython-310.pyc +0 -0
- tsr/models/tokenizers/__pycache__/image.cpython-310.pyc +0 -0
- tsr/models/tokenizers/__pycache__/triplane.cpython-310.pyc +0 -0
- tsr/models/tokenizers/image.py +67 -0
- tsr/models/tokenizers/triplane.py +45 -0
- tsr/models/transformer/__pycache__/attention.cpython-310.pyc +0 -0
- tsr/models/transformer/__pycache__/basic_transformer_block.cpython-310.pyc +0 -0
- tsr/models/transformer/__pycache__/transformer_1d.cpython-310.pyc +0 -0
- tsr/models/transformer/attention.py +628 -0
- tsr/models/transformer/basic_transformer_block.py +314 -0
- tsr/models/transformer/transformer_1d.py +216 -0
- tsr/system.py +203 -0
- tsr/utils.py +482 -0
tsr/.DS_Store
ADDED
Binary file (6.15 kB). View file
|
|
tsr/__pycache__/system.cpython-310.pyc
ADDED
Binary file (5.41 kB). View file
|
|
tsr/__pycache__/utils.cpython-310.pyc
ADDED
Binary file (13.8 kB). View file
|
|
tsr/models/.DS_Store
ADDED
Binary file (6.15 kB). View file
|
|
tsr/models/__pycache__/camera.cpython-310.pyc
ADDED
Binary file (1.48 kB). View file
|
|
tsr/models/__pycache__/isosurface.cpython-310.pyc
ADDED
Binary file (2.04 kB). View file
|
|
tsr/models/__pycache__/nerf_renderer.cpython-310.pyc
ADDED
Binary file (5.29 kB). View file
|
|
tsr/models/__pycache__/network_utils.cpython-310.pyc
ADDED
Binary file (3.42 kB). View file
|
|
tsr/models/isosurface.py
ADDED
@@ -0,0 +1,48 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
from typing import Callable, Optional, Tuple
|
2 |
+
|
3 |
+
import numpy as np
|
4 |
+
import torch
|
5 |
+
import torch.nn as nn
|
6 |
+
from torchmcubes import marching_cubes
|
7 |
+
|
8 |
+
|
9 |
+
class IsosurfaceHelper(nn.Module):
|
10 |
+
points_range: Tuple[float, float] = (0, 1)
|
11 |
+
|
12 |
+
@property
|
13 |
+
def grid_vertices(self) -> torch.FloatTensor:
|
14 |
+
raise NotImplementedError
|
15 |
+
|
16 |
+
|
17 |
+
class MarchingCubeHelper(IsosurfaceHelper):
|
18 |
+
def __init__(self, resolution: int) -> None:
|
19 |
+
super().__init__()
|
20 |
+
self.resolution = resolution
|
21 |
+
self.mc_func: Callable = marching_cubes
|
22 |
+
self._grid_vertices: Optional[torch.FloatTensor] = None
|
23 |
+
|
24 |
+
@property
|
25 |
+
def grid_vertices(self) -> torch.FloatTensor:
|
26 |
+
if self._grid_vertices is None:
|
27 |
+
# keep the vertices on CPU so that we can support very large resolution
|
28 |
+
x, y, z = (
|
29 |
+
torch.linspace(*self.points_range, self.resolution),
|
30 |
+
torch.linspace(*self.points_range, self.resolution),
|
31 |
+
torch.linspace(*self.points_range, self.resolution),
|
32 |
+
)
|
33 |
+
x, y, z = torch.meshgrid(x, y, z, indexing="ij")
|
34 |
+
verts = torch.cat(
|
35 |
+
[x.reshape(-1, 1), y.reshape(-1, 1), z.reshape(-1, 1)], dim=-1
|
36 |
+
).reshape(-1, 3)
|
37 |
+
self._grid_vertices = verts
|
38 |
+
return self._grid_vertices
|
39 |
+
|
40 |
+
def forward(
|
41 |
+
self,
|
42 |
+
level: torch.FloatTensor,
|
43 |
+
) -> Tuple[torch.FloatTensor, torch.LongTensor]:
|
44 |
+
level = -level.view(self.resolution, self.resolution, self.resolution)
|
45 |
+
v_pos, t_pos_idx = self.mc_func(level.detach(), 0.0)
|
46 |
+
v_pos = v_pos[..., [2, 1, 0]]
|
47 |
+
v_pos = v_pos / (self.resolution - 1.0)
|
48 |
+
return v_pos, t_pos_idx
|
tsr/models/nerf_renderer.py
ADDED
@@ -0,0 +1,180 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
from dataclasses import dataclass, field
|
2 |
+
from typing import Dict
|
3 |
+
|
4 |
+
import torch
|
5 |
+
import torch.nn.functional as F
|
6 |
+
from einops import rearrange, reduce
|
7 |
+
|
8 |
+
from ..utils import (
|
9 |
+
BaseModule,
|
10 |
+
chunk_batch,
|
11 |
+
get_activation,
|
12 |
+
rays_intersect_bbox,
|
13 |
+
scale_tensor,
|
14 |
+
)
|
15 |
+
|
16 |
+
|
17 |
+
class TriplaneNeRFRenderer(BaseModule):
|
18 |
+
@dataclass
|
19 |
+
class Config(BaseModule.Config):
|
20 |
+
radius: float
|
21 |
+
|
22 |
+
feature_reduction: str = "concat"
|
23 |
+
density_activation: str = "trunc_exp"
|
24 |
+
density_bias: float = -1.0
|
25 |
+
color_activation: str = "sigmoid"
|
26 |
+
num_samples_per_ray: int = 128
|
27 |
+
randomized: bool = False
|
28 |
+
|
29 |
+
cfg: Config
|
30 |
+
|
31 |
+
def configure(self) -> None:
|
32 |
+
assert self.cfg.feature_reduction in ["concat", "mean"]
|
33 |
+
self.chunk_size = 0
|
34 |
+
|
35 |
+
def set_chunk_size(self, chunk_size: int):
|
36 |
+
assert (
|
37 |
+
chunk_size >= 0
|
38 |
+
), "chunk_size must be a non-negative integer (0 for no chunking)."
|
39 |
+
self.chunk_size = chunk_size
|
40 |
+
|
41 |
+
def query_triplane(
|
42 |
+
self,
|
43 |
+
decoder: torch.nn.Module,
|
44 |
+
positions: torch.Tensor,
|
45 |
+
triplane: torch.Tensor,
|
46 |
+
) -> Dict[str, torch.Tensor]:
|
47 |
+
input_shape = positions.shape[:-1]
|
48 |
+
positions = positions.view(-1, 3)
|
49 |
+
|
50 |
+
# positions in (-radius, radius)
|
51 |
+
# normalized to (-1, 1) for grid sample
|
52 |
+
positions = scale_tensor(
|
53 |
+
positions, (-self.cfg.radius, self.cfg.radius), (-1, 1)
|
54 |
+
)
|
55 |
+
|
56 |
+
def _query_chunk(x):
|
57 |
+
indices2D: torch.Tensor = torch.stack(
|
58 |
+
(x[..., [0, 1]], x[..., [0, 2]], x[..., [1, 2]]),
|
59 |
+
dim=-3,
|
60 |
+
)
|
61 |
+
out: torch.Tensor = F.grid_sample(
|
62 |
+
rearrange(triplane, "Np Cp Hp Wp -> Np Cp Hp Wp", Np=3),
|
63 |
+
rearrange(indices2D, "Np N Nd -> Np () N Nd", Np=3),
|
64 |
+
align_corners=False,
|
65 |
+
mode="bilinear",
|
66 |
+
)
|
67 |
+
if self.cfg.feature_reduction == "concat":
|
68 |
+
out = rearrange(out, "Np Cp () N -> N (Np Cp)", Np=3)
|
69 |
+
elif self.cfg.feature_reduction == "mean":
|
70 |
+
out = reduce(out, "Np Cp () N -> N Cp", Np=3, reduction="mean")
|
71 |
+
else:
|
72 |
+
raise NotImplementedError
|
73 |
+
|
74 |
+
net_out: Dict[str, torch.Tensor] = decoder(out)
|
75 |
+
return net_out
|
76 |
+
|
77 |
+
if self.chunk_size > 0:
|
78 |
+
net_out = chunk_batch(_query_chunk, self.chunk_size, positions)
|
79 |
+
else:
|
80 |
+
net_out = _query_chunk(positions)
|
81 |
+
|
82 |
+
net_out["density_act"] = get_activation(self.cfg.density_activation)(
|
83 |
+
net_out["density"] + self.cfg.density_bias
|
84 |
+
)
|
85 |
+
net_out["color"] = get_activation(self.cfg.color_activation)(
|
86 |
+
net_out["features"]
|
87 |
+
)
|
88 |
+
|
89 |
+
net_out = {k: v.view(*input_shape, -1) for k, v in net_out.items()}
|
90 |
+
|
91 |
+
return net_out
|
92 |
+
|
93 |
+
def _forward(
|
94 |
+
self,
|
95 |
+
decoder: torch.nn.Module,
|
96 |
+
triplane: torch.Tensor,
|
97 |
+
rays_o: torch.Tensor,
|
98 |
+
rays_d: torch.Tensor,
|
99 |
+
**kwargs,
|
100 |
+
):
|
101 |
+
rays_shape = rays_o.shape[:-1]
|
102 |
+
rays_o = rays_o.view(-1, 3)
|
103 |
+
rays_d = rays_d.view(-1, 3)
|
104 |
+
n_rays = rays_o.shape[0]
|
105 |
+
|
106 |
+
t_near, t_far, rays_valid = rays_intersect_bbox(rays_o, rays_d, self.cfg.radius)
|
107 |
+
t_near, t_far = t_near[rays_valid], t_far[rays_valid]
|
108 |
+
|
109 |
+
t_vals = torch.linspace(
|
110 |
+
0, 1, self.cfg.num_samples_per_ray + 1, device=triplane.device
|
111 |
+
)
|
112 |
+
t_mid = (t_vals[:-1] + t_vals[1:]) / 2.0
|
113 |
+
z_vals = t_near * (1 - t_mid[None]) + t_far * t_mid[None] # (N_rays, N_samples)
|
114 |
+
|
115 |
+
xyz = (
|
116 |
+
rays_o[:, None, :] + z_vals[..., None] * rays_d[..., None, :]
|
117 |
+
) # (N_rays, N_sample, 3)
|
118 |
+
|
119 |
+
mlp_out = self.query_triplane(
|
120 |
+
decoder=decoder,
|
121 |
+
positions=xyz,
|
122 |
+
triplane=triplane,
|
123 |
+
)
|
124 |
+
|
125 |
+
eps = 1e-10
|
126 |
+
# deltas = z_vals[:, 1:] - z_vals[:, :-1] # (N_rays, N_samples)
|
127 |
+
deltas = t_vals[1:] - t_vals[:-1] # (N_rays, N_samples)
|
128 |
+
alpha = 1 - torch.exp(
|
129 |
+
-deltas * mlp_out["density_act"][..., 0]
|
130 |
+
) # (N_rays, N_samples)
|
131 |
+
accum_prod = torch.cat(
|
132 |
+
[
|
133 |
+
torch.ones_like(alpha[:, :1]),
|
134 |
+
torch.cumprod(1 - alpha[:, :-1] + eps, dim=-1),
|
135 |
+
],
|
136 |
+
dim=-1,
|
137 |
+
)
|
138 |
+
weights = alpha * accum_prod # (N_rays, N_samples)
|
139 |
+
comp_rgb_ = (weights[..., None] * mlp_out["color"]).sum(dim=-2) # (N_rays, 3)
|
140 |
+
opacity_ = weights.sum(dim=-1) # (N_rays)
|
141 |
+
|
142 |
+
comp_rgb = torch.zeros(
|
143 |
+
n_rays, 3, dtype=comp_rgb_.dtype, device=comp_rgb_.device
|
144 |
+
)
|
145 |
+
opacity = torch.zeros(n_rays, dtype=opacity_.dtype, device=opacity_.device)
|
146 |
+
comp_rgb[rays_valid] = comp_rgb_
|
147 |
+
opacity[rays_valid] = opacity_
|
148 |
+
|
149 |
+
comp_rgb += 1 - opacity[..., None]
|
150 |
+
comp_rgb = comp_rgb.view(*rays_shape, 3)
|
151 |
+
|
152 |
+
return comp_rgb
|
153 |
+
|
154 |
+
def forward(
|
155 |
+
self,
|
156 |
+
decoder: torch.nn.Module,
|
157 |
+
triplane: torch.Tensor,
|
158 |
+
rays_o: torch.Tensor,
|
159 |
+
rays_d: torch.Tensor,
|
160 |
+
) -> Dict[str, torch.Tensor]:
|
161 |
+
if triplane.ndim == 4:
|
162 |
+
comp_rgb = self._forward(decoder, triplane, rays_o, rays_d)
|
163 |
+
else:
|
164 |
+
comp_rgb = torch.stack(
|
165 |
+
[
|
166 |
+
self._forward(decoder, triplane[i], rays_o[i], rays_d[i])
|
167 |
+
for i in range(triplane.shape[0])
|
168 |
+
],
|
169 |
+
dim=0,
|
170 |
+
)
|
171 |
+
|
172 |
+
return comp_rgb
|
173 |
+
|
174 |
+
def train(self, mode=True):
|
175 |
+
self.randomized = mode and self.cfg.randomized
|
176 |
+
return super().train(mode=mode)
|
177 |
+
|
178 |
+
def eval(self):
|
179 |
+
self.randomized = False
|
180 |
+
return super().eval()
|
tsr/models/network_utils.py
ADDED
@@ -0,0 +1,124 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
from dataclasses import dataclass, field
|
2 |
+
from typing import Optional
|
3 |
+
|
4 |
+
import torch
|
5 |
+
import torch.nn as nn
|
6 |
+
from einops import rearrange
|
7 |
+
|
8 |
+
from ..utils import BaseModule
|
9 |
+
|
10 |
+
|
11 |
+
class TriplaneUpsampleNetwork(BaseModule):
|
12 |
+
@dataclass
|
13 |
+
class Config(BaseModule.Config):
|
14 |
+
in_channels: int
|
15 |
+
out_channels: int
|
16 |
+
|
17 |
+
cfg: Config
|
18 |
+
|
19 |
+
def configure(self) -> None:
|
20 |
+
self.upsample = nn.ConvTranspose2d(
|
21 |
+
self.cfg.in_channels, self.cfg.out_channels, kernel_size=2, stride=2
|
22 |
+
)
|
23 |
+
|
24 |
+
def forward(self, triplanes: torch.Tensor) -> torch.Tensor:
|
25 |
+
triplanes_up = rearrange(
|
26 |
+
self.upsample(
|
27 |
+
rearrange(triplanes, "B Np Ci Hp Wp -> (B Np) Ci Hp Wp", Np=3)
|
28 |
+
),
|
29 |
+
"(B Np) Co Hp Wp -> B Np Co Hp Wp",
|
30 |
+
Np=3,
|
31 |
+
)
|
32 |
+
return triplanes_up
|
33 |
+
|
34 |
+
|
35 |
+
class NeRFMLP(BaseModule):
|
36 |
+
@dataclass
|
37 |
+
class Config(BaseModule.Config):
|
38 |
+
in_channels: int
|
39 |
+
n_neurons: int
|
40 |
+
n_hidden_layers: int
|
41 |
+
activation: str = "relu"
|
42 |
+
bias: bool = True
|
43 |
+
weight_init: Optional[str] = "kaiming_uniform"
|
44 |
+
bias_init: Optional[str] = None
|
45 |
+
|
46 |
+
cfg: Config
|
47 |
+
|
48 |
+
def configure(self) -> None:
|
49 |
+
layers = [
|
50 |
+
self.make_linear(
|
51 |
+
self.cfg.in_channels,
|
52 |
+
self.cfg.n_neurons,
|
53 |
+
bias=self.cfg.bias,
|
54 |
+
weight_init=self.cfg.weight_init,
|
55 |
+
bias_init=self.cfg.bias_init,
|
56 |
+
),
|
57 |
+
self.make_activation(self.cfg.activation),
|
58 |
+
]
|
59 |
+
for i in range(self.cfg.n_hidden_layers - 1):
|
60 |
+
layers += [
|
61 |
+
self.make_linear(
|
62 |
+
self.cfg.n_neurons,
|
63 |
+
self.cfg.n_neurons,
|
64 |
+
bias=self.cfg.bias,
|
65 |
+
weight_init=self.cfg.weight_init,
|
66 |
+
bias_init=self.cfg.bias_init,
|
67 |
+
),
|
68 |
+
self.make_activation(self.cfg.activation),
|
69 |
+
]
|
70 |
+
layers += [
|
71 |
+
self.make_linear(
|
72 |
+
self.cfg.n_neurons,
|
73 |
+
4, # density 1 + features 3
|
74 |
+
bias=self.cfg.bias,
|
75 |
+
weight_init=self.cfg.weight_init,
|
76 |
+
bias_init=self.cfg.bias_init,
|
77 |
+
)
|
78 |
+
]
|
79 |
+
self.layers = nn.Sequential(*layers)
|
80 |
+
|
81 |
+
def make_linear(
|
82 |
+
self,
|
83 |
+
dim_in,
|
84 |
+
dim_out,
|
85 |
+
bias=True,
|
86 |
+
weight_init=None,
|
87 |
+
bias_init=None,
|
88 |
+
):
|
89 |
+
layer = nn.Linear(dim_in, dim_out, bias=bias)
|
90 |
+
|
91 |
+
if weight_init is None:
|
92 |
+
pass
|
93 |
+
elif weight_init == "kaiming_uniform":
|
94 |
+
torch.nn.init.kaiming_uniform_(layer.weight, nonlinearity="relu")
|
95 |
+
else:
|
96 |
+
raise NotImplementedError
|
97 |
+
|
98 |
+
if bias:
|
99 |
+
if bias_init is None:
|
100 |
+
pass
|
101 |
+
elif bias_init == "zero":
|
102 |
+
torch.nn.init.zeros_(layer.bias)
|
103 |
+
else:
|
104 |
+
raise NotImplementedError
|
105 |
+
|
106 |
+
return layer
|
107 |
+
|
108 |
+
def make_activation(self, activation):
|
109 |
+
if activation == "relu":
|
110 |
+
return nn.ReLU(inplace=True)
|
111 |
+
elif activation == "silu":
|
112 |
+
return nn.SiLU(inplace=True)
|
113 |
+
else:
|
114 |
+
raise NotImplementedError
|
115 |
+
|
116 |
+
def forward(self, x):
|
117 |
+
inp_shape = x.shape[:-1]
|
118 |
+
x = x.reshape(-1, x.shape[-1])
|
119 |
+
|
120 |
+
features = self.layers(x)
|
121 |
+
features = features.reshape(*inp_shape, -1)
|
122 |
+
out = {"density": features[..., 0:1], "features": features[..., 1:4]}
|
123 |
+
|
124 |
+
return out
|
tsr/models/tokenizers/__pycache__/dino.cpython-310.pyc
ADDED
Binary file (18.6 kB). View file
|
|
tsr/models/tokenizers/__pycache__/image.cpython-310.pyc
ADDED
Binary file (2.42 kB). View file
|
|
tsr/models/tokenizers/__pycache__/triplane.cpython-310.pyc
ADDED
Binary file (1.76 kB). View file
|
|
tsr/models/tokenizers/image.py
ADDED
@@ -0,0 +1,67 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
from dataclasses import dataclass
|
2 |
+
from typing import Optional
|
3 |
+
|
4 |
+
import torch
|
5 |
+
import torch.nn as nn
|
6 |
+
from einops import rearrange
|
7 |
+
from huggingface_hub import hf_hub_download
|
8 |
+
from transformers.models.vit.modeling_vit import ViTModel
|
9 |
+
|
10 |
+
from ...utils import BaseModule
|
11 |
+
|
12 |
+
|
13 |
+
class DINOSingleImageTokenizer(BaseModule):
|
14 |
+
@dataclass
|
15 |
+
class Config(BaseModule.Config):
|
16 |
+
pretrained_model_name_or_path: str = "facebook/dino-vitb16"
|
17 |
+
enable_gradient_checkpointing: bool = False
|
18 |
+
|
19 |
+
cfg: Config
|
20 |
+
|
21 |
+
def configure(self) -> None:
|
22 |
+
self.model: ViTModel = ViTModel(
|
23 |
+
ViTModel.config_class.from_pretrained(
|
24 |
+
hf_hub_download(
|
25 |
+
repo_id=self.cfg.pretrained_model_name_or_path,
|
26 |
+
filename="config.json",
|
27 |
+
)
|
28 |
+
)
|
29 |
+
)
|
30 |
+
|
31 |
+
if self.cfg.enable_gradient_checkpointing:
|
32 |
+
self.model.encoder.gradient_checkpointing = True
|
33 |
+
|
34 |
+
self.register_buffer(
|
35 |
+
"image_mean",
|
36 |
+
torch.as_tensor([0.485, 0.456, 0.406]).reshape(1, 1, 3, 1, 1),
|
37 |
+
persistent=False,
|
38 |
+
)
|
39 |
+
self.register_buffer(
|
40 |
+
"image_std",
|
41 |
+
torch.as_tensor([0.229, 0.224, 0.225]).reshape(1, 1, 3, 1, 1),
|
42 |
+
persistent=False,
|
43 |
+
)
|
44 |
+
|
45 |
+
def forward(self, images: torch.FloatTensor, **kwargs) -> torch.FloatTensor:
|
46 |
+
packed = False
|
47 |
+
if images.ndim == 4:
|
48 |
+
packed = True
|
49 |
+
images = images.unsqueeze(1)
|
50 |
+
|
51 |
+
batch_size, n_input_views = images.shape[:2]
|
52 |
+
images = (images - self.image_mean) / self.image_std
|
53 |
+
out = self.model(
|
54 |
+
rearrange(images, "B N C H W -> (B N) C H W"), interpolate_pos_encoding=True
|
55 |
+
)
|
56 |
+
local_features, global_features = out.last_hidden_state, out.pooler_output
|
57 |
+
local_features = local_features.permute(0, 2, 1)
|
58 |
+
local_features = rearrange(
|
59 |
+
local_features, "(B N) Ct Nt -> B N Ct Nt", B=batch_size
|
60 |
+
)
|
61 |
+
if packed:
|
62 |
+
local_features = local_features.squeeze(1)
|
63 |
+
|
64 |
+
return local_features
|
65 |
+
|
66 |
+
def detokenize(self, *args, **kwargs):
|
67 |
+
raise NotImplementedError
|
tsr/models/tokenizers/triplane.py
ADDED
@@ -0,0 +1,45 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
import math
|
2 |
+
from dataclasses import dataclass
|
3 |
+
|
4 |
+
import torch
|
5 |
+
import torch.nn as nn
|
6 |
+
from einops import rearrange, repeat
|
7 |
+
|
8 |
+
from ...utils import BaseModule
|
9 |
+
|
10 |
+
|
11 |
+
class Triplane1DTokenizer(BaseModule):
|
12 |
+
@dataclass
|
13 |
+
class Config(BaseModule.Config):
|
14 |
+
plane_size: int
|
15 |
+
num_channels: int
|
16 |
+
|
17 |
+
cfg: Config
|
18 |
+
|
19 |
+
def configure(self) -> None:
|
20 |
+
self.embeddings = nn.Parameter(
|
21 |
+
torch.randn(
|
22 |
+
(3, self.cfg.num_channels, self.cfg.plane_size, self.cfg.plane_size),
|
23 |
+
dtype=torch.float32,
|
24 |
+
)
|
25 |
+
* 1
|
26 |
+
/ math.sqrt(self.cfg.num_channels)
|
27 |
+
)
|
28 |
+
|
29 |
+
def forward(self, batch_size: int) -> torch.Tensor:
|
30 |
+
return rearrange(
|
31 |
+
repeat(self.embeddings, "Np Ct Hp Wp -> B Np Ct Hp Wp", B=batch_size),
|
32 |
+
"B Np Ct Hp Wp -> B Ct (Np Hp Wp)",
|
33 |
+
)
|
34 |
+
|
35 |
+
def detokenize(self, tokens: torch.Tensor) -> torch.Tensor:
|
36 |
+
batch_size, Ct, Nt = tokens.shape
|
37 |
+
assert Nt == self.cfg.plane_size**2 * 3
|
38 |
+
assert Ct == self.cfg.num_channels
|
39 |
+
return rearrange(
|
40 |
+
tokens,
|
41 |
+
"B Ct (Np Hp Wp) -> B Np Ct Hp Wp",
|
42 |
+
Np=3,
|
43 |
+
Hp=self.cfg.plane_size,
|
44 |
+
Wp=self.cfg.plane_size,
|
45 |
+
)
|
tsr/models/transformer/__pycache__/attention.cpython-310.pyc
ADDED
Binary file (15.3 kB). View file
|
|
tsr/models/transformer/__pycache__/basic_transformer_block.cpython-310.pyc
ADDED
Binary file (9.96 kB). View file
|
|
tsr/models/transformer/__pycache__/transformer_1d.cpython-310.pyc
ADDED
Binary file (7.47 kB). View file
|
|
tsr/models/transformer/attention.py
ADDED
@@ -0,0 +1,628 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Copyright 2023 The HuggingFace Team. All rights reserved.
|
2 |
+
#
|
3 |
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
4 |
+
# you may not use this file except in compliance with the License.
|
5 |
+
# You may obtain a copy of the License at
|
6 |
+
#
|
7 |
+
# http://www.apache.org/licenses/LICENSE-2.0
|
8 |
+
#
|
9 |
+
# Unless required by applicable law or agreed to in writing, software
|
10 |
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
11 |
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
12 |
+
# See the License for the specific language governing permissions and
|
13 |
+
# limitations under the License.
|
14 |
+
from typing import Optional
|
15 |
+
|
16 |
+
import torch
|
17 |
+
import torch.nn.functional as F
|
18 |
+
from torch import nn
|
19 |
+
|
20 |
+
|
21 |
+
class Attention(nn.Module):
|
22 |
+
r"""
|
23 |
+
A cross attention layer.
|
24 |
+
|
25 |
+
Parameters:
|
26 |
+
query_dim (`int`):
|
27 |
+
The number of channels in the query.
|
28 |
+
cross_attention_dim (`int`, *optional*):
|
29 |
+
The number of channels in the encoder_hidden_states. If not given, defaults to `query_dim`.
|
30 |
+
heads (`int`, *optional*, defaults to 8):
|
31 |
+
The number of heads to use for multi-head attention.
|
32 |
+
dim_head (`int`, *optional*, defaults to 64):
|
33 |
+
The number of channels in each head.
|
34 |
+
dropout (`float`, *optional*, defaults to 0.0):
|
35 |
+
The dropout probability to use.
|
36 |
+
bias (`bool`, *optional*, defaults to False):
|
37 |
+
Set to `True` for the query, key, and value linear layers to contain a bias parameter.
|
38 |
+
upcast_attention (`bool`, *optional*, defaults to False):
|
39 |
+
Set to `True` to upcast the attention computation to `float32`.
|
40 |
+
upcast_softmax (`bool`, *optional*, defaults to False):
|
41 |
+
Set to `True` to upcast the softmax computation to `float32`.
|
42 |
+
cross_attention_norm (`str`, *optional*, defaults to `None`):
|
43 |
+
The type of normalization to use for the cross attention. Can be `None`, `layer_norm`, or `group_norm`.
|
44 |
+
cross_attention_norm_num_groups (`int`, *optional*, defaults to 32):
|
45 |
+
The number of groups to use for the group norm in the cross attention.
|
46 |
+
added_kv_proj_dim (`int`, *optional*, defaults to `None`):
|
47 |
+
The number of channels to use for the added key and value projections. If `None`, no projection is used.
|
48 |
+
norm_num_groups (`int`, *optional*, defaults to `None`):
|
49 |
+
The number of groups to use for the group norm in the attention.
|
50 |
+
spatial_norm_dim (`int`, *optional*, defaults to `None`):
|
51 |
+
The number of channels to use for the spatial normalization.
|
52 |
+
out_bias (`bool`, *optional*, defaults to `True`):
|
53 |
+
Set to `True` to use a bias in the output linear layer.
|
54 |
+
scale_qk (`bool`, *optional*, defaults to `True`):
|
55 |
+
Set to `True` to scale the query and key by `1 / sqrt(dim_head)`.
|
56 |
+
only_cross_attention (`bool`, *optional*, defaults to `False`):
|
57 |
+
Set to `True` to only use cross attention and not added_kv_proj_dim. Can only be set to `True` if
|
58 |
+
`added_kv_proj_dim` is not `None`.
|
59 |
+
eps (`float`, *optional*, defaults to 1e-5):
|
60 |
+
An additional value added to the denominator in group normalization that is used for numerical stability.
|
61 |
+
rescale_output_factor (`float`, *optional*, defaults to 1.0):
|
62 |
+
A factor to rescale the output by dividing it with this value.
|
63 |
+
residual_connection (`bool`, *optional*, defaults to `False`):
|
64 |
+
Set to `True` to add the residual connection to the output.
|
65 |
+
_from_deprecated_attn_block (`bool`, *optional*, defaults to `False`):
|
66 |
+
Set to `True` if the attention block is loaded from a deprecated state dict.
|
67 |
+
processor (`AttnProcessor`, *optional*, defaults to `None`):
|
68 |
+
The attention processor to use. If `None`, defaults to `AttnProcessor2_0` if `torch 2.x` is used and
|
69 |
+
`AttnProcessor` otherwise.
|
70 |
+
"""
|
71 |
+
|
72 |
+
def __init__(
|
73 |
+
self,
|
74 |
+
query_dim: int,
|
75 |
+
cross_attention_dim: Optional[int] = None,
|
76 |
+
heads: int = 8,
|
77 |
+
dim_head: int = 64,
|
78 |
+
dropout: float = 0.0,
|
79 |
+
bias: bool = False,
|
80 |
+
upcast_attention: bool = False,
|
81 |
+
upcast_softmax: bool = False,
|
82 |
+
cross_attention_norm: Optional[str] = None,
|
83 |
+
cross_attention_norm_num_groups: int = 32,
|
84 |
+
added_kv_proj_dim: Optional[int] = None,
|
85 |
+
norm_num_groups: Optional[int] = None,
|
86 |
+
out_bias: bool = True,
|
87 |
+
scale_qk: bool = True,
|
88 |
+
only_cross_attention: bool = False,
|
89 |
+
eps: float = 1e-5,
|
90 |
+
rescale_output_factor: float = 1.0,
|
91 |
+
residual_connection: bool = False,
|
92 |
+
_from_deprecated_attn_block: bool = False,
|
93 |
+
processor: Optional["AttnProcessor"] = None,
|
94 |
+
out_dim: int = None,
|
95 |
+
):
|
96 |
+
super().__init__()
|
97 |
+
self.inner_dim = out_dim if out_dim is not None else dim_head * heads
|
98 |
+
self.query_dim = query_dim
|
99 |
+
self.cross_attention_dim = (
|
100 |
+
cross_attention_dim if cross_attention_dim is not None else query_dim
|
101 |
+
)
|
102 |
+
self.upcast_attention = upcast_attention
|
103 |
+
self.upcast_softmax = upcast_softmax
|
104 |
+
self.rescale_output_factor = rescale_output_factor
|
105 |
+
self.residual_connection = residual_connection
|
106 |
+
self.dropout = dropout
|
107 |
+
self.fused_projections = False
|
108 |
+
self.out_dim = out_dim if out_dim is not None else query_dim
|
109 |
+
|
110 |
+
# we make use of this private variable to know whether this class is loaded
|
111 |
+
# with an deprecated state dict so that we can convert it on the fly
|
112 |
+
self._from_deprecated_attn_block = _from_deprecated_attn_block
|
113 |
+
|
114 |
+
self.scale_qk = scale_qk
|
115 |
+
self.scale = dim_head**-0.5 if self.scale_qk else 1.0
|
116 |
+
|
117 |
+
self.heads = out_dim // dim_head if out_dim is not None else heads
|
118 |
+
# for slice_size > 0 the attention score computation
|
119 |
+
# is split across the batch axis to save memory
|
120 |
+
# You can set slice_size with `set_attention_slice`
|
121 |
+
self.sliceable_head_dim = heads
|
122 |
+
|
123 |
+
self.added_kv_proj_dim = added_kv_proj_dim
|
124 |
+
self.only_cross_attention = only_cross_attention
|
125 |
+
|
126 |
+
if self.added_kv_proj_dim is None and self.only_cross_attention:
|
127 |
+
raise ValueError(
|
128 |
+
"`only_cross_attention` can only be set to True if `added_kv_proj_dim` is not None. Make sure to set either `only_cross_attention=False` or define `added_kv_proj_dim`."
|
129 |
+
)
|
130 |
+
|
131 |
+
if norm_num_groups is not None:
|
132 |
+
self.group_norm = nn.GroupNorm(
|
133 |
+
num_channels=query_dim, num_groups=norm_num_groups, eps=eps, affine=True
|
134 |
+
)
|
135 |
+
else:
|
136 |
+
self.group_norm = None
|
137 |
+
|
138 |
+
self.spatial_norm = None
|
139 |
+
|
140 |
+
if cross_attention_norm is None:
|
141 |
+
self.norm_cross = None
|
142 |
+
elif cross_attention_norm == "layer_norm":
|
143 |
+
self.norm_cross = nn.LayerNorm(self.cross_attention_dim)
|
144 |
+
elif cross_attention_norm == "group_norm":
|
145 |
+
if self.added_kv_proj_dim is not None:
|
146 |
+
# The given `encoder_hidden_states` are initially of shape
|
147 |
+
# (batch_size, seq_len, added_kv_proj_dim) before being projected
|
148 |
+
# to (batch_size, seq_len, cross_attention_dim). The norm is applied
|
149 |
+
# before the projection, so we need to use `added_kv_proj_dim` as
|
150 |
+
# the number of channels for the group norm.
|
151 |
+
norm_cross_num_channels = added_kv_proj_dim
|
152 |
+
else:
|
153 |
+
norm_cross_num_channels = self.cross_attention_dim
|
154 |
+
|
155 |
+
self.norm_cross = nn.GroupNorm(
|
156 |
+
num_channels=norm_cross_num_channels,
|
157 |
+
num_groups=cross_attention_norm_num_groups,
|
158 |
+
eps=1e-5,
|
159 |
+
affine=True,
|
160 |
+
)
|
161 |
+
else:
|
162 |
+
raise ValueError(
|
163 |
+
f"unknown cross_attention_norm: {cross_attention_norm}. Should be None, 'layer_norm' or 'group_norm'"
|
164 |
+
)
|
165 |
+
|
166 |
+
linear_cls = nn.Linear
|
167 |
+
|
168 |
+
self.linear_cls = linear_cls
|
169 |
+
self.to_q = linear_cls(query_dim, self.inner_dim, bias=bias)
|
170 |
+
|
171 |
+
if not self.only_cross_attention:
|
172 |
+
# only relevant for the `AddedKVProcessor` classes
|
173 |
+
self.to_k = linear_cls(self.cross_attention_dim, self.inner_dim, bias=bias)
|
174 |
+
self.to_v = linear_cls(self.cross_attention_dim, self.inner_dim, bias=bias)
|
175 |
+
else:
|
176 |
+
self.to_k = None
|
177 |
+
self.to_v = None
|
178 |
+
|
179 |
+
if self.added_kv_proj_dim is not None:
|
180 |
+
self.add_k_proj = linear_cls(added_kv_proj_dim, self.inner_dim)
|
181 |
+
self.add_v_proj = linear_cls(added_kv_proj_dim, self.inner_dim)
|
182 |
+
|
183 |
+
self.to_out = nn.ModuleList([])
|
184 |
+
self.to_out.append(linear_cls(self.inner_dim, self.out_dim, bias=out_bias))
|
185 |
+
self.to_out.append(nn.Dropout(dropout))
|
186 |
+
|
187 |
+
# set attention processor
|
188 |
+
# We use the AttnProcessor2_0 by default when torch 2.x is used which uses
|
189 |
+
# torch.nn.functional.scaled_dot_product_attention for native Flash/memory_efficient_attention
|
190 |
+
# but only if it has the default `scale` argument. TODO remove scale_qk check when we move to torch 2.1
|
191 |
+
if processor is None:
|
192 |
+
processor = (
|
193 |
+
AttnProcessor2_0()
|
194 |
+
if hasattr(F, "scaled_dot_product_attention") and self.scale_qk
|
195 |
+
else AttnProcessor()
|
196 |
+
)
|
197 |
+
self.set_processor(processor)
|
198 |
+
|
199 |
+
def set_processor(self, processor: "AttnProcessor") -> None:
|
200 |
+
self.processor = processor
|
201 |
+
|
202 |
+
def forward(
|
203 |
+
self,
|
204 |
+
hidden_states: torch.FloatTensor,
|
205 |
+
encoder_hidden_states: Optional[torch.FloatTensor] = None,
|
206 |
+
attention_mask: Optional[torch.FloatTensor] = None,
|
207 |
+
**cross_attention_kwargs,
|
208 |
+
) -> torch.Tensor:
|
209 |
+
r"""
|
210 |
+
The forward method of the `Attention` class.
|
211 |
+
|
212 |
+
Args:
|
213 |
+
hidden_states (`torch.Tensor`):
|
214 |
+
The hidden states of the query.
|
215 |
+
encoder_hidden_states (`torch.Tensor`, *optional*):
|
216 |
+
The hidden states of the encoder.
|
217 |
+
attention_mask (`torch.Tensor`, *optional*):
|
218 |
+
The attention mask to use. If `None`, no mask is applied.
|
219 |
+
**cross_attention_kwargs:
|
220 |
+
Additional keyword arguments to pass along to the cross attention.
|
221 |
+
|
222 |
+
Returns:
|
223 |
+
`torch.Tensor`: The output of the attention layer.
|
224 |
+
"""
|
225 |
+
# The `Attention` class can call different attention processors / attention functions
|
226 |
+
# here we simply pass along all tensors to the selected processor class
|
227 |
+
# For standard processors that are defined here, `**cross_attention_kwargs` is empty
|
228 |
+
return self.processor(
|
229 |
+
self,
|
230 |
+
hidden_states,
|
231 |
+
encoder_hidden_states=encoder_hidden_states,
|
232 |
+
attention_mask=attention_mask,
|
233 |
+
**cross_attention_kwargs,
|
234 |
+
)
|
235 |
+
|
236 |
+
def batch_to_head_dim(self, tensor: torch.Tensor) -> torch.Tensor:
|
237 |
+
r"""
|
238 |
+
Reshape the tensor from `[batch_size, seq_len, dim]` to `[batch_size // heads, seq_len, dim * heads]`. `heads`
|
239 |
+
is the number of heads initialized while constructing the `Attention` class.
|
240 |
+
|
241 |
+
Args:
|
242 |
+
tensor (`torch.Tensor`): The tensor to reshape.
|
243 |
+
|
244 |
+
Returns:
|
245 |
+
`torch.Tensor`: The reshaped tensor.
|
246 |
+
"""
|
247 |
+
head_size = self.heads
|
248 |
+
batch_size, seq_len, dim = tensor.shape
|
249 |
+
tensor = tensor.reshape(batch_size // head_size, head_size, seq_len, dim)
|
250 |
+
tensor = tensor.permute(0, 2, 1, 3).reshape(
|
251 |
+
batch_size // head_size, seq_len, dim * head_size
|
252 |
+
)
|
253 |
+
return tensor
|
254 |
+
|
255 |
+
def head_to_batch_dim(self, tensor: torch.Tensor, out_dim: int = 3) -> torch.Tensor:
|
256 |
+
r"""
|
257 |
+
Reshape the tensor from `[batch_size, seq_len, dim]` to `[batch_size, seq_len, heads, dim // heads]` `heads` is
|
258 |
+
the number of heads initialized while constructing the `Attention` class.
|
259 |
+
|
260 |
+
Args:
|
261 |
+
tensor (`torch.Tensor`): The tensor to reshape.
|
262 |
+
out_dim (`int`, *optional*, defaults to `3`): The output dimension of the tensor. If `3`, the tensor is
|
263 |
+
reshaped to `[batch_size * heads, seq_len, dim // heads]`.
|
264 |
+
|
265 |
+
Returns:
|
266 |
+
`torch.Tensor`: The reshaped tensor.
|
267 |
+
"""
|
268 |
+
head_size = self.heads
|
269 |
+
batch_size, seq_len, dim = tensor.shape
|
270 |
+
tensor = tensor.reshape(batch_size, seq_len, head_size, dim // head_size)
|
271 |
+
tensor = tensor.permute(0, 2, 1, 3)
|
272 |
+
|
273 |
+
if out_dim == 3:
|
274 |
+
tensor = tensor.reshape(batch_size * head_size, seq_len, dim // head_size)
|
275 |
+
|
276 |
+
return tensor
|
277 |
+
|
278 |
+
def get_attention_scores(
|
279 |
+
self,
|
280 |
+
query: torch.Tensor,
|
281 |
+
key: torch.Tensor,
|
282 |
+
attention_mask: torch.Tensor = None,
|
283 |
+
) -> torch.Tensor:
|
284 |
+
r"""
|
285 |
+
Compute the attention scores.
|
286 |
+
|
287 |
+
Args:
|
288 |
+
query (`torch.Tensor`): The query tensor.
|
289 |
+
key (`torch.Tensor`): The key tensor.
|
290 |
+
attention_mask (`torch.Tensor`, *optional*): The attention mask to use. If `None`, no mask is applied.
|
291 |
+
|
292 |
+
Returns:
|
293 |
+
`torch.Tensor`: The attention probabilities/scores.
|
294 |
+
"""
|
295 |
+
dtype = query.dtype
|
296 |
+
if self.upcast_attention:
|
297 |
+
query = query.float()
|
298 |
+
key = key.float()
|
299 |
+
|
300 |
+
if attention_mask is None:
|
301 |
+
baddbmm_input = torch.empty(
|
302 |
+
query.shape[0],
|
303 |
+
query.shape[1],
|
304 |
+
key.shape[1],
|
305 |
+
dtype=query.dtype,
|
306 |
+
device=query.device,
|
307 |
+
)
|
308 |
+
beta = 0
|
309 |
+
else:
|
310 |
+
baddbmm_input = attention_mask
|
311 |
+
beta = 1
|
312 |
+
|
313 |
+
attention_scores = torch.baddbmm(
|
314 |
+
baddbmm_input,
|
315 |
+
query,
|
316 |
+
key.transpose(-1, -2),
|
317 |
+
beta=beta,
|
318 |
+
alpha=self.scale,
|
319 |
+
)
|
320 |
+
del baddbmm_input
|
321 |
+
|
322 |
+
if self.upcast_softmax:
|
323 |
+
attention_scores = attention_scores.float()
|
324 |
+
|
325 |
+
attention_probs = attention_scores.softmax(dim=-1)
|
326 |
+
del attention_scores
|
327 |
+
|
328 |
+
attention_probs = attention_probs.to(dtype)
|
329 |
+
|
330 |
+
return attention_probs
|
331 |
+
|
332 |
+
def prepare_attention_mask(
|
333 |
+
self,
|
334 |
+
attention_mask: torch.Tensor,
|
335 |
+
target_length: int,
|
336 |
+
batch_size: int,
|
337 |
+
out_dim: int = 3,
|
338 |
+
) -> torch.Tensor:
|
339 |
+
r"""
|
340 |
+
Prepare the attention mask for the attention computation.
|
341 |
+
|
342 |
+
Args:
|
343 |
+
attention_mask (`torch.Tensor`):
|
344 |
+
The attention mask to prepare.
|
345 |
+
target_length (`int`):
|
346 |
+
The target length of the attention mask. This is the length of the attention mask after padding.
|
347 |
+
batch_size (`int`):
|
348 |
+
The batch size, which is used to repeat the attention mask.
|
349 |
+
out_dim (`int`, *optional*, defaults to `3`):
|
350 |
+
The output dimension of the attention mask. Can be either `3` or `4`.
|
351 |
+
|
352 |
+
Returns:
|
353 |
+
`torch.Tensor`: The prepared attention mask.
|
354 |
+
"""
|
355 |
+
head_size = self.heads
|
356 |
+
if attention_mask is None:
|
357 |
+
return attention_mask
|
358 |
+
|
359 |
+
current_length: int = attention_mask.shape[-1]
|
360 |
+
if current_length != target_length:
|
361 |
+
if attention_mask.device.type == "mps":
|
362 |
+
# HACK: MPS: Does not support padding by greater than dimension of input tensor.
|
363 |
+
# Instead, we can manually construct the padding tensor.
|
364 |
+
padding_shape = (
|
365 |
+
attention_mask.shape[0],
|
366 |
+
attention_mask.shape[1],
|
367 |
+
target_length,
|
368 |
+
)
|
369 |
+
padding = torch.zeros(
|
370 |
+
padding_shape,
|
371 |
+
dtype=attention_mask.dtype,
|
372 |
+
device=attention_mask.device,
|
373 |
+
)
|
374 |
+
attention_mask = torch.cat([attention_mask, padding], dim=2)
|
375 |
+
else:
|
376 |
+
# TODO: for pipelines such as stable-diffusion, padding cross-attn mask:
|
377 |
+
# we want to instead pad by (0, remaining_length), where remaining_length is:
|
378 |
+
# remaining_length: int = target_length - current_length
|
379 |
+
# TODO: re-enable tests/models/test_models_unet_2d_condition.py#test_model_xattn_padding
|
380 |
+
attention_mask = F.pad(attention_mask, (0, target_length), value=0.0)
|
381 |
+
|
382 |
+
if out_dim == 3:
|
383 |
+
if attention_mask.shape[0] < batch_size * head_size:
|
384 |
+
attention_mask = attention_mask.repeat_interleave(head_size, dim=0)
|
385 |
+
elif out_dim == 4:
|
386 |
+
attention_mask = attention_mask.unsqueeze(1)
|
387 |
+
attention_mask = attention_mask.repeat_interleave(head_size, dim=1)
|
388 |
+
|
389 |
+
return attention_mask
|
390 |
+
|
391 |
+
def norm_encoder_hidden_states(
|
392 |
+
self, encoder_hidden_states: torch.Tensor
|
393 |
+
) -> torch.Tensor:
|
394 |
+
r"""
|
395 |
+
Normalize the encoder hidden states. Requires `self.norm_cross` to be specified when constructing the
|
396 |
+
`Attention` class.
|
397 |
+
|
398 |
+
Args:
|
399 |
+
encoder_hidden_states (`torch.Tensor`): Hidden states of the encoder.
|
400 |
+
|
401 |
+
Returns:
|
402 |
+
`torch.Tensor`: The normalized encoder hidden states.
|
403 |
+
"""
|
404 |
+
assert (
|
405 |
+
self.norm_cross is not None
|
406 |
+
), "self.norm_cross must be defined to call self.norm_encoder_hidden_states"
|
407 |
+
|
408 |
+
if isinstance(self.norm_cross, nn.LayerNorm):
|
409 |
+
encoder_hidden_states = self.norm_cross(encoder_hidden_states)
|
410 |
+
elif isinstance(self.norm_cross, nn.GroupNorm):
|
411 |
+
# Group norm norms along the channels dimension and expects
|
412 |
+
# input to be in the shape of (N, C, *). In this case, we want
|
413 |
+
# to norm along the hidden dimension, so we need to move
|
414 |
+
# (batch_size, sequence_length, hidden_size) ->
|
415 |
+
# (batch_size, hidden_size, sequence_length)
|
416 |
+
encoder_hidden_states = encoder_hidden_states.transpose(1, 2)
|
417 |
+
encoder_hidden_states = self.norm_cross(encoder_hidden_states)
|
418 |
+
encoder_hidden_states = encoder_hidden_states.transpose(1, 2)
|
419 |
+
else:
|
420 |
+
assert False
|
421 |
+
|
422 |
+
return encoder_hidden_states
|
423 |
+
|
424 |
+
@torch.no_grad()
|
425 |
+
def fuse_projections(self, fuse=True):
|
426 |
+
is_cross_attention = self.cross_attention_dim != self.query_dim
|
427 |
+
device = self.to_q.weight.data.device
|
428 |
+
dtype = self.to_q.weight.data.dtype
|
429 |
+
|
430 |
+
if not is_cross_attention:
|
431 |
+
# fetch weight matrices.
|
432 |
+
concatenated_weights = torch.cat(
|
433 |
+
[self.to_q.weight.data, self.to_k.weight.data, self.to_v.weight.data]
|
434 |
+
)
|
435 |
+
in_features = concatenated_weights.shape[1]
|
436 |
+
out_features = concatenated_weights.shape[0]
|
437 |
+
|
438 |
+
# create a new single projection layer and copy over the weights.
|
439 |
+
self.to_qkv = self.linear_cls(
|
440 |
+
in_features, out_features, bias=False, device=device, dtype=dtype
|
441 |
+
)
|
442 |
+
self.to_qkv.weight.copy_(concatenated_weights)
|
443 |
+
|
444 |
+
else:
|
445 |
+
concatenated_weights = torch.cat(
|
446 |
+
[self.to_k.weight.data, self.to_v.weight.data]
|
447 |
+
)
|
448 |
+
in_features = concatenated_weights.shape[1]
|
449 |
+
out_features = concatenated_weights.shape[0]
|
450 |
+
|
451 |
+
self.to_kv = self.linear_cls(
|
452 |
+
in_features, out_features, bias=False, device=device, dtype=dtype
|
453 |
+
)
|
454 |
+
self.to_kv.weight.copy_(concatenated_weights)
|
455 |
+
|
456 |
+
self.fused_projections = fuse
|
457 |
+
|
458 |
+
|
459 |
+
class AttnProcessor:
|
460 |
+
r"""
|
461 |
+
Default processor for performing attention-related computations.
|
462 |
+
"""
|
463 |
+
|
464 |
+
def __call__(
|
465 |
+
self,
|
466 |
+
attn: Attention,
|
467 |
+
hidden_states: torch.FloatTensor,
|
468 |
+
encoder_hidden_states: Optional[torch.FloatTensor] = None,
|
469 |
+
attention_mask: Optional[torch.FloatTensor] = None,
|
470 |
+
) -> torch.Tensor:
|
471 |
+
residual = hidden_states
|
472 |
+
|
473 |
+
input_ndim = hidden_states.ndim
|
474 |
+
|
475 |
+
if input_ndim == 4:
|
476 |
+
batch_size, channel, height, width = hidden_states.shape
|
477 |
+
hidden_states = hidden_states.view(
|
478 |
+
batch_size, channel, height * width
|
479 |
+
).transpose(1, 2)
|
480 |
+
|
481 |
+
batch_size, sequence_length, _ = (
|
482 |
+
hidden_states.shape
|
483 |
+
if encoder_hidden_states is None
|
484 |
+
else encoder_hidden_states.shape
|
485 |
+
)
|
486 |
+
attention_mask = attn.prepare_attention_mask(
|
487 |
+
attention_mask, sequence_length, batch_size
|
488 |
+
)
|
489 |
+
|
490 |
+
if attn.group_norm is not None:
|
491 |
+
hidden_states = attn.group_norm(hidden_states.transpose(1, 2)).transpose(
|
492 |
+
1, 2
|
493 |
+
)
|
494 |
+
|
495 |
+
query = attn.to_q(hidden_states)
|
496 |
+
|
497 |
+
if encoder_hidden_states is None:
|
498 |
+
encoder_hidden_states = hidden_states
|
499 |
+
elif attn.norm_cross:
|
500 |
+
encoder_hidden_states = attn.norm_encoder_hidden_states(
|
501 |
+
encoder_hidden_states
|
502 |
+
)
|
503 |
+
|
504 |
+
key = attn.to_k(encoder_hidden_states)
|
505 |
+
value = attn.to_v(encoder_hidden_states)
|
506 |
+
|
507 |
+
query = attn.head_to_batch_dim(query)
|
508 |
+
key = attn.head_to_batch_dim(key)
|
509 |
+
value = attn.head_to_batch_dim(value)
|
510 |
+
|
511 |
+
attention_probs = attn.get_attention_scores(query, key, attention_mask)
|
512 |
+
hidden_states = torch.bmm(attention_probs, value)
|
513 |
+
hidden_states = attn.batch_to_head_dim(hidden_states)
|
514 |
+
|
515 |
+
# linear proj
|
516 |
+
hidden_states = attn.to_out[0](hidden_states)
|
517 |
+
# dropout
|
518 |
+
hidden_states = attn.to_out[1](hidden_states)
|
519 |
+
|
520 |
+
if input_ndim == 4:
|
521 |
+
hidden_states = hidden_states.transpose(-1, -2).reshape(
|
522 |
+
batch_size, channel, height, width
|
523 |
+
)
|
524 |
+
|
525 |
+
if attn.residual_connection:
|
526 |
+
hidden_states = hidden_states + residual
|
527 |
+
|
528 |
+
hidden_states = hidden_states / attn.rescale_output_factor
|
529 |
+
|
530 |
+
return hidden_states
|
531 |
+
|
532 |
+
|
533 |
+
class AttnProcessor2_0:
|
534 |
+
r"""
|
535 |
+
Processor for implementing scaled dot-product attention (enabled by default if you're using PyTorch 2.0).
|
536 |
+
"""
|
537 |
+
|
538 |
+
def __init__(self):
|
539 |
+
if not hasattr(F, "scaled_dot_product_attention"):
|
540 |
+
raise ImportError(
|
541 |
+
"AttnProcessor2_0 requires PyTorch 2.0, to use it, please upgrade PyTorch to 2.0."
|
542 |
+
)
|
543 |
+
|
544 |
+
def __call__(
|
545 |
+
self,
|
546 |
+
attn: Attention,
|
547 |
+
hidden_states: torch.FloatTensor,
|
548 |
+
encoder_hidden_states: Optional[torch.FloatTensor] = None,
|
549 |
+
attention_mask: Optional[torch.FloatTensor] = None,
|
550 |
+
) -> torch.FloatTensor:
|
551 |
+
residual = hidden_states
|
552 |
+
|
553 |
+
input_ndim = hidden_states.ndim
|
554 |
+
|
555 |
+
if input_ndim == 4:
|
556 |
+
batch_size, channel, height, width = hidden_states.shape
|
557 |
+
hidden_states = hidden_states.view(
|
558 |
+
batch_size, channel, height * width
|
559 |
+
).transpose(1, 2)
|
560 |
+
|
561 |
+
batch_size, sequence_length, _ = (
|
562 |
+
hidden_states.shape
|
563 |
+
if encoder_hidden_states is None
|
564 |
+
else encoder_hidden_states.shape
|
565 |
+
)
|
566 |
+
|
567 |
+
if attention_mask is not None:
|
568 |
+
attention_mask = attn.prepare_attention_mask(
|
569 |
+
attention_mask, sequence_length, batch_size
|
570 |
+
)
|
571 |
+
# scaled_dot_product_attention expects attention_mask shape to be
|
572 |
+
# (batch, heads, source_length, target_length)
|
573 |
+
attention_mask = attention_mask.view(
|
574 |
+
batch_size, attn.heads, -1, attention_mask.shape[-1]
|
575 |
+
)
|
576 |
+
|
577 |
+
if attn.group_norm is not None:
|
578 |
+
hidden_states = attn.group_norm(hidden_states.transpose(1, 2)).transpose(
|
579 |
+
1, 2
|
580 |
+
)
|
581 |
+
|
582 |
+
query = attn.to_q(hidden_states)
|
583 |
+
|
584 |
+
if encoder_hidden_states is None:
|
585 |
+
encoder_hidden_states = hidden_states
|
586 |
+
elif attn.norm_cross:
|
587 |
+
encoder_hidden_states = attn.norm_encoder_hidden_states(
|
588 |
+
encoder_hidden_states
|
589 |
+
)
|
590 |
+
|
591 |
+
key = attn.to_k(encoder_hidden_states)
|
592 |
+
value = attn.to_v(encoder_hidden_states)
|
593 |
+
|
594 |
+
inner_dim = key.shape[-1]
|
595 |
+
head_dim = inner_dim // attn.heads
|
596 |
+
|
597 |
+
query = query.view(batch_size, -1, attn.heads, head_dim).transpose(1, 2)
|
598 |
+
|
599 |
+
key = key.view(batch_size, -1, attn.heads, head_dim).transpose(1, 2)
|
600 |
+
value = value.view(batch_size, -1, attn.heads, head_dim).transpose(1, 2)
|
601 |
+
|
602 |
+
# the output of sdp = (batch, num_heads, seq_len, head_dim)
|
603 |
+
# TODO: add support for attn.scale when we move to Torch 2.1
|
604 |
+
hidden_states = F.scaled_dot_product_attention(
|
605 |
+
query, key, value, attn_mask=attention_mask, dropout_p=0.0, is_causal=False
|
606 |
+
)
|
607 |
+
|
608 |
+
hidden_states = hidden_states.transpose(1, 2).reshape(
|
609 |
+
batch_size, -1, attn.heads * head_dim
|
610 |
+
)
|
611 |
+
hidden_states = hidden_states.to(query.dtype)
|
612 |
+
|
613 |
+
# linear proj
|
614 |
+
hidden_states = attn.to_out[0](hidden_states)
|
615 |
+
# dropout
|
616 |
+
hidden_states = attn.to_out[1](hidden_states)
|
617 |
+
|
618 |
+
if input_ndim == 4:
|
619 |
+
hidden_states = hidden_states.transpose(-1, -2).reshape(
|
620 |
+
batch_size, channel, height, width
|
621 |
+
)
|
622 |
+
|
623 |
+
if attn.residual_connection:
|
624 |
+
hidden_states = hidden_states + residual
|
625 |
+
|
626 |
+
hidden_states = hidden_states / attn.rescale_output_factor
|
627 |
+
|
628 |
+
return hidden_states
|
tsr/models/transformer/basic_transformer_block.py
ADDED
@@ -0,0 +1,314 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Copyright 2023 The HuggingFace Team. All rights reserved.
|
2 |
+
#
|
3 |
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
4 |
+
# you may not use this file except in compliance with the License.
|
5 |
+
# You may obtain a copy of the License at
|
6 |
+
#
|
7 |
+
# http://www.apache.org/licenses/LICENSE-2.0
|
8 |
+
#
|
9 |
+
# Unless required by applicable law or agreed to in writing, software
|
10 |
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
11 |
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
12 |
+
# See the License for the specific language governing permissions and
|
13 |
+
# limitations under the License.
|
14 |
+
|
15 |
+
from typing import Any, Dict, Optional
|
16 |
+
|
17 |
+
import torch
|
18 |
+
import torch.nn.functional as F
|
19 |
+
from torch import nn
|
20 |
+
|
21 |
+
from .attention import Attention
|
22 |
+
|
23 |
+
|
24 |
+
class BasicTransformerBlock(nn.Module):
|
25 |
+
r"""
|
26 |
+
A basic Transformer block.
|
27 |
+
|
28 |
+
Parameters:
|
29 |
+
dim (`int`): The number of channels in the input and output.
|
30 |
+
num_attention_heads (`int`): The number of heads to use for multi-head attention.
|
31 |
+
attention_head_dim (`int`): The number of channels in each head.
|
32 |
+
dropout (`float`, *optional*, defaults to 0.0): The dropout probability to use.
|
33 |
+
cross_attention_dim (`int`, *optional*): The size of the encoder_hidden_states vector for cross attention.
|
34 |
+
activation_fn (`str`, *optional*, defaults to `"geglu"`): Activation function to be used in feed-forward.
|
35 |
+
num_embeds_ada_norm (:
|
36 |
+
obj: `int`, *optional*): The number of diffusion steps used during training. See `Transformer2DModel`.
|
37 |
+
attention_bias (:
|
38 |
+
obj: `bool`, *optional*, defaults to `False`): Configure if the attentions should contain a bias parameter.
|
39 |
+
only_cross_attention (`bool`, *optional*):
|
40 |
+
Whether to use only cross-attention layers. In this case two cross attention layers are used.
|
41 |
+
double_self_attention (`bool`, *optional*):
|
42 |
+
Whether to use two self-attention layers. In this case no cross attention layers are used.
|
43 |
+
upcast_attention (`bool`, *optional*):
|
44 |
+
Whether to upcast the attention computation to float32. This is useful for mixed precision training.
|
45 |
+
norm_elementwise_affine (`bool`, *optional*, defaults to `True`):
|
46 |
+
Whether to use learnable elementwise affine parameters for normalization.
|
47 |
+
norm_type (`str`, *optional*, defaults to `"layer_norm"`):
|
48 |
+
The normalization layer to use. Can be `"layer_norm"`, `"ada_norm"` or `"ada_norm_zero"`.
|
49 |
+
final_dropout (`bool` *optional*, defaults to False):
|
50 |
+
Whether to apply a final dropout after the last feed-forward layer.
|
51 |
+
attention_type (`str`, *optional*, defaults to `"default"`):
|
52 |
+
The type of attention to use. Can be `"default"` or `"gated"` or `"gated-text-image"`.
|
53 |
+
"""
|
54 |
+
|
55 |
+
def __init__(
|
56 |
+
self,
|
57 |
+
dim: int,
|
58 |
+
num_attention_heads: int,
|
59 |
+
attention_head_dim: int,
|
60 |
+
dropout=0.0,
|
61 |
+
cross_attention_dim: Optional[int] = None,
|
62 |
+
activation_fn: str = "geglu",
|
63 |
+
attention_bias: bool = False,
|
64 |
+
only_cross_attention: bool = False,
|
65 |
+
double_self_attention: bool = False,
|
66 |
+
upcast_attention: bool = False,
|
67 |
+
norm_elementwise_affine: bool = True,
|
68 |
+
norm_type: str = "layer_norm",
|
69 |
+
final_dropout: bool = False,
|
70 |
+
):
|
71 |
+
super().__init__()
|
72 |
+
self.only_cross_attention = only_cross_attention
|
73 |
+
|
74 |
+
assert norm_type == "layer_norm"
|
75 |
+
|
76 |
+
# Define 3 blocks. Each block has its own normalization layer.
|
77 |
+
# 1. Self-Attn
|
78 |
+
self.norm1 = nn.LayerNorm(dim, elementwise_affine=norm_elementwise_affine)
|
79 |
+
self.attn1 = Attention(
|
80 |
+
query_dim=dim,
|
81 |
+
heads=num_attention_heads,
|
82 |
+
dim_head=attention_head_dim,
|
83 |
+
dropout=dropout,
|
84 |
+
bias=attention_bias,
|
85 |
+
cross_attention_dim=cross_attention_dim if only_cross_attention else None,
|
86 |
+
upcast_attention=upcast_attention,
|
87 |
+
)
|
88 |
+
|
89 |
+
# 2. Cross-Attn
|
90 |
+
if cross_attention_dim is not None or double_self_attention:
|
91 |
+
# We currently only use AdaLayerNormZero for self attention where there will only be one attention block.
|
92 |
+
# I.e. the number of returned modulation chunks from AdaLayerZero would not make sense if returned during
|
93 |
+
# the second cross attention block.
|
94 |
+
self.norm2 = nn.LayerNorm(dim, elementwise_affine=norm_elementwise_affine)
|
95 |
+
|
96 |
+
self.attn2 = Attention(
|
97 |
+
query_dim=dim,
|
98 |
+
cross_attention_dim=cross_attention_dim
|
99 |
+
if not double_self_attention
|
100 |
+
else None,
|
101 |
+
heads=num_attention_heads,
|
102 |
+
dim_head=attention_head_dim,
|
103 |
+
dropout=dropout,
|
104 |
+
bias=attention_bias,
|
105 |
+
upcast_attention=upcast_attention,
|
106 |
+
) # is self-attn if encoder_hidden_states is none
|
107 |
+
else:
|
108 |
+
self.norm2 = None
|
109 |
+
self.attn2 = None
|
110 |
+
|
111 |
+
# 3. Feed-forward
|
112 |
+
self.norm3 = nn.LayerNorm(dim, elementwise_affine=norm_elementwise_affine)
|
113 |
+
self.ff = FeedForward(
|
114 |
+
dim,
|
115 |
+
dropout=dropout,
|
116 |
+
activation_fn=activation_fn,
|
117 |
+
final_dropout=final_dropout,
|
118 |
+
)
|
119 |
+
|
120 |
+
# let chunk size default to None
|
121 |
+
self._chunk_size = None
|
122 |
+
self._chunk_dim = 0
|
123 |
+
|
124 |
+
def set_chunk_feed_forward(self, chunk_size: Optional[int], dim: int):
|
125 |
+
# Sets chunk feed-forward
|
126 |
+
self._chunk_size = chunk_size
|
127 |
+
self._chunk_dim = dim
|
128 |
+
|
129 |
+
def forward(
|
130 |
+
self,
|
131 |
+
hidden_states: torch.FloatTensor,
|
132 |
+
attention_mask: Optional[torch.FloatTensor] = None,
|
133 |
+
encoder_hidden_states: Optional[torch.FloatTensor] = None,
|
134 |
+
encoder_attention_mask: Optional[torch.FloatTensor] = None,
|
135 |
+
) -> torch.FloatTensor:
|
136 |
+
# Notice that normalization is always applied before the real computation in the following blocks.
|
137 |
+
# 0. Self-Attention
|
138 |
+
norm_hidden_states = self.norm1(hidden_states)
|
139 |
+
|
140 |
+
attn_output = self.attn1(
|
141 |
+
norm_hidden_states,
|
142 |
+
encoder_hidden_states=encoder_hidden_states
|
143 |
+
if self.only_cross_attention
|
144 |
+
else None,
|
145 |
+
attention_mask=attention_mask,
|
146 |
+
)
|
147 |
+
|
148 |
+
hidden_states = attn_output + hidden_states
|
149 |
+
|
150 |
+
# 3. Cross-Attention
|
151 |
+
if self.attn2 is not None:
|
152 |
+
norm_hidden_states = self.norm2(hidden_states)
|
153 |
+
|
154 |
+
attn_output = self.attn2(
|
155 |
+
norm_hidden_states,
|
156 |
+
encoder_hidden_states=encoder_hidden_states,
|
157 |
+
attention_mask=encoder_attention_mask,
|
158 |
+
)
|
159 |
+
hidden_states = attn_output + hidden_states
|
160 |
+
|
161 |
+
# 4. Feed-forward
|
162 |
+
norm_hidden_states = self.norm3(hidden_states)
|
163 |
+
|
164 |
+
if self._chunk_size is not None:
|
165 |
+
# "feed_forward_chunk_size" can be used to save memory
|
166 |
+
if norm_hidden_states.shape[self._chunk_dim] % self._chunk_size != 0:
|
167 |
+
raise ValueError(
|
168 |
+
f"`hidden_states` dimension to be chunked: {norm_hidden_states.shape[self._chunk_dim]} has to be divisible by chunk size: {self._chunk_size}. Make sure to set an appropriate `chunk_size` when calling `unet.enable_forward_chunking`."
|
169 |
+
)
|
170 |
+
|
171 |
+
num_chunks = norm_hidden_states.shape[self._chunk_dim] // self._chunk_size
|
172 |
+
ff_output = torch.cat(
|
173 |
+
[
|
174 |
+
self.ff(hid_slice)
|
175 |
+
for hid_slice in norm_hidden_states.chunk(
|
176 |
+
num_chunks, dim=self._chunk_dim
|
177 |
+
)
|
178 |
+
],
|
179 |
+
dim=self._chunk_dim,
|
180 |
+
)
|
181 |
+
else:
|
182 |
+
ff_output = self.ff(norm_hidden_states)
|
183 |
+
|
184 |
+
hidden_states = ff_output + hidden_states
|
185 |
+
|
186 |
+
return hidden_states
|
187 |
+
|
188 |
+
|
189 |
+
class FeedForward(nn.Module):
|
190 |
+
r"""
|
191 |
+
A feed-forward layer.
|
192 |
+
|
193 |
+
Parameters:
|
194 |
+
dim (`int`): The number of channels in the input.
|
195 |
+
dim_out (`int`, *optional*): The number of channels in the output. If not given, defaults to `dim`.
|
196 |
+
mult (`int`, *optional*, defaults to 4): The multiplier to use for the hidden dimension.
|
197 |
+
dropout (`float`, *optional*, defaults to 0.0): The dropout probability to use.
|
198 |
+
activation_fn (`str`, *optional*, defaults to `"geglu"`): Activation function to be used in feed-forward.
|
199 |
+
final_dropout (`bool` *optional*, defaults to False): Apply a final dropout.
|
200 |
+
"""
|
201 |
+
|
202 |
+
def __init__(
|
203 |
+
self,
|
204 |
+
dim: int,
|
205 |
+
dim_out: Optional[int] = None,
|
206 |
+
mult: int = 4,
|
207 |
+
dropout: float = 0.0,
|
208 |
+
activation_fn: str = "geglu",
|
209 |
+
final_dropout: bool = False,
|
210 |
+
):
|
211 |
+
super().__init__()
|
212 |
+
inner_dim = int(dim * mult)
|
213 |
+
dim_out = dim_out if dim_out is not None else dim
|
214 |
+
linear_cls = nn.Linear
|
215 |
+
|
216 |
+
if activation_fn == "gelu":
|
217 |
+
act_fn = GELU(dim, inner_dim)
|
218 |
+
if activation_fn == "gelu-approximate":
|
219 |
+
act_fn = GELU(dim, inner_dim, approximate="tanh")
|
220 |
+
elif activation_fn == "geglu":
|
221 |
+
act_fn = GEGLU(dim, inner_dim)
|
222 |
+
elif activation_fn == "geglu-approximate":
|
223 |
+
act_fn = ApproximateGELU(dim, inner_dim)
|
224 |
+
|
225 |
+
self.net = nn.ModuleList([])
|
226 |
+
# project in
|
227 |
+
self.net.append(act_fn)
|
228 |
+
# project dropout
|
229 |
+
self.net.append(nn.Dropout(dropout))
|
230 |
+
# project out
|
231 |
+
self.net.append(linear_cls(inner_dim, dim_out))
|
232 |
+
# FF as used in Vision Transformer, MLP-Mixer, etc. have a final dropout
|
233 |
+
if final_dropout:
|
234 |
+
self.net.append(nn.Dropout(dropout))
|
235 |
+
|
236 |
+
def forward(self, hidden_states: torch.Tensor) -> torch.Tensor:
|
237 |
+
for module in self.net:
|
238 |
+
hidden_states = module(hidden_states)
|
239 |
+
return hidden_states
|
240 |
+
|
241 |
+
|
242 |
+
class GELU(nn.Module):
|
243 |
+
r"""
|
244 |
+
GELU activation function with tanh approximation support with `approximate="tanh"`.
|
245 |
+
|
246 |
+
Parameters:
|
247 |
+
dim_in (`int`): The number of channels in the input.
|
248 |
+
dim_out (`int`): The number of channels in the output.
|
249 |
+
approximate (`str`, *optional*, defaults to `"none"`): If `"tanh"`, use tanh approximation.
|
250 |
+
"""
|
251 |
+
|
252 |
+
def __init__(self, dim_in: int, dim_out: int, approximate: str = "none"):
|
253 |
+
super().__init__()
|
254 |
+
self.proj = nn.Linear(dim_in, dim_out)
|
255 |
+
self.approximate = approximate
|
256 |
+
|
257 |
+
def gelu(self, gate: torch.Tensor) -> torch.Tensor:
|
258 |
+
if gate.device.type != "mps":
|
259 |
+
return F.gelu(gate, approximate=self.approximate)
|
260 |
+
# mps: gelu is not implemented for float16
|
261 |
+
return F.gelu(gate.to(dtype=torch.float32), approximate=self.approximate).to(
|
262 |
+
dtype=gate.dtype
|
263 |
+
)
|
264 |
+
|
265 |
+
def forward(self, hidden_states):
|
266 |
+
hidden_states = self.proj(hidden_states)
|
267 |
+
hidden_states = self.gelu(hidden_states)
|
268 |
+
return hidden_states
|
269 |
+
|
270 |
+
|
271 |
+
class GEGLU(nn.Module):
|
272 |
+
r"""
|
273 |
+
A variant of the gated linear unit activation function from https://arxiv.org/abs/2002.05202.
|
274 |
+
|
275 |
+
Parameters:
|
276 |
+
dim_in (`int`): The number of channels in the input.
|
277 |
+
dim_out (`int`): The number of channels in the output.
|
278 |
+
"""
|
279 |
+
|
280 |
+
def __init__(self, dim_in: int, dim_out: int):
|
281 |
+
super().__init__()
|
282 |
+
linear_cls = nn.Linear
|
283 |
+
|
284 |
+
self.proj = linear_cls(dim_in, dim_out * 2)
|
285 |
+
|
286 |
+
def gelu(self, gate: torch.Tensor) -> torch.Tensor:
|
287 |
+
if gate.device.type != "mps":
|
288 |
+
return F.gelu(gate)
|
289 |
+
# mps: gelu is not implemented for float16
|
290 |
+
return F.gelu(gate.to(dtype=torch.float32)).to(dtype=gate.dtype)
|
291 |
+
|
292 |
+
def forward(self, hidden_states, scale: float = 1.0):
|
293 |
+
args = ()
|
294 |
+
hidden_states, gate = self.proj(hidden_states, *args).chunk(2, dim=-1)
|
295 |
+
return hidden_states * self.gelu(gate)
|
296 |
+
|
297 |
+
|
298 |
+
class ApproximateGELU(nn.Module):
|
299 |
+
r"""
|
300 |
+
The approximate form of Gaussian Error Linear Unit (GELU). For more details, see section 2:
|
301 |
+
https://arxiv.org/abs/1606.08415.
|
302 |
+
|
303 |
+
Parameters:
|
304 |
+
dim_in (`int`): The number of channels in the input.
|
305 |
+
dim_out (`int`): The number of channels in the output.
|
306 |
+
"""
|
307 |
+
|
308 |
+
def __init__(self, dim_in: int, dim_out: int):
|
309 |
+
super().__init__()
|
310 |
+
self.proj = nn.Linear(dim_in, dim_out)
|
311 |
+
|
312 |
+
def forward(self, x: torch.Tensor) -> torch.Tensor:
|
313 |
+
x = self.proj(x)
|
314 |
+
return x * torch.sigmoid(1.702 * x)
|
tsr/models/transformer/transformer_1d.py
ADDED
@@ -0,0 +1,216 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
from dataclasses import dataclass, field
|
2 |
+
from typing import Any, Dict, Optional
|
3 |
+
|
4 |
+
import torch
|
5 |
+
import torch.nn.functional as F
|
6 |
+
from torch import nn
|
7 |
+
|
8 |
+
from ...utils import BaseModule
|
9 |
+
from .basic_transformer_block import BasicTransformerBlock
|
10 |
+
|
11 |
+
|
12 |
+
class Transformer1D(BaseModule):
|
13 |
+
"""
|
14 |
+
A 1D Transformer model for sequence data.
|
15 |
+
|
16 |
+
Parameters:
|
17 |
+
num_attention_heads (`int`, *optional*, defaults to 16): The number of heads to use for multi-head attention.
|
18 |
+
attention_head_dim (`int`, *optional*, defaults to 88): The number of channels in each head.
|
19 |
+
in_channels (`int`, *optional*):
|
20 |
+
The number of channels in the input and output (specify if the input is **continuous**).
|
21 |
+
num_layers (`int`, *optional*, defaults to 1): The number of layers of Transformer blocks to use.
|
22 |
+
dropout (`float`, *optional*, defaults to 0.0): The dropout probability to use.
|
23 |
+
cross_attention_dim (`int`, *optional*): The number of `encoder_hidden_states` dimensions to use.
|
24 |
+
activation_fn (`str`, *optional*, defaults to `"geglu"`): Activation function to use in feed-forward.
|
25 |
+
num_embeds_ada_norm ( `int`, *optional*):
|
26 |
+
The number of diffusion steps used during training. Pass if at least one of the norm_layers is
|
27 |
+
`AdaLayerNorm`. This is fixed during training since it is used to learn a number of embeddings that are
|
28 |
+
added to the hidden states.
|
29 |
+
|
30 |
+
During inference, you can denoise for up to but not more steps than `num_embeds_ada_norm`.
|
31 |
+
attention_bias (`bool`, *optional*):
|
32 |
+
Configure if the `TransformerBlocks` attention should contain a bias parameter.
|
33 |
+
"""
|
34 |
+
|
35 |
+
@dataclass
|
36 |
+
class Config(BaseModule.Config):
|
37 |
+
num_attention_heads: int = 16
|
38 |
+
attention_head_dim: int = 88
|
39 |
+
in_channels: Optional[int] = None
|
40 |
+
out_channels: Optional[int] = None
|
41 |
+
num_layers: int = 1
|
42 |
+
dropout: float = 0.0
|
43 |
+
norm_num_groups: int = 32
|
44 |
+
cross_attention_dim: Optional[int] = None
|
45 |
+
attention_bias: bool = False
|
46 |
+
activation_fn: str = "geglu"
|
47 |
+
only_cross_attention: bool = False
|
48 |
+
double_self_attention: bool = False
|
49 |
+
upcast_attention: bool = False
|
50 |
+
norm_type: str = "layer_norm"
|
51 |
+
norm_elementwise_affine: bool = True
|
52 |
+
gradient_checkpointing: bool = False
|
53 |
+
|
54 |
+
cfg: Config
|
55 |
+
|
56 |
+
def configure(self) -> None:
|
57 |
+
self.num_attention_heads = self.cfg.num_attention_heads
|
58 |
+
self.attention_head_dim = self.cfg.attention_head_dim
|
59 |
+
inner_dim = self.num_attention_heads * self.attention_head_dim
|
60 |
+
|
61 |
+
linear_cls = nn.Linear
|
62 |
+
|
63 |
+
# 2. Define input layers
|
64 |
+
self.in_channels = self.cfg.in_channels
|
65 |
+
|
66 |
+
self.norm = torch.nn.GroupNorm(
|
67 |
+
num_groups=self.cfg.norm_num_groups,
|
68 |
+
num_channels=self.cfg.in_channels,
|
69 |
+
eps=1e-6,
|
70 |
+
affine=True,
|
71 |
+
)
|
72 |
+
self.proj_in = linear_cls(self.cfg.in_channels, inner_dim)
|
73 |
+
|
74 |
+
# 3. Define transformers blocks
|
75 |
+
self.transformer_blocks = nn.ModuleList(
|
76 |
+
[
|
77 |
+
BasicTransformerBlock(
|
78 |
+
inner_dim,
|
79 |
+
self.num_attention_heads,
|
80 |
+
self.attention_head_dim,
|
81 |
+
dropout=self.cfg.dropout,
|
82 |
+
cross_attention_dim=self.cfg.cross_attention_dim,
|
83 |
+
activation_fn=self.cfg.activation_fn,
|
84 |
+
attention_bias=self.cfg.attention_bias,
|
85 |
+
only_cross_attention=self.cfg.only_cross_attention,
|
86 |
+
double_self_attention=self.cfg.double_self_attention,
|
87 |
+
upcast_attention=self.cfg.upcast_attention,
|
88 |
+
norm_type=self.cfg.norm_type,
|
89 |
+
norm_elementwise_affine=self.cfg.norm_elementwise_affine,
|
90 |
+
)
|
91 |
+
for d in range(self.cfg.num_layers)
|
92 |
+
]
|
93 |
+
)
|
94 |
+
|
95 |
+
# 4. Define output layers
|
96 |
+
self.out_channels = (
|
97 |
+
self.cfg.in_channels
|
98 |
+
if self.cfg.out_channels is None
|
99 |
+
else self.cfg.out_channels
|
100 |
+
)
|
101 |
+
|
102 |
+
self.proj_out = linear_cls(inner_dim, self.cfg.in_channels)
|
103 |
+
|
104 |
+
self.gradient_checkpointing = self.cfg.gradient_checkpointing
|
105 |
+
|
106 |
+
def forward(
|
107 |
+
self,
|
108 |
+
hidden_states: torch.Tensor,
|
109 |
+
encoder_hidden_states: Optional[torch.Tensor] = None,
|
110 |
+
attention_mask: Optional[torch.Tensor] = None,
|
111 |
+
encoder_attention_mask: Optional[torch.Tensor] = None,
|
112 |
+
):
|
113 |
+
"""
|
114 |
+
The [`Transformer1DModel`] forward method.
|
115 |
+
|
116 |
+
Args:
|
117 |
+
hidden_states (`torch.LongTensor` of shape `(batch size, num latent pixels)` if discrete, `torch.FloatTensor` of shape `(batch size, channel, height, width)` if continuous):
|
118 |
+
Input `hidden_states`.
|
119 |
+
encoder_hidden_states ( `torch.FloatTensor` of shape `(batch size, sequence len, embed dims)`, *optional*):
|
120 |
+
Conditional embeddings for cross attention layer. If not given, cross-attention defaults to
|
121 |
+
self-attention.
|
122 |
+
timestep ( `torch.LongTensor`, *optional*):
|
123 |
+
Used to indicate denoising step. Optional timestep to be applied as an embedding in `AdaLayerNorm`.
|
124 |
+
class_labels ( `torch.LongTensor` of shape `(batch size, num classes)`, *optional*):
|
125 |
+
Used to indicate class labels conditioning. Optional class labels to be applied as an embedding in
|
126 |
+
`AdaLayerZeroNorm`.
|
127 |
+
cross_attention_kwargs ( `Dict[str, Any]`, *optional*):
|
128 |
+
A kwargs dictionary that if specified is passed along to the `AttentionProcessor` as defined under
|
129 |
+
`self.processor` in
|
130 |
+
[diffusers.models.attention_processor](https://github.com/huggingface/diffusers/blob/main/src/diffusers/models/attention_processor.py).
|
131 |
+
attention_mask ( `torch.Tensor`, *optional*):
|
132 |
+
An attention mask of shape `(batch, key_tokens)` is applied to `encoder_hidden_states`. If `1` the mask
|
133 |
+
is kept, otherwise if `0` it is discarded. Mask will be converted into a bias, which adds large
|
134 |
+
negative values to the attention scores corresponding to "discard" tokens.
|
135 |
+
encoder_attention_mask ( `torch.Tensor`, *optional*):
|
136 |
+
Cross-attention mask applied to `encoder_hidden_states`. Two formats supported:
|
137 |
+
|
138 |
+
* Mask `(batch, sequence_length)` True = keep, False = discard.
|
139 |
+
* Bias `(batch, 1, sequence_length)` 0 = keep, -10000 = discard.
|
140 |
+
|
141 |
+
If `ndim == 2`: will be interpreted as a mask, then converted into a bias consistent with the format
|
142 |
+
above. This bias will be added to the cross-attention scores.
|
143 |
+
return_dict (`bool`, *optional*, defaults to `True`):
|
144 |
+
Whether or not to return a [`~models.unet_2d_condition.UNet2DConditionOutput`] instead of a plain
|
145 |
+
tuple.
|
146 |
+
|
147 |
+
Returns:
|
148 |
+
If `return_dict` is True, an [`~models.transformer_2d.Transformer2DModelOutput`] is returned, otherwise a
|
149 |
+
`tuple` where the first element is the sample tensor.
|
150 |
+
"""
|
151 |
+
# ensure attention_mask is a bias, and give it a singleton query_tokens dimension.
|
152 |
+
# we may have done this conversion already, e.g. if we came here via UNet2DConditionModel#forward.
|
153 |
+
# we can tell by counting dims; if ndim == 2: it's a mask rather than a bias.
|
154 |
+
# expects mask of shape:
|
155 |
+
# [batch, key_tokens]
|
156 |
+
# adds singleton query_tokens dimension:
|
157 |
+
# [batch, 1, key_tokens]
|
158 |
+
# this helps to broadcast it as a bias over attention scores, which will be in one of the following shapes:
|
159 |
+
# [batch, heads, query_tokens, key_tokens] (e.g. torch sdp attn)
|
160 |
+
# [batch * heads, query_tokens, key_tokens] (e.g. xformers or classic attn)
|
161 |
+
if attention_mask is not None and attention_mask.ndim == 2:
|
162 |
+
# assume that mask is expressed as:
|
163 |
+
# (1 = keep, 0 = discard)
|
164 |
+
# convert mask into a bias that can be added to attention scores:
|
165 |
+
# (keep = +0, discard = -10000.0)
|
166 |
+
attention_mask = (1 - attention_mask.to(hidden_states.dtype)) * -10000.0
|
167 |
+
attention_mask = attention_mask.unsqueeze(1)
|
168 |
+
|
169 |
+
# convert encoder_attention_mask to a bias the same way we do for attention_mask
|
170 |
+
if encoder_attention_mask is not None and encoder_attention_mask.ndim == 2:
|
171 |
+
encoder_attention_mask = (
|
172 |
+
1 - encoder_attention_mask.to(hidden_states.dtype)
|
173 |
+
) * -10000.0
|
174 |
+
encoder_attention_mask = encoder_attention_mask.unsqueeze(1)
|
175 |
+
|
176 |
+
# 1. Input
|
177 |
+
batch, _, seq_len = hidden_states.shape
|
178 |
+
residual = hidden_states
|
179 |
+
|
180 |
+
hidden_states = self.norm(hidden_states)
|
181 |
+
inner_dim = hidden_states.shape[1]
|
182 |
+
hidden_states = hidden_states.permute(0, 2, 1).reshape(
|
183 |
+
batch, seq_len, inner_dim
|
184 |
+
)
|
185 |
+
hidden_states = self.proj_in(hidden_states)
|
186 |
+
|
187 |
+
# 2. Blocks
|
188 |
+
for block in self.transformer_blocks:
|
189 |
+
if self.training and self.gradient_checkpointing:
|
190 |
+
hidden_states = torch.utils.checkpoint.checkpoint(
|
191 |
+
block,
|
192 |
+
hidden_states,
|
193 |
+
attention_mask,
|
194 |
+
encoder_hidden_states,
|
195 |
+
encoder_attention_mask,
|
196 |
+
use_reentrant=False,
|
197 |
+
)
|
198 |
+
else:
|
199 |
+
hidden_states = block(
|
200 |
+
hidden_states,
|
201 |
+
attention_mask=attention_mask,
|
202 |
+
encoder_hidden_states=encoder_hidden_states,
|
203 |
+
encoder_attention_mask=encoder_attention_mask,
|
204 |
+
)
|
205 |
+
|
206 |
+
# 3. Output
|
207 |
+
hidden_states = self.proj_out(hidden_states)
|
208 |
+
hidden_states = (
|
209 |
+
hidden_states.reshape(batch, seq_len, inner_dim)
|
210 |
+
.permute(0, 2, 1)
|
211 |
+
.contiguous()
|
212 |
+
)
|
213 |
+
|
214 |
+
output = hidden_states + residual
|
215 |
+
|
216 |
+
return output
|
tsr/system.py
ADDED
@@ -0,0 +1,203 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
import math
|
2 |
+
import os
|
3 |
+
from dataclasses import dataclass, field
|
4 |
+
from typing import List, Union
|
5 |
+
|
6 |
+
import numpy as np
|
7 |
+
import PIL.Image
|
8 |
+
import torch
|
9 |
+
import torch.nn.functional as F
|
10 |
+
import trimesh
|
11 |
+
from einops import rearrange
|
12 |
+
from huggingface_hub import hf_hub_download
|
13 |
+
from omegaconf import OmegaConf
|
14 |
+
from PIL import Image
|
15 |
+
|
16 |
+
from .models.isosurface import MarchingCubeHelper
|
17 |
+
from .utils import (
|
18 |
+
BaseModule,
|
19 |
+
ImagePreprocessor,
|
20 |
+
find_class,
|
21 |
+
get_spherical_cameras,
|
22 |
+
scale_tensor,
|
23 |
+
)
|
24 |
+
|
25 |
+
|
26 |
+
class TSR(BaseModule):
|
27 |
+
@dataclass
|
28 |
+
class Config(BaseModule.Config):
|
29 |
+
cond_image_size: int
|
30 |
+
|
31 |
+
image_tokenizer_cls: str
|
32 |
+
image_tokenizer: dict
|
33 |
+
|
34 |
+
tokenizer_cls: str
|
35 |
+
tokenizer: dict
|
36 |
+
|
37 |
+
backbone_cls: str
|
38 |
+
backbone: dict
|
39 |
+
|
40 |
+
post_processor_cls: str
|
41 |
+
post_processor: dict
|
42 |
+
|
43 |
+
decoder_cls: str
|
44 |
+
decoder: dict
|
45 |
+
|
46 |
+
renderer_cls: str
|
47 |
+
renderer: dict
|
48 |
+
|
49 |
+
cfg: Config
|
50 |
+
|
51 |
+
@classmethod
|
52 |
+
def from_pretrained(
|
53 |
+
cls, pretrained_model_name_or_path: str, config_name: str, weight_name: str, token=None
|
54 |
+
):
|
55 |
+
if os.path.isdir(pretrained_model_name_or_path):
|
56 |
+
config_path = os.path.join(pretrained_model_name_or_path, config_name)
|
57 |
+
weight_path = os.path.join(pretrained_model_name_or_path, weight_name)
|
58 |
+
else:
|
59 |
+
config_path = hf_hub_download(
|
60 |
+
repo_id=pretrained_model_name_or_path, filename=config_name, token=token
|
61 |
+
)
|
62 |
+
weight_path = hf_hub_download(
|
63 |
+
repo_id=pretrained_model_name_or_path, filename=weight_name, token=token
|
64 |
+
)
|
65 |
+
|
66 |
+
cfg = OmegaConf.load(config_path)
|
67 |
+
OmegaConf.resolve(cfg)
|
68 |
+
model = cls(cfg)
|
69 |
+
ckpt = torch.load(weight_path, map_location="cpu")
|
70 |
+
model.load_state_dict(ckpt)
|
71 |
+
return model
|
72 |
+
|
73 |
+
def configure(self):
|
74 |
+
self.image_tokenizer = find_class(self.cfg.image_tokenizer_cls)(
|
75 |
+
self.cfg.image_tokenizer
|
76 |
+
)
|
77 |
+
self.tokenizer = find_class(self.cfg.tokenizer_cls)(self.cfg.tokenizer)
|
78 |
+
self.backbone = find_class(self.cfg.backbone_cls)(self.cfg.backbone)
|
79 |
+
self.post_processor = find_class(self.cfg.post_processor_cls)(
|
80 |
+
self.cfg.post_processor
|
81 |
+
)
|
82 |
+
self.decoder = find_class(self.cfg.decoder_cls)(self.cfg.decoder)
|
83 |
+
self.renderer = find_class(self.cfg.renderer_cls)(self.cfg.renderer)
|
84 |
+
self.image_processor = ImagePreprocessor()
|
85 |
+
self.isosurface_helper = None
|
86 |
+
|
87 |
+
def forward(
|
88 |
+
self,
|
89 |
+
image: Union[
|
90 |
+
PIL.Image.Image,
|
91 |
+
np.ndarray,
|
92 |
+
torch.FloatTensor,
|
93 |
+
List[PIL.Image.Image],
|
94 |
+
List[np.ndarray],
|
95 |
+
List[torch.FloatTensor],
|
96 |
+
],
|
97 |
+
device: str,
|
98 |
+
) -> torch.FloatTensor:
|
99 |
+
rgb_cond = self.image_processor(image, self.cfg.cond_image_size)[:, None].to(
|
100 |
+
device
|
101 |
+
)
|
102 |
+
batch_size = rgb_cond.shape[0]
|
103 |
+
|
104 |
+
input_image_tokens: torch.Tensor = self.image_tokenizer(
|
105 |
+
rearrange(rgb_cond, "B Nv H W C -> B Nv C H W", Nv=1),
|
106 |
+
)
|
107 |
+
|
108 |
+
input_image_tokens = rearrange(
|
109 |
+
input_image_tokens, "B Nv C Nt -> B (Nv Nt) C", Nv=1
|
110 |
+
)
|
111 |
+
|
112 |
+
tokens: torch.Tensor = self.tokenizer(batch_size)
|
113 |
+
|
114 |
+
tokens = self.backbone(
|
115 |
+
tokens,
|
116 |
+
encoder_hidden_states=input_image_tokens,
|
117 |
+
)
|
118 |
+
|
119 |
+
scene_codes = self.post_processor(self.tokenizer.detokenize(tokens))
|
120 |
+
return scene_codes
|
121 |
+
|
122 |
+
def render(
|
123 |
+
self,
|
124 |
+
scene_codes,
|
125 |
+
n_views: int,
|
126 |
+
elevation_deg: float = 0.0,
|
127 |
+
camera_distance: float = 1.9,
|
128 |
+
fovy_deg: float = 40.0,
|
129 |
+
height: int = 256,
|
130 |
+
width: int = 256,
|
131 |
+
return_type: str = "pil",
|
132 |
+
):
|
133 |
+
rays_o, rays_d = get_spherical_cameras(
|
134 |
+
n_views, elevation_deg, camera_distance, fovy_deg, height, width
|
135 |
+
)
|
136 |
+
rays_o, rays_d = rays_o.to(scene_codes.device), rays_d.to(scene_codes.device)
|
137 |
+
|
138 |
+
def process_output(image: torch.FloatTensor):
|
139 |
+
if return_type == "pt":
|
140 |
+
return image
|
141 |
+
elif return_type == "np":
|
142 |
+
return image.detach().cpu().numpy()
|
143 |
+
elif return_type == "pil":
|
144 |
+
return Image.fromarray(
|
145 |
+
(image.detach().cpu().numpy() * 255.0).astype(np.uint8)
|
146 |
+
)
|
147 |
+
else:
|
148 |
+
raise NotImplementedError
|
149 |
+
|
150 |
+
images = []
|
151 |
+
for scene_code in scene_codes:
|
152 |
+
images_ = []
|
153 |
+
for i in range(n_views):
|
154 |
+
with torch.no_grad():
|
155 |
+
image = self.renderer(
|
156 |
+
self.decoder, scene_code, rays_o[i], rays_d[i]
|
157 |
+
)
|
158 |
+
images_.append(process_output(image))
|
159 |
+
images.append(images_)
|
160 |
+
|
161 |
+
return images
|
162 |
+
|
163 |
+
def set_marching_cubes_resolution(self, resolution: int):
|
164 |
+
if (
|
165 |
+
self.isosurface_helper is not None
|
166 |
+
and self.isosurface_helper.resolution == resolution
|
167 |
+
):
|
168 |
+
return
|
169 |
+
self.isosurface_helper = MarchingCubeHelper(resolution)
|
170 |
+
|
171 |
+
def extract_mesh(self, scene_codes, resolution: int = 256, threshold: float = 25.0):
|
172 |
+
self.set_marching_cubes_resolution(resolution)
|
173 |
+
meshes = []
|
174 |
+
for scene_code in scene_codes:
|
175 |
+
with torch.no_grad():
|
176 |
+
density = self.renderer.query_triplane(
|
177 |
+
self.decoder,
|
178 |
+
scale_tensor(
|
179 |
+
self.isosurface_helper.grid_vertices.to(scene_codes.device),
|
180 |
+
self.isosurface_helper.points_range,
|
181 |
+
(-self.renderer.cfg.radius, self.renderer.cfg.radius),
|
182 |
+
),
|
183 |
+
scene_code,
|
184 |
+
)["density_act"]
|
185 |
+
v_pos, t_pos_idx = self.isosurface_helper(-(density - threshold))
|
186 |
+
v_pos = scale_tensor(
|
187 |
+
v_pos,
|
188 |
+
self.isosurface_helper.points_range,
|
189 |
+
(-self.renderer.cfg.radius, self.renderer.cfg.radius),
|
190 |
+
)
|
191 |
+
with torch.no_grad():
|
192 |
+
color = self.renderer.query_triplane(
|
193 |
+
self.decoder,
|
194 |
+
v_pos,
|
195 |
+
scene_code,
|
196 |
+
)["color"]
|
197 |
+
mesh = trimesh.Trimesh(
|
198 |
+
vertices=v_pos.cpu().numpy(),
|
199 |
+
faces=t_pos_idx.cpu().numpy(),
|
200 |
+
vertex_colors=color.cpu().numpy(),
|
201 |
+
)
|
202 |
+
meshes.append(mesh)
|
203 |
+
return meshes
|
tsr/utils.py
ADDED
@@ -0,0 +1,482 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
import importlib
|
2 |
+
import math
|
3 |
+
from collections import defaultdict
|
4 |
+
from dataclasses import dataclass
|
5 |
+
from typing import Any, Callable, Dict, List, Optional, Tuple, Union
|
6 |
+
|
7 |
+
import imageio
|
8 |
+
import numpy as np
|
9 |
+
import PIL.Image
|
10 |
+
import rembg
|
11 |
+
import torch
|
12 |
+
import torch.nn as nn
|
13 |
+
import torch.nn.functional as F
|
14 |
+
import trimesh
|
15 |
+
from omegaconf import DictConfig, OmegaConf
|
16 |
+
from PIL import Image
|
17 |
+
|
18 |
+
|
19 |
+
def parse_structured(fields: Any, cfg: Optional[Union[dict, DictConfig]] = None) -> Any:
|
20 |
+
scfg = OmegaConf.merge(OmegaConf.structured(fields), cfg)
|
21 |
+
return scfg
|
22 |
+
|
23 |
+
|
24 |
+
def find_class(cls_string):
|
25 |
+
module_string = ".".join(cls_string.split(".")[:-1])
|
26 |
+
cls_name = cls_string.split(".")[-1]
|
27 |
+
module = importlib.import_module(module_string, package=None)
|
28 |
+
cls = getattr(module, cls_name)
|
29 |
+
return cls
|
30 |
+
|
31 |
+
|
32 |
+
def get_intrinsic_from_fov(fov, H, W, bs=-1):
|
33 |
+
focal_length = 0.5 * H / np.tan(0.5 * fov)
|
34 |
+
intrinsic = np.identity(3, dtype=np.float32)
|
35 |
+
intrinsic[0, 0] = focal_length
|
36 |
+
intrinsic[1, 1] = focal_length
|
37 |
+
intrinsic[0, 2] = W / 2.0
|
38 |
+
intrinsic[1, 2] = H / 2.0
|
39 |
+
|
40 |
+
if bs > 0:
|
41 |
+
intrinsic = intrinsic[None].repeat(bs, axis=0)
|
42 |
+
|
43 |
+
return torch.from_numpy(intrinsic)
|
44 |
+
|
45 |
+
|
46 |
+
class BaseModule(nn.Module):
|
47 |
+
@dataclass
|
48 |
+
class Config:
|
49 |
+
pass
|
50 |
+
|
51 |
+
cfg: Config # add this to every subclass of BaseModule to enable static type checking
|
52 |
+
|
53 |
+
def __init__(
|
54 |
+
self, cfg: Optional[Union[dict, DictConfig]] = None, *args, **kwargs
|
55 |
+
) -> None:
|
56 |
+
super().__init__()
|
57 |
+
self.cfg = parse_structured(self.Config, cfg)
|
58 |
+
self.configure(*args, **kwargs)
|
59 |
+
|
60 |
+
def configure(self, *args, **kwargs) -> None:
|
61 |
+
raise NotImplementedError
|
62 |
+
|
63 |
+
|
64 |
+
class ImagePreprocessor:
|
65 |
+
def convert_and_resize(
|
66 |
+
self,
|
67 |
+
image: Union[PIL.Image.Image, np.ndarray, torch.Tensor],
|
68 |
+
size: int,
|
69 |
+
):
|
70 |
+
if isinstance(image, PIL.Image.Image):
|
71 |
+
image = torch.from_numpy(np.array(image).astype(np.float32) / 255.0)
|
72 |
+
elif isinstance(image, np.ndarray):
|
73 |
+
if image.dtype == np.uint8:
|
74 |
+
image = torch.from_numpy(image.astype(np.float32) / 255.0)
|
75 |
+
else:
|
76 |
+
image = torch.from_numpy(image)
|
77 |
+
elif isinstance(image, torch.Tensor):
|
78 |
+
pass
|
79 |
+
|
80 |
+
batched = image.ndim == 4
|
81 |
+
|
82 |
+
if not batched:
|
83 |
+
image = image[None, ...]
|
84 |
+
image = F.interpolate(
|
85 |
+
image.permute(0, 3, 1, 2),
|
86 |
+
(size, size),
|
87 |
+
mode="bilinear",
|
88 |
+
align_corners=False,
|
89 |
+
antialias=True,
|
90 |
+
).permute(0, 2, 3, 1)
|
91 |
+
if not batched:
|
92 |
+
image = image[0]
|
93 |
+
return image
|
94 |
+
|
95 |
+
def __call__(
|
96 |
+
self,
|
97 |
+
image: Union[
|
98 |
+
PIL.Image.Image,
|
99 |
+
np.ndarray,
|
100 |
+
torch.FloatTensor,
|
101 |
+
List[PIL.Image.Image],
|
102 |
+
List[np.ndarray],
|
103 |
+
List[torch.FloatTensor],
|
104 |
+
],
|
105 |
+
size: int,
|
106 |
+
) -> Any:
|
107 |
+
if isinstance(image, (np.ndarray, torch.FloatTensor)) and image.ndim == 4:
|
108 |
+
image = self.convert_and_resize(image, size)
|
109 |
+
else:
|
110 |
+
if not isinstance(image, list):
|
111 |
+
image = [image]
|
112 |
+
image = [self.convert_and_resize(im, size) for im in image]
|
113 |
+
image = torch.stack(image, dim=0)
|
114 |
+
return image
|
115 |
+
|
116 |
+
|
117 |
+
def rays_intersect_bbox(
|
118 |
+
rays_o: torch.Tensor,
|
119 |
+
rays_d: torch.Tensor,
|
120 |
+
radius: float,
|
121 |
+
near: float = 0.0,
|
122 |
+
valid_thresh: float = 0.01,
|
123 |
+
):
|
124 |
+
input_shape = rays_o.shape[:-1]
|
125 |
+
rays_o, rays_d = rays_o.view(-1, 3), rays_d.view(-1, 3)
|
126 |
+
rays_d_valid = torch.where(
|
127 |
+
rays_d.abs() < 1e-6, torch.full_like(rays_d, 1e-6), rays_d
|
128 |
+
)
|
129 |
+
if type(radius) in [int, float]:
|
130 |
+
radius = torch.FloatTensor(
|
131 |
+
[[-radius, radius], [-radius, radius], [-radius, radius]]
|
132 |
+
).to(rays_o.device)
|
133 |
+
radius = (
|
134 |
+
1.0 - 1.0e-3
|
135 |
+
) * radius # tighten the radius to make sure the intersection point lies in the bounding box
|
136 |
+
interx0 = (radius[..., 1] - rays_o) / rays_d_valid
|
137 |
+
interx1 = (radius[..., 0] - rays_o) / rays_d_valid
|
138 |
+
t_near = torch.minimum(interx0, interx1).amax(dim=-1).clamp_min(near)
|
139 |
+
t_far = torch.maximum(interx0, interx1).amin(dim=-1)
|
140 |
+
|
141 |
+
# check wheter a ray intersects the bbox or not
|
142 |
+
rays_valid = t_far - t_near > valid_thresh
|
143 |
+
|
144 |
+
t_near[torch.where(~rays_valid)] = 0.0
|
145 |
+
t_far[torch.where(~rays_valid)] = 0.0
|
146 |
+
|
147 |
+
t_near = t_near.view(*input_shape, 1)
|
148 |
+
t_far = t_far.view(*input_shape, 1)
|
149 |
+
rays_valid = rays_valid.view(*input_shape)
|
150 |
+
|
151 |
+
return t_near, t_far, rays_valid
|
152 |
+
|
153 |
+
|
154 |
+
def chunk_batch(func: Callable, chunk_size: int, *args, **kwargs) -> Any:
|
155 |
+
if chunk_size <= 0:
|
156 |
+
return func(*args, **kwargs)
|
157 |
+
B = None
|
158 |
+
for arg in list(args) + list(kwargs.values()):
|
159 |
+
if isinstance(arg, torch.Tensor):
|
160 |
+
B = arg.shape[0]
|
161 |
+
break
|
162 |
+
assert (
|
163 |
+
B is not None
|
164 |
+
), "No tensor found in args or kwargs, cannot determine batch size."
|
165 |
+
out = defaultdict(list)
|
166 |
+
out_type = None
|
167 |
+
# max(1, B) to support B == 0
|
168 |
+
for i in range(0, max(1, B), chunk_size):
|
169 |
+
out_chunk = func(
|
170 |
+
*[
|
171 |
+
arg[i : i + chunk_size] if isinstance(arg, torch.Tensor) else arg
|
172 |
+
for arg in args
|
173 |
+
],
|
174 |
+
**{
|
175 |
+
k: arg[i : i + chunk_size] if isinstance(arg, torch.Tensor) else arg
|
176 |
+
for k, arg in kwargs.items()
|
177 |
+
},
|
178 |
+
)
|
179 |
+
if out_chunk is None:
|
180 |
+
continue
|
181 |
+
out_type = type(out_chunk)
|
182 |
+
if isinstance(out_chunk, torch.Tensor):
|
183 |
+
out_chunk = {0: out_chunk}
|
184 |
+
elif isinstance(out_chunk, tuple) or isinstance(out_chunk, list):
|
185 |
+
chunk_length = len(out_chunk)
|
186 |
+
out_chunk = {i: chunk for i, chunk in enumerate(out_chunk)}
|
187 |
+
elif isinstance(out_chunk, dict):
|
188 |
+
pass
|
189 |
+
else:
|
190 |
+
print(
|
191 |
+
f"Return value of func must be in type [torch.Tensor, list, tuple, dict], get {type(out_chunk)}."
|
192 |
+
)
|
193 |
+
exit(1)
|
194 |
+
for k, v in out_chunk.items():
|
195 |
+
v = v if torch.is_grad_enabled() else v.detach()
|
196 |
+
out[k].append(v)
|
197 |
+
|
198 |
+
if out_type is None:
|
199 |
+
return None
|
200 |
+
|
201 |
+
out_merged: Dict[Any, Optional[torch.Tensor]] = {}
|
202 |
+
for k, v in out.items():
|
203 |
+
if all([vv is None for vv in v]):
|
204 |
+
# allow None in return value
|
205 |
+
out_merged[k] = None
|
206 |
+
elif all([isinstance(vv, torch.Tensor) for vv in v]):
|
207 |
+
out_merged[k] = torch.cat(v, dim=0)
|
208 |
+
else:
|
209 |
+
raise TypeError(
|
210 |
+
f"Unsupported types in return value of func: {[type(vv) for vv in v if not isinstance(vv, torch.Tensor)]}"
|
211 |
+
)
|
212 |
+
|
213 |
+
if out_type is torch.Tensor:
|
214 |
+
return out_merged[0]
|
215 |
+
elif out_type in [tuple, list]:
|
216 |
+
return out_type([out_merged[i] for i in range(chunk_length)])
|
217 |
+
elif out_type is dict:
|
218 |
+
return out_merged
|
219 |
+
|
220 |
+
|
221 |
+
ValidScale = Union[Tuple[float, float], torch.FloatTensor]
|
222 |
+
|
223 |
+
|
224 |
+
def scale_tensor(dat: torch.FloatTensor, inp_scale: ValidScale, tgt_scale: ValidScale):
|
225 |
+
if inp_scale is None:
|
226 |
+
inp_scale = (0, 1)
|
227 |
+
if tgt_scale is None:
|
228 |
+
tgt_scale = (0, 1)
|
229 |
+
if isinstance(tgt_scale, torch.FloatTensor):
|
230 |
+
assert dat.shape[-1] == tgt_scale.shape[-1]
|
231 |
+
dat = (dat - inp_scale[0]) / (inp_scale[1] - inp_scale[0])
|
232 |
+
dat = dat * (tgt_scale[1] - tgt_scale[0]) + tgt_scale[0]
|
233 |
+
return dat
|
234 |
+
|
235 |
+
|
236 |
+
def get_activation(name) -> Callable:
|
237 |
+
if name is None:
|
238 |
+
return lambda x: x
|
239 |
+
name = name.lower()
|
240 |
+
if name == "none":
|
241 |
+
return lambda x: x
|
242 |
+
elif name == "exp":
|
243 |
+
return lambda x: torch.exp(x)
|
244 |
+
elif name == "sigmoid":
|
245 |
+
return lambda x: torch.sigmoid(x)
|
246 |
+
elif name == "tanh":
|
247 |
+
return lambda x: torch.tanh(x)
|
248 |
+
elif name == "softplus":
|
249 |
+
return lambda x: F.softplus(x)
|
250 |
+
else:
|
251 |
+
try:
|
252 |
+
return getattr(F, name)
|
253 |
+
except AttributeError:
|
254 |
+
raise ValueError(f"Unknown activation function: {name}")
|
255 |
+
|
256 |
+
|
257 |
+
def get_ray_directions(
|
258 |
+
H: int,
|
259 |
+
W: int,
|
260 |
+
focal: Union[float, Tuple[float, float]],
|
261 |
+
principal: Optional[Tuple[float, float]] = None,
|
262 |
+
use_pixel_centers: bool = True,
|
263 |
+
normalize: bool = True,
|
264 |
+
) -> torch.FloatTensor:
|
265 |
+
"""
|
266 |
+
Get ray directions for all pixels in camera coordinate.
|
267 |
+
Reference: https://www.scratchapixel.com/lessons/3d-basic-rendering/
|
268 |
+
ray-tracing-generating-camera-rays/standard-coordinate-systems
|
269 |
+
|
270 |
+
Inputs:
|
271 |
+
H, W, focal, principal, use_pixel_centers: image height, width, focal length, principal point and whether use pixel centers
|
272 |
+
Outputs:
|
273 |
+
directions: (H, W, 3), the direction of the rays in camera coordinate
|
274 |
+
"""
|
275 |
+
pixel_center = 0.5 if use_pixel_centers else 0
|
276 |
+
|
277 |
+
if isinstance(focal, float):
|
278 |
+
fx, fy = focal, focal
|
279 |
+
cx, cy = W / 2, H / 2
|
280 |
+
else:
|
281 |
+
fx, fy = focal
|
282 |
+
assert principal is not None
|
283 |
+
cx, cy = principal
|
284 |
+
|
285 |
+
i, j = torch.meshgrid(
|
286 |
+
torch.arange(W, dtype=torch.float32) + pixel_center,
|
287 |
+
torch.arange(H, dtype=torch.float32) + pixel_center,
|
288 |
+
indexing="xy",
|
289 |
+
)
|
290 |
+
|
291 |
+
directions = torch.stack([(i - cx) / fx, -(j - cy) / fy, -torch.ones_like(i)], -1)
|
292 |
+
|
293 |
+
if normalize:
|
294 |
+
directions = F.normalize(directions, dim=-1)
|
295 |
+
|
296 |
+
return directions
|
297 |
+
|
298 |
+
|
299 |
+
def get_rays(
|
300 |
+
directions,
|
301 |
+
c2w,
|
302 |
+
keepdim=False,
|
303 |
+
noise_scale=0.0,
|
304 |
+
normalize=False,
|
305 |
+
) -> Tuple[torch.FloatTensor, torch.FloatTensor]:
|
306 |
+
# Rotate ray directions from camera coordinate to the world coordinate
|
307 |
+
assert directions.shape[-1] == 3
|
308 |
+
|
309 |
+
if directions.ndim == 2: # (N_rays, 3)
|
310 |
+
if c2w.ndim == 2: # (4, 4)
|
311 |
+
c2w = c2w[None, :, :]
|
312 |
+
assert c2w.ndim == 3 # (N_rays, 4, 4) or (1, 4, 4)
|
313 |
+
rays_d = (directions[:, None, :] * c2w[:, :3, :3]).sum(-1) # (N_rays, 3)
|
314 |
+
rays_o = c2w[:, :3, 3].expand(rays_d.shape)
|
315 |
+
elif directions.ndim == 3: # (H, W, 3)
|
316 |
+
assert c2w.ndim in [2, 3]
|
317 |
+
if c2w.ndim == 2: # (4, 4)
|
318 |
+
rays_d = (directions[:, :, None, :] * c2w[None, None, :3, :3]).sum(
|
319 |
+
-1
|
320 |
+
) # (H, W, 3)
|
321 |
+
rays_o = c2w[None, None, :3, 3].expand(rays_d.shape)
|
322 |
+
elif c2w.ndim == 3: # (B, 4, 4)
|
323 |
+
rays_d = (directions[None, :, :, None, :] * c2w[:, None, None, :3, :3]).sum(
|
324 |
+
-1
|
325 |
+
) # (B, H, W, 3)
|
326 |
+
rays_o = c2w[:, None, None, :3, 3].expand(rays_d.shape)
|
327 |
+
elif directions.ndim == 4: # (B, H, W, 3)
|
328 |
+
assert c2w.ndim == 3 # (B, 4, 4)
|
329 |
+
rays_d = (directions[:, :, :, None, :] * c2w[:, None, None, :3, :3]).sum(
|
330 |
+
-1
|
331 |
+
) # (B, H, W, 3)
|
332 |
+
rays_o = c2w[:, None, None, :3, 3].expand(rays_d.shape)
|
333 |
+
|
334 |
+
# add camera noise to avoid grid-like artifect
|
335 |
+
# https://github.com/ashawkey/stable-dreamfusion/blob/49c3d4fa01d68a4f027755acf94e1ff6020458cc/nerf/utils.py#L373
|
336 |
+
if noise_scale > 0:
|
337 |
+
rays_o = rays_o + torch.randn(3, device=rays_o.device) * noise_scale
|
338 |
+
rays_d = rays_d + torch.randn(3, device=rays_d.device) * noise_scale
|
339 |
+
|
340 |
+
if normalize:
|
341 |
+
rays_d = F.normalize(rays_d, dim=-1)
|
342 |
+
if not keepdim:
|
343 |
+
rays_o, rays_d = rays_o.reshape(-1, 3), rays_d.reshape(-1, 3)
|
344 |
+
|
345 |
+
return rays_o, rays_d
|
346 |
+
|
347 |
+
|
348 |
+
def get_spherical_cameras(
|
349 |
+
n_views: int,
|
350 |
+
elevation_deg: float,
|
351 |
+
camera_distance: float,
|
352 |
+
fovy_deg: float,
|
353 |
+
height: int,
|
354 |
+
width: int,
|
355 |
+
):
|
356 |
+
azimuth_deg = torch.linspace(0, 360.0, n_views + 1)[:n_views]
|
357 |
+
elevation_deg = torch.full_like(azimuth_deg, elevation_deg)
|
358 |
+
camera_distances = torch.full_like(elevation_deg, camera_distance)
|
359 |
+
|
360 |
+
elevation = elevation_deg * math.pi / 180
|
361 |
+
azimuth = azimuth_deg * math.pi / 180
|
362 |
+
|
363 |
+
# convert spherical coordinates to cartesian coordinates
|
364 |
+
# right hand coordinate system, x back, y right, z up
|
365 |
+
# elevation in (-90, 90), azimuth from +x to +y in (-180, 180)
|
366 |
+
camera_positions = torch.stack(
|
367 |
+
[
|
368 |
+
camera_distances * torch.cos(elevation) * torch.cos(azimuth),
|
369 |
+
camera_distances * torch.cos(elevation) * torch.sin(azimuth),
|
370 |
+
camera_distances * torch.sin(elevation),
|
371 |
+
],
|
372 |
+
dim=-1,
|
373 |
+
)
|
374 |
+
|
375 |
+
# default scene center at origin
|
376 |
+
center = torch.zeros_like(camera_positions)
|
377 |
+
# default camera up direction as +z
|
378 |
+
up = torch.as_tensor([0, 0, 1], dtype=torch.float32)[None, :].repeat(n_views, 1)
|
379 |
+
|
380 |
+
fovy = torch.full_like(elevation_deg, fovy_deg) * math.pi / 180
|
381 |
+
|
382 |
+
lookat = F.normalize(center - camera_positions, dim=-1)
|
383 |
+
right = F.normalize(torch.cross(lookat, up), dim=-1)
|
384 |
+
up = F.normalize(torch.cross(right, lookat), dim=-1)
|
385 |
+
c2w3x4 = torch.cat(
|
386 |
+
[torch.stack([right, up, -lookat], dim=-1), camera_positions[:, :, None]],
|
387 |
+
dim=-1,
|
388 |
+
)
|
389 |
+
c2w = torch.cat([c2w3x4, torch.zeros_like(c2w3x4[:, :1])], dim=1)
|
390 |
+
c2w[:, 3, 3] = 1.0
|
391 |
+
|
392 |
+
# get directions by dividing directions_unit_focal by focal length
|
393 |
+
focal_length = 0.5 * height / torch.tan(0.5 * fovy)
|
394 |
+
directions_unit_focal = get_ray_directions(
|
395 |
+
H=height,
|
396 |
+
W=width,
|
397 |
+
focal=1.0,
|
398 |
+
)
|
399 |
+
directions = directions_unit_focal[None, :, :, :].repeat(n_views, 1, 1, 1)
|
400 |
+
directions[:, :, :, :2] = (
|
401 |
+
directions[:, :, :, :2] / focal_length[:, None, None, None]
|
402 |
+
)
|
403 |
+
# must use normalize=True to normalize directions here
|
404 |
+
rays_o, rays_d = get_rays(directions, c2w, keepdim=True, normalize=True)
|
405 |
+
|
406 |
+
return rays_o, rays_d
|
407 |
+
|
408 |
+
|
409 |
+
def remove_background(
|
410 |
+
image: PIL.Image.Image,
|
411 |
+
rembg_session: Any = None,
|
412 |
+
force: bool = False,
|
413 |
+
**rembg_kwargs,
|
414 |
+
) -> PIL.Image.Image:
|
415 |
+
do_remove = True
|
416 |
+
if image.mode == "RGBA" and image.getextrema()[3][0] < 255:
|
417 |
+
do_remove = False
|
418 |
+
do_remove = do_remove or force
|
419 |
+
if do_remove:
|
420 |
+
image = rembg.remove(image, session=rembg_session, **rembg_kwargs)
|
421 |
+
return image
|
422 |
+
|
423 |
+
|
424 |
+
def resize_foreground(
|
425 |
+
image: PIL.Image.Image,
|
426 |
+
ratio: float,
|
427 |
+
) -> PIL.Image.Image:
|
428 |
+
image = np.array(image)
|
429 |
+
assert image.shape[-1] == 4
|
430 |
+
alpha = np.where(image[..., 3] > 0)
|
431 |
+
y1, y2, x1, x2 = (
|
432 |
+
alpha[0].min(),
|
433 |
+
alpha[0].max(),
|
434 |
+
alpha[1].min(),
|
435 |
+
alpha[1].max(),
|
436 |
+
)
|
437 |
+
# crop the foreground
|
438 |
+
fg = image[y1:y2, x1:x2]
|
439 |
+
# pad to square
|
440 |
+
size = max(fg.shape[0], fg.shape[1])
|
441 |
+
ph0, pw0 = (size - fg.shape[0]) // 2, (size - fg.shape[1]) // 2
|
442 |
+
ph1, pw1 = size - fg.shape[0] - ph0, size - fg.shape[1] - pw0
|
443 |
+
new_image = np.pad(
|
444 |
+
fg,
|
445 |
+
((ph0, ph1), (pw0, pw1), (0, 0)),
|
446 |
+
mode="constant",
|
447 |
+
constant_values=((0, 0), (0, 0), (0, 0)),
|
448 |
+
)
|
449 |
+
|
450 |
+
# compute padding according to the ratio
|
451 |
+
new_size = int(new_image.shape[0] / ratio)
|
452 |
+
# pad to size, double side
|
453 |
+
ph0, pw0 = (new_size - size) // 2, (new_size - size) // 2
|
454 |
+
ph1, pw1 = new_size - size - ph0, new_size - size - pw0
|
455 |
+
new_image = np.pad(
|
456 |
+
new_image,
|
457 |
+
((ph0, ph1), (pw0, pw1), (0, 0)),
|
458 |
+
mode="constant",
|
459 |
+
constant_values=((0, 0), (0, 0), (0, 0)),
|
460 |
+
)
|
461 |
+
new_image = PIL.Image.fromarray(new_image)
|
462 |
+
return new_image
|
463 |
+
|
464 |
+
|
465 |
+
def save_video(
|
466 |
+
frames: List[PIL.Image.Image],
|
467 |
+
output_path: str,
|
468 |
+
fps: int = 30,
|
469 |
+
):
|
470 |
+
# use imageio to save video
|
471 |
+
frames = [np.array(frame) for frame in frames]
|
472 |
+
writer = imageio.get_writer(output_path, fps=fps)
|
473 |
+
for frame in frames:
|
474 |
+
writer.append_data(frame)
|
475 |
+
writer.close()
|
476 |
+
|
477 |
+
|
478 |
+
def to_gradio_3d_orientation(mesh):
|
479 |
+
mesh.apply_transform(trimesh.transformations.rotation_matrix(-np.pi/2, [1, 0, 0]))
|
480 |
+
# mesh.apply_scale([1, 1, -1])
|
481 |
+
mesh.apply_transform(trimesh.transformations.rotation_matrix(np.pi/2, [0, 1, 0]))
|
482 |
+
return mesh
|