-
Notifications
You must be signed in to change notification settings - Fork 4
/
conv_wasm.ts
233 lines (208 loc) · 7.91 KB
/
conv_wasm.ts
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
/**
* This example shows how to use convolutional layers to extract features from an image.
*/
import {
AveragePool2DLayer,
Conv2DLayer,
Cost,
MaxPool2DLayer,
type Rank,
Sequential,
setupBackend,
Tensor,
tensor4D,
WASM,
} from "../../packages/core/mod.ts";
import { decode } from "https://deno.land/x/pngs@0.1.1/mod.ts";
import { createCanvas } from "https://deno.land/x/canvas@v1.4.1/mod.ts";
import type { Layer } from "../../packages/core/src/core/api/layer.ts";
const canvas = createCanvas(600, 600);
const ctx = canvas.getContext("2d");
ctx.fillStyle = "white";
ctx.fillRect(0, 0, 600, 600);
/**
* Size of the input tensor for the neural network.
*/
const dim = 28;
/**
* 3D array used as a filter for convolutional layers in a neural network.
* The filter is a 3x3 matrix with values that can be used to detect vertical edges in an image.
* The values in the filter are set such that the center column has negative values and the left and right columns have positive values.
* This means that when the filter is convolved with an image, it will highlight vertical edges in the image.
*/
const kernel = [
[
[
[0, -1, 1],
[0, -1, 1],
[0, -1, 1],
],
],
];
//Credit: Hashrock (https://github.com/hashrock)
const img = decode(Deno.readFileSync("./examples/filters/deno.png")).image;
const buffer = new Float32Array(dim * dim);
for (let i = 0; i < dim * dim; i++) {
buffer[i] = img[i * 4];
}
/**
* Setup the WASM backend. This backend is slower than the CPU backend but works on the Edge.
*/
await setupBackend(WASM);
/**
* Draw the original image on the canvas.
*/
drawPixels(buffer, dim);
const conv = await feedForward([
/**
* Creates a convolutional layer in a neural network.
*/
Conv2DLayer({
/**
* The kernel is set to the filter defined previously.
*/
kernel: tensor4D(kernel),
/**
* Sets the kernel size for a convolutional layer in a neural network.
* The kernel size is a 4D array that specifies the dimensions of the kernel that will be used to convolve the input tensor.
* In this case, the kernel size is set to [1, 1, 3, 3],
* which means that the kernel will have a depth of 1, a height of 3, and a width of 3.
* This kernel size is used in the creation of two different convolutional layers in the code below the selection.
*/
kernelSize: [1, 1, 3, 3],
/**
* Sets the padding for a convolutional layer in a neural network.
* Padding is used to add extra pixels around the edges of an image
* to ensure that the output of the convolution operation has the same dimensions as the input.
* In this case, the padding is set to [1, 1],
* which means that one pixel of padding will be added to the top, bottom, left, and right edges of the input tensor.
*/
padding: [1, 1],
}),
]);
drawPixels(conv.data, conv.shape[2], 280);
/**
* Creates a network with a convolutional layer and a max pooling layer.
*/
const pool = await feedForward([
Conv2DLayer({
/**
* The kernel is set to the filter defined previously.
*/
kernel: tensor4D(kernel),
/**
* Sets the kernel size for a convolutional layer in a neural network.
* The kernel size is a 4D array that specifies the dimensions of the kernel that will be used to convolve the input tensor.
* In this case, the kernel size is set to [1, 1, 3, 3],
* which means that the kernel will have a depth of 1, a height of 3, and a width of 3.
* This kernel size is used in the creation of two different convolutional layers in the code below the selection.
*/
kernelSize: [1, 1, 3, 3],
/**
* Sets the padding for a convolutional layer in a neural network.
* Padding is used to add extra pixels around the edges of an image
* to ensure that the output of the convolution operation has the same dimensions as the input.
* In this case, the padding is set to [1, 1],
* which means that one pixel of padding will be added to the top, bottom, left, and right edges of the input tensor.
*/
padding: [1, 1],
}),
/**
* Creates a Max Pooling layer in a neural network.
* Max Pooling is a technique used in Convolutional Neural Networks (CNNs) to reduce the spatial dimensions of the output volume.
* It is used to downsample the input along the spatial dimensions (width and height) while keeping the depth constant.
* The Max Pooling layer takes a tensor as input and applies a max filter to non-overlapping subregions of the tensor.
* The output of the Max Pooling layer is a tensor with reduced spatial dimensions and the same depth as the input tensor.
*/
MaxPool2DLayer({
/**
* The strides specifies the stride of the pooling operation along the spatial dimensions.
* In this case, the stride is set to [2, 2],
* which means that the pooling operation will be applied to non-overlapping 2x2 regions of the input tensor.
*/
strides: [2, 2],
}),
]);
/**
* Draw the output of the Max Pooling layer example on the canvas.
*/
drawPixels(pool.data, pool.shape[2], 0, 280, 2);
/**
* Creates a network with a convolutional layer and an average pooling layer.
*/
const pool2 = await feedForward([
Conv2DLayer({
/**
* The kernel is set to the filter defined previously.
*/
kernel: tensor4D(kernel),
/**
* Sets the kernel size for a convolutional layer in a neural network.
* The kernel size is a 4D array that specifies the dimensions of the kernel that will be used to convolve the input tensor.
* In this case, the kernel size is set to [1, 1, 3, 3],
* which means that the kernel will have a depth of 1, a height of 3, and a width of 3.
* This kernel size is used in the creation of two different convolutional layers in the code below the selection.
*/
kernelSize: [1, 1, 3, 3],
/**
* Sets the padding for a convolutional layer in a neural network.
* Padding is used to add extra pixels around the edges of an image
* to ensure that the output of the convolution operation has the same dimensions as the input.
* In this case, the padding is set to [1, 1],
* which means that one pixel of padding will be added to the top, bottom, left, and right edges of the input tensor.
*/
padding: [1, 1],
}),
/**
* Creates an average pooling layer in a neural network.
* Average pooling is a technique used in Convolutional Neural Networks (CNNs) to reduce the spatial dimensions of the output volume.
* It is used to downsample the input along the spatial dimensions (width and height) while keeping the depth constant.
* The average pooling layer takes a tensor as input and applies an average filter to non-overlapping subregions of the tensor.
* The output of the average pooling layer is a tensor with reduced spatial dimensions and the same depth as the input tensor.
*/
AveragePool2DLayer({ strides: [2, 2] }),
]);
/**
* Draw the output of the Average Pooling layer example on the canvas.
*/
drawPixels(pool2.data, pool2.shape[2], 280, 280, 2);
/**
* Creates a network with the given layers and feeds forward the input data.
*/
async function feedForward(layers: Layer[]) {
const net = new Sequential({
size: [1, 1, dim, dim],
silent: true,
layers,
cost: Cost.MSE,
});
const data = new Tensor(buffer, [1, dim, dim]);
return (await net.predict(data)) as Tensor<Rank.R4>;
}
/**
* Function to draw the pixels on the canvas.
*/
function drawPixels(
buffer: Float32Array,
dim: number,
offsetX = 0,
offsetY = 0,
scale = 1,
) {
for (let i = 0; i < dim; i++) {
for (let j = 0; j < dim; j++) {
const pixel = buffer[j * dim + i];
ctx.fillStyle = `rgb(${pixel}, ${pixel}, ${pixel})`;
ctx.fillRect(
i * 10 * scale + offsetX,
j * 10 * scale + offsetY,
10 * scale,
10 * scale,
);
}
}
}
/**
* Save the canvas as a PNG file.
*/
await Deno.writeFile("./examples/filters/output.png", canvas.toBuffer());