diff --git a/index.html b/index.html new file mode 100644 index 0000000..a99aec5 --- /dev/null +++ b/index.html @@ -0,0 +1,191 @@ + + + +
+ + + + ++ Split Computing (SC), where a Deep Neural Network (DNN) is intelligently split with a part of it deployed on an edge device and the rest on a remote server is emerging as a promising approach. + It allows the power of DNNs to be leveraged for latency-sensitive applications that do not allow the entire DNN to be deployed remotely, while not having sufficient computation bandwidth available locally. + In many such embedded systems scenarios, such as those in the automotive domain, computational resource constraints also necessitate Multi-Task Learning (MTL), where the same DNN is used for multiple inference tasks instead of having dedicated DNNs for each ask, which would need more computing bandwidth. + However, how to partition such a multi-tasking DNN to be deployed within a SC framework has not been sufficiently studied. + This paper studies this problem, and MTL-Split, our novel proposed architecture, shows encouraging results on both synthetic and real-world data. +
+
+
+ TBA.
+
+
+