Skip to main content Link Menu Expand (external link) Document Search Copy Copied
Jinhee Kim
I'm a 4th-year undergraduate student advised by Prof. Jong Hwan Ko in the IRIS Lab at Sungkyunkwan University. My research interests are:
On-device AI · Dynamic Neural Networks · Model Quantization.


Today, we live in a world where artificial intelligence shows performance on par, or even better, than humans. Despite this, fields outside the science community still struggle to harness the full power of AI and bring it to their own domains, due to the high computational and memory demands. To make AI accessible to a wider audience, it's important to:
  1. Reduce the cost of inference, which impacts key user requirements (latency & performance)
  2. Streamline the development process.
To address this, my research focuses on using quantization techniques to dynamically compress NNs at runtime based on resource availability. Specifically, I am exploring weight sharing and flexible bit-precision within dynamic neural networks to enable more efficient deployment on edge devices.

What are dynamic NNs? More here.

RECENT FOCUSES

More efficient flexible bit precision models.
During runtime, dynamic neural networks generally use quantization as a means to change the bit precision of a model. However, the existing quantization method does not allow the sharing of most-significant-bits (MSBs), which leads to the model having to reference the full-precision value every time it wants to convert to a lower precision version of itself. This incurs memory and computational overhead, which can be reduced by a simple solution- bit-shifting. My current research explores how neural networks react to straightfowardly bit-shifting their weights, and the impact this has to the overall performance.

Shorter training time for dynamic neural networks.
Dynamic neural networks achieve great robustness during runtime by training for every scenario beforehand. Naturally, this leads to an exponential growth in training time, which 1) not only hinders the deployment of these models but also 2) results in huge carbon emissions. Interestingly, dynamic neural networks need not to train for every setting it will encounter in runtime, but just a judiciously selected few which can maintain performance for most settings of deployment. My current research focuses on how to find these selected few, and ultimately optimize the training process for dynamic neural networks.

RECENT POSTS