As a part of Google Health, I led the project "Graph Convolutional Transformer", which aimed to capture the hidden graphical structure of electronic health records (EHR) to gain better prediction performance. This work was presented at AAAI 2020. I also participated in a project to analyze how the model uncertainty affects the neural network prediction models for EHR data, which was presented at CHIL 2020.Research Internship at Google Research, Mountain View, California, May 2017 - Aug 2017
I was a member of the project team named MorphNet. The objective was to automatically learn the structure of neural networks given some resource constraint (e.g. number of parameters, number of FLOPs), using various regularization methods. My specific task was related to NLP applications and I modified several variable selection algorithms such as Smoothly Clipped Absolute Deviation (SCAD) to activate/deactivate groups of parameters. MorphNet was presented at CVPR 2018.Research Internship at DeepMind, London, U.K., Feb 2017 - May 2017
My first project was to train an embodied agent to find out the heaviest object in a virtual environment. This was an extended work of "Which is heavier?" experiment from Learning to Perform Physics Experiments via Deep Reinforcement Learning (Denil et al. ICLR 2017). The agent was equipped with a hammer to probe the objects, and a positive reward was given when the hammer was in contact with the heaviest object (hence the project name Pinata). The agent successfully learned to interact with the objects and stick to the heaviest one (example video 1, example video 2).
My second project was training neural agents to develop compositional language purely from raw pixels by playing an image description game. By employing a communication strategy named Obverter, which is motivated by the theory of mind, we confirmed the two agents could develop a highly structured/patterned communication protocol. This work will be presented in ICLR 2018.
In my second internship at Sutter Health, I focused on developing interpretable deep learning models for predictive healthcare. Specifically, using the neural attention mechanism combined with RNN and MLP, I was able to design a sequence prediction model that demonstrated similar AUC as RNN but completely interpretable; the model allowed precise calculation of how much each diagnosis/medication/procedure in the past visits contributed to the final prediction. This work, named RETAIN, was presented in NIPS 2016.Internship at Research, Development and Dissemination (RD&D), Sutter Health, California, May 2015 - August 2015
During the internship, I explored the potential of applying deep learning methods to health care problems, specifically predicting the future heart failure diagnosis. Applying stacked de-noising auto encoders to heart failure prediction enabled sophisticated analysis of the relation between patient features and heart failure diagnosis. Furthermore, through the combination of the word embedding technique and recurrent neural networks, I was able to improve the heart failure prediction performance from 0.81 AUC to 0.86 AUC. This work was published in JAMIA.Research and development at Knowledge Mining Research Team, Electronics and Telecommunications Research Institute (ETRI), South Korea, February 2010 - April 2014
My first task was to build a Hadoop and HBase cluster to process Big Data (news, blogs, Tweets). Then I taught myself to write MapReduce codes for text analysis. This changed the way my team process text data, from dealing several tens of thousand documents on a few separate machines to systematically analyzing millions of documents per day on a cluster and saving the results to a distributed database. Then I worked on Named Entity Recognition module and Event Extraction module, while simultaneously studying machine learning techniques. I also developed solvers for Binary SVM, Structural SVM, One-class SVM, Ranking SVM in C++ and Java.Project Participation