I am currently a ML-Engineer at Apple Inc., where I am
involved in scaling
and optimization of Siri invocation stack involving ML-algorithms across all Apple Devices.
Previously, I was a researcher/computer-scientist at Stanford
Research Institute (SRI) International, where I worked on cross-disciplinary machine learning
& deep learning
applications.
At SRI Intl. I've worked on Model Quantization/Pruning (SRI spin-off LatentAI), AMD and DASL.
I'm interested in applied machine learning, optimization, computer vision, model
quantization/pruning.
I delve much into cross-disciplinary projects and these are some of my contributions.
We improve the voice trigger detection performance of smartphone assistant using small
number of utterances from target speaker in a multi-task learning framework to output a
personalized embedding for each utterance.
We introduce a modified Transformer based architecture to take in atom-level feature vectors as
inputs and aim to encode a a better molecule representation space which helps effectively include
the structure-property relations of a molecule.
We propose Common-Sense based Neuro-Symbolic Loss (CSNL) that formulates prior knowledge as novel
neuro-symbolic loss functions that regularize visual-semantic embedding. CSNL forces visual features
in the VSE to obey common-sense rules relating to hypernyms and attributes. This uses the DASL framework.
We present the comparison of data-free post-training quantization methods for state-of-art models
trained on datasets like ImageNet. To better analyze quantization results, we describe the overall
range and local sparsity of valuesafforded through various quantization schemes.
Prior Experience/Positions
Stanford Research Institute (SRI) Intl., San Diego
CA [Feb'20 - Mar'21] Computer Scientist
Advanced Analytics/Aritifical Intelligence Center (AIC) group; Research focus on
Molecular-Property prediction,
DASL, and Zero-Shot Learning.
Stanford Research Institute (SRI) Intl., Princeton
NJ [Feb'19 - Aug'19] Deep Learning & Optimization Intern
Computer Vision Group; Research focus on Post-Training and Training-Aware DNN Model Quantization
techniques.
Developed framework for post-training quantization for weights-only 8-bit or lower precision for
previously
trained in Tensorflow, Darknet(Yolo)
Univ. of Minnesota, Twin Cities [Aug'18 -
Dec'18] Graduate Teaching Assistant
Responsible for grading, preparing and teaching course materials for Linear Control Systems
EE-4231
Boston Scientific, Arden Hills MN [May'18 -
Aug'18] Machine Learning Intern
Corporate Research Team; Developed end-to-end machine learning pipelines for analysing time-series
data obtained from wearable devices of Chronic-Pain patients to find correlation between the
physiological signals and reported pain scores.
Aindra Systems, Bangalore, IND [Feb'17 -
Jun'17] Engineering Intern
First taste of Machine-Learning & Deep Learning domain while working with R&D engineers on
detection of cervical
cancerous cells in high-res pap smear images. Contributed parts of the pipeline mainly focused on
image processing algorithms for robust segmentation of cell nuclei.