Prakash Bisht

Machine Learning Engineer at IU-Health · NLP Researcher at Eller, University of Arizona · Phoenix, AZ

A passionate learner, with deep rooted interest in Science.

Hi there! I currently work as an ML Engineer at IU-Health, Indianapolis, IN. I graduated in Dec '23 from The University of Arizona with an M.S. in Data Science. My expertise spans Retrieval-Augmented Generation (RAG), open-source Large Language Models (LLMs), NLP, Deep Learning, Agentic Workflows, and ML System Design. I have hands-on experience with frameworks like Langchain and Llamaindex for building RAG pipelines and implementing Agentic Workflows. I also possess knowledge in Reinforcement Learning (RL) agents and technology for aligning LLMs with human values using Reinforcement Learning from Human Feedback (RLHF) techniques.
With over 6 years of experience in roles such as Data Scientist-I, ML Engineer, and Open-Source ML Developer, I have contributed to industries including Healthcare, EdTech, and Open-Source research. I am skilled in Python, SQL, GCP, AWS, Data Analytics, R, Data Visualization, and Statistical Modeling. My proficiency extends to popular ML frameworks like TensorFlow, PyTorch, and Hugging Face, and I have hands-on experience in API Development, MLOps, and cloud platforms like AWS and Google Cloud. In my current role, I’ve led the development and deployment of impactful ML models that improved patient outcomes and streamlined business operations, resulting in significant cost savings.

Outside work, I enjoy hiking and working out. : )

I am passionate about advancements in ML and Generative AI (GenAI) and am keen on roles in AI Research, Machine Learning Engineering, and ML Research Engineering. If you find my profile aligned with your needs or if your organization is hiring, feel free to reach out at my email

Education

University of Arizona

School of Information, University of Arizona, Tucson, Arizona

Master of Science in Data Science


CGPA: 3.80 / 4.00
Aug, 2022 - Dec, 2023

Sharda University

School of Engineering & Technology, Sharda University, Uttar Pradesh, India

Bachelor of Technology in Mechanical Engineering


CGPA: 3.80 / 4.00
First Division (equivalent to magna cum laude)
Aug 2012 - May 2016

Experience

Machine Learning Engineer

Indiana University Health, Indianapolis, IN

  • Developed a two-part deep learning solution using 2D CNNs to predict Cardiac Ablation Sites from ECG data, achieving 95% accuracy in PVC Identification and 74% accuracy in site prediction. This reduced ventricle mapping time by 1 hour per procedure, increased daily procedure capacity by 50%, enhanced patient outcomes, and generated an additional $28K in revenue per surgery, while saving up to $30,800 in manual labor costs.
  • Implemented a distributed DL model predicting hospital length of stay with 87.3% accuracy, saving $55M annually by enhancing resource management. Ingested and processed 940M data points from Google Cloud Storage into BigQuery, using TensorFlow-io on PySpark clusters with GPUs on DataProc for training, and integrated MLflow for model tracking, streamlining patient throughput and care..
  • Implemented a Retrieval Augmented Generation (RAG) pipeline using Phi-3-Mini model, converting thousands of pharmacy contracts into a vector store. Automated the extraction of critical business information for decision making and vendor negotiation by fine-tuning a FLAN-T5 model for QnA task, reducing manual review time by hundred of hours, enhancing efficiency, and improving contract management and negotiation processes.
Nov 2023 - Present

Machine Learning Intern

Lightsense Technology, Inc., Tucson, AZ

  • Developed a robust fact-checking automation pipeline using OpenAI’s Large Language Models alongside FAISS (Facebook AI Similarity Search) library to validate LLMs generated outputs, minimizing erroneous outputs by 20%, by devising custom NLP models and techniques to identify and mitigate the hallucinating tendency of LLMs in producing unverified information.
  • Fostered the potential of LLMs in scientific research, and anticipated 4 times increase in grant writing procedure efficiency owing to LLMs superior Natural Language Understanding, text summarizing, reason based probing & data generation capabilities, by exploring and implementing their use in spectroscopic technology and research grant development.
  • Utilized classic ML techniques to uncover latent patterns, explore anomalies in dataset encompassing findings from various Spectroscopic technologies such as UV-Vis absorbance, emissions & scattering and Fourier Transform Infrared spectroscopy with an objective to advance efforts to address fast diagnosis for variety of medical conditions using ML in spectroscopy..
June 2023 - Aug 2023

NLP Summer Research Intern

Deep Target NLP Research Group, Tucson, AZ

  • Improved precision by 15% over baseline models by fine-tuning and fusing BioBERT and ClinicalBERT models for Electronic Health Records on ASD. Enhanced this by addressing data imbalance, leading to a boost in precision and recall across diverse label distributions and an increased token representation quality for EHR texts.
  • Circumvented GPU memory limitations using gradient accumulation and summarization, combined with a dynamic learning rate scheduler for SGD, resulting in 15% faster convergence. Further streamlined data processing, improving throughput by 25% and reduced inference time by 20% ensuring robust real-time response.
  • Delved deep into transformer model behaviors, revealing insights leading to a 10% precision boost on edge cases, while deploying techniques such as knowledge distillation to ensure swift real-time responses in production environments.
  • Spearheaded initiatives that reduced clinicians' diagnosis time by 50% by collaborating cross-functionally and integrating insights from data engineering, UI/UX teams, and expert clinicians.
May 2023 - Aug 2023

Data Scientist

Embibe, Bangalore, India

  • Implemented Deep Item Response Theory (D-IRT) data-driven models via scalable deep learning frameworks that gauged latent traits of 8 lakh students & million questions at once dropping annotation efforts by over 90 percent, as measured by the increased efficiency and accuracy of the data analysis, by utilising Neural Architecture Search (NAS) for model search, custom data generators for handling very large datasets, and exporting models to ONNX for model interoperability.
  • Engineered “Model Distillation Techniques” to interpret blackbox DL models and applied them to D-IRT models, resulting in the successful extraction of crucial item characteristics such as 'Difficulty' and ‘Discrimination’, and validated the effectiveness through extensive experimentation.
  • Built a unified “Automatic Test Generation Utility” for faculties to generate thousands of tests on-demand & validate before commissioning.
  • Collaborated with 6 members agile team to develop features and REST APIs for ATG in a timely manner.
Jan 2020 - Mar 2021

ML Engineer

mlsquare, Bangalore, India

  • Developed a python package to facilitate Interoperablity amongst different ML frameworks- (Sklearn-Tensorflow, Surpriselib-Pytorch. etc)
  • Implemented an sklearn compatible interface for SVD via an appropriate DL model to support interoperable machine learning.
  • Built a NAS interface for AutoML.
  • Designed and Implemented a surpriselib compatible interface for collaborative filtering via Deep Factorisation Machines.
  • Analysed multiple public datasets in Bayesian workflow using Pyro & contributing chapters to “Probabilistic Programming with Pyro”.
  • Investigated meta & curriculum learning strategies to immune training regime to predictor outliers, fostering resource- efficient machine learning.
Apr 2021 - Mar 2022

ML Engineer

IcarusNova Discovery Pvt Ltd, Bangalore, India

  • Collaborated on “diet analysis & moderating diet patterns of underprivileged diabetic patients using Deep learning”:
    • Conceptualized a DL pipeline constituting-- data collection, clustering, training and deployment on cloud as MaaS (Model as a Service) supplied search space.
  • Implemented a variant of RCNN for food detection, and deployed it as Flask web-app using Heroku.
  • July 2017 - July 2019

    Research Experience

    Graduate Research Assistant (NLP)

    Eller College of Management, Tucson, AZ

    Leveraging Natural Language Understanding to develop NLP models for “Autism Spectrum Disorder (ASD) Risk Assessment for Early Diagnosis” by analysing the semantics of Electronic Health Records (NIH-NIMH Grant).

    • Implemented a boilerplate wrapper which was capable of loading several individual label Bi-GRU models, test dataset passed in different formats and output predictions from ensemble of models.
    • Developed a multilabel model, which outperformed the Ensemble of 7 individual models by reducing the prediction time by factor 130, the outcome was a result of employing more vectorised operations thus optimizing the preprocessing phase significantly.
    • Experimented with various adversarial retraining configurations for built multilabel model and benchmarked performance against ensemble of individual models as follows, Few of the following modifications to the model configuration were done to pose adversaries to the model and make it more robust:
      • GloVe 42B word vector used in embedding instead of GloVe 6B.
      • Synonym Replacement Attack: a method to generate adversarial examples by replacing words in the input text with their synonyms. This approach can potentially confuse the classifier by maintaining the overall meaning of the sentence while changing specific words by utilizing ‘WordNet’ to find synonyms.
      • Random Character Swap Attack: It generates adversarial examples by randomly swapping characters in the input text. This method aims to introduce noise into the text, making it more challenging for the classifier to recognize patterns. It swaps a specified number of character pairs to introduce noise into the text.
      • Oversampling: It is used to balance the training dataset by oversampling the underrepresented classes.
    • The final multilabel model trained using adversarial retraining in combination with oversampling was robust to adversarial samples and gave significant improvement in precision for class labels which had low representation in training dataset therefore poor performance.
    Sept 2022 - Dec 2023

    Projects

    Implementing ML techniques to retrieve the underlying factors for a given medical conditions

    Named Entity Recognition Natural Language Processing Spacy Python TensorFlow Latent Dirichlet Allocation

    Using patient records dataset comprising of several .txt files that contained information with pertinent sections such as the 'Discharge Diagnosis of patient', 'Chief Complaint' and 'History of Present Illness of patient'. With each text file there were annotations representing the output of a named entity recognition process to help compliment the factors found during modelling. I applied Machine Learning techniques to uncover the common underlying factors for a given medical condition in the given patient records.

    Robust NLP Model for Early Diagnosis of Autism Spectrum Disorder

    Natural Language Processing Robust ML Python TensorFlow Deep Learning

    Using clinical text data obtained from patients’ records and notes, Our project aimed to develop a robust deep- learning model for early diagnosis of Autism Spectrum Disorder (ASD). The main challenge in this task is the scarcity of labeled data, which limits the model’s ability to generalize well across various clinical scenarios. To address this issue, adversarial train- ing techniques are used to improve the model’s generalization and performance on unseen data. Furthermore, the proposed model employs advanced natural language processing techniques to capture the nuances and patterns found in clinical texts. This approach is expected to improve the accuracy, reliability, and robustness of clinical decision support systems for ASD diagnosis, resulting in better patient outcomes and more efficient healthcare delivery.

    The project was done towards the completion of my course ECE 696: Advanced topics in Electrical Eng. - Differential Privacy, Robustness & Fairness of ML models.

    Detecting all the time expressions from given corpora of text data

    Natural Language Processing KerasTuner Python TensorFlow Optimization

    We were given a thousand text samples for training & some 130 for test scrapped from emails in Anafora XML format. The task was to develope and optimize a NLP model to first identify and predict all the pieces of time-expressions from the given texts into 28 target label used to express or hint time in various contexts of a natural langauge, such as 'Month-Of-Year', 'Day-Of-Month', 'time-of-the-day', 'AMPM-Of-Day', 'Hour-Of-Day', 'Minute-Of-Hour', 'Month-Of-Year', 'Part-Of-Week', 'Season-Of-Year', 'Second-Of-Minute', 'Year' etc.

    The project was done towards the completion of my course INFO 557: Neural Networks.

    Open Source Contributions

    mlsquare: a framework to democratise AI [link]

    Python TensorFlow ML Engineering ML Systems

    ML Square is a python library that utilises deep learning techniques to

    • Enable interoperability between existing standard machine learning frameworks.
    • Provide explainability as a first-class function.
    • Make ML self learnable.


    Skills

    Core Skills
    • Natural Language Processing
    • Computer Vision
    • Deep Learning


    • Machine Learning
    • Exploratory Data Analysis


    • Statistical Modelling & Visualisation
    • Image Processing


    • Database Management
    Languages & OS
    • Python
    • R
    • C/C++
    • SQL
    • MATLAB
    • Pyro


    • MacOS
    • Linux
    • Windows
    Tools, Frameworks & Technologies
    • Pytorch
    • Tensorflow
    • Pyro
    • SciKit-Learn
    • REST APIs
    • Paperspace


    • Google Cloud Platform
    • AWS
    • SQLAlchemy
    • ElasticSearch


    • Docker
    • CI/CD
    • Keras
    • NLTK
    • OpenCV
    • Genism
    • SpaCy
    • Jupyter


    • Git
    • Flask
    • Postman
    • PyCharm
    • RStudio
    • JIRA
    Languages and OS
    • Python
    • R
    • C/C++
    • SQL
    • MATLAB
    • Pyro


    • MacOS
    • linux
    • Ubuntu
    • Windows
    Database Technologies
    • MySQL
    • PostgreSQL
    • MongoDB
    • SQLLite
    • Redis
    Dev Ops
    • Heroku
    • Travis
    • AWS
    Organization
    • Trello
    • Scrum

    Publications

    Crowdsourcing with Enhanced Data Quality Assurance: An Efficient Approach to Mitigate Resource Scarcity Challenges in Training Large Language Models for Healthcare [DOI]

    P. Barai, Prakash Bisht, G. Leroy, J. M. Rothman, S. Lee, J. Andrews, S. A. Rice, A. Ahmed.
    American Medical Informatics Association (AMIA) Summit, Boston, 2024 (Distinguished paper award – data science and AI category)



    Utilizing Large Language Models to Generate Synthetic Data to Increase the Performance of BERT-Based Neural Networks [DOI]

    Chancellor R. Woolsey, Prakash Bisht, Joshua Rothman, Gondy Leroy.
    American Medical Informatics Association (AMIA) Summit, Boston, 2024



    A Deep Learning framework for Interoperable Machine Learning [DOI]

    Shakkeel A, Prakash Bisht, Ravi M, and Soma S D.
    AIMLSystems 2021. Association for Computing Machinery, New York, NY, USA, Article 23, 1–7



    Probabilistic Programming with Pyro [ebook, url*]

    Prakash Bisht, and Soma S Dhavala.



    Nifty tech tag lists from Wouter Beeftink