Skip to content
View appliedml42's full-sized avatar
🤖
Learning
🤖
Learning
Block or Report

Block or report appliedml42

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Please don't include any personal information such as legal names or email addresses. Maximum 100 characters, markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
appliedml42/README.md

Please reach out if you would like to consult on your AI/ML project.

Screenshot 2022-12-03 at 8 10 52 PM

     

I have used Machine Learning, Deep Learning, and Natural Language Processing to bring value to organizations like Amazon and Tinder since 2011. Andrew Ng's lecture at Amazon introduced me to Deep Learning in the early 2010s. I have been hooked ever since.

AI Self Learning

Experience

2024-Present

Coming Soon

2023-2024

AI Technical Lead, Various Startups

In early 2023, I decided to resign from Tinder and explore the AI startup landscape by working at different startups. During this exploration, I worked at a few startups and got hands-on experience on how Large Language Models can be leveraged to build products.

  • Features:
    • Record Matching between Salesforce records and incomplete leads using a cascade of GPT 3.5 Turbo and GPT 4 using Langchain for async batching and calling the API.
    • Rate professionals for job placements based on their work experience using GPT 4.
  • Fine-tuning: Fine-tuned open-source LLMs with LoRA, replacing costly proprietary systems to improve cost and stability.
  • Inference: I advised infrastructure engineers, did node local benchmarks for LLM inference engines like vLLM, and evaluated services like Together.ai and Fireworks.ai.
  • Knowledge: Leadership: Stayed abreast of the latest SOTA transformer architectures, such as KV Cache, Multi-Query Attention, and RoPE, ensuring the dissemination of this knowledge.
  • Evaluation: Built a UI tool using Streamlit and ideas from different evaluation repositories like LLMJudge to build a tool to compare the performance of different LLMs on a given task.

Staff MLE & Technical Lead, Tinder

Technical Lead for a Machine Learning Engineers & Risk Analysts team. I started the group and grew it to 9 people. This team owns detection models/algorithms for identifying Trust & Safety violations.

  • Responsible for hiring, mentoring, and team charter.
  • Organized Risk Analysts under the Escalations and Early Warning team. Collaborated with the team to invent KPIs for the org. This team optimizes detection strategies using existing signals and provides feedback highlighting detection gaps.
  • Removed data scarcity by establishing best practices for leveraging moderation logs for training and evaluation datasets.
  • Established modeling best practices. Transformers for text classification by pre-training on Tinder’s unique data; Text pre-processing pipelines to handle adversarial attacks; Metadata embeddings to improve multi-lingual performance; ConvNeXT family of models for effective and efficient image classifiers.
  • Enabled real-time model inference. TFLite and Quantization for CPU-based models; TensorRT and Triton for GPU-based models; Collaborated with Infra team to establish requirements for the K8-based model endpoint framework.
2019-2023

Staff MLE & Technical Lead, Tinder

Technical Lead for a Machine Learning Engineers & Risk Analysts team. I started the group and grew it to 9 people. This team owns detection models/algorithms for identifying Trust & Safety violations.

  • Responsible for hiring, mentoring, and team charter.
  • Organized Risk Analysts under the Escalations and Early Warning team. Collaborated with the team to invent KPIs for the org. This team optimizes detection strategies using existing signals and provides feedback highlighting detection gaps.
  • Removed data scarcity by establishing best practices for leveraging moderation logs for training and evaluation datasets.
  • Established modeling best practices. Transformers for text classification by pre-training on Tinder’s unique data; Text pre-processing pipelines to handle adversarial attacks; Metadata embeddings to improve multi-lingual performance; ConvNeXT family of models for effective and efficient image classifiers.
  • Enabled real-time model inference. TFLite and Quantization for CPU-based models; TensorRT and Triton for GPU-based models; Collaborated with Infra team to establish requirements for the K8-based model endpoint framework.
2015-2019

Senior Applied Scientist, Amazon

Technical lead for the query understanding science team.

  • Led the Research & Development of the first query -> shopping intent Recurrent Neural Network model. This model became the foundation for multiple shopping experiences that required mapping of queries to shopping intents.
  • Established best practices for training Deep Learning models and their deployment for the team. Built GPU-based training and batch-inference infrastructure over AWS Batch.
  • Collaborated closely with the infrastructure team to establish requirements for the inference framework. Made sure the team invested in a Python-based framework to make model deployments painless.
2011-2015

Senior Software Development Engineer

Founding engineer on the X-Ray team. X-Ray identifies a book’s topics, characters, images, and essential passages.

  • Developed algorithms to identify characters and all their mentions in a book. For example, the algorithm will ensure that Mr. Potter and Harry Potter are associated with the same character Harry Potter.
  • Led the Book Maps teams and developed algorithms to identify images and critical passages in a book.
  • Developed testing at scale methodologies to go code complete months before the deadline. Concretely, each new algorithm change was tested against thousands of books, and the most common bugs were sorted by frequency and prioritized.
  • Generated n-gram count statistics across the whole Kindle catalog using Map-Reduce and Dynamo DB. This data is used for topic modeling.
  • Developed the Kindle N-gram Corpus. N-gram frequency table using Map-Reduce and DynamodDB across all Kindle books circa 2012. Essential topics in books are identified using this dataset.
Patents & Publications

Technologies & Tools

Pinned

  1. MultiHeadAttention Implementation us... MultiHeadAttention Implementation using einops and Pytorch
    1
    '''
    2
    I am reading this amazing series(https://uvadlc-notebooks.readthedocs.io/en/latest). I always struggle with revisiting 
    3
    my old code that has a lot of tensor manipulation. Experimented with reimplementing their MultiHeadAttention layer using 
    4
    einops syntax that feels more human readable.
    5
    '''
  2. LLMs LLMs Public

    Transformer Language Models using Pytorch Lightning.

  3. docker docker Public

    Development environment for all other projects.

    Dockerfile