📚 Natural Language Processing from Stanford University: Distilled Notes
👉🏼
- NLP is one of the most popular
#AI
domains, widely used from language translation to auto-complete to voice assistants.
- Presenting notes from Stanf…
📚 Ilya Sutskever's Top 30 AI Papers •
- Ilya Sutskever shared a list of 30 papers with John Carmack, saying, “If you really learn all of these, you’ll know 90% of what matters today.”
- In this article on , we have reviewed all of
🚀 Launching - your one-stop shop for free AI resources:
- This project has been a work in progress for over a year, and I'm thrilled to finally share it with you.
- is a comprehensive platform offering a
🎉 Thrilled to announce that we received an Outstanding Paper Award at
#EMNLP2023
!
🔷Counter Turing Test (CT^2): AI-Generated Text Detection is Not as Easy as You May Think - Introducing AI Detectability Index
➡️
@emnlpmeeting
#ArtificialIntelligence
#AI
📝 Announcing our paper surveying Multimodal AI Architectures -- with a comprehensive taxonomy and analysis of their pros/cons & applications in any-to-any modality model development
➡️ 𝐂𝐨𝐦𝐩𝐫𝐞𝐡𝐞𝐧𝐬𝐢𝐯𝐞 𝐓𝐚𝐱𝐨𝐧𝐨𝐦𝐲: First work to explicitly identify and categorize
I’m
#hiring
for full-time Generative AI roles (Applied Scientists and Deep Learning Architects) in Amazon Web Services (
@awscloud
). We're seeking exceptional talent skilled in deep learning, eager to work on cutting-edge AI, and interested in building the next generation of
🧠 All you need to know about Vision-Language Models (VLMs) / Multimodal Large Language Models (MLLMs) •
• Overview
• Architecture
• Architecture of Vision-Language Models
• Examples of Popular VLMs and Their Architectural Choices
• Differences from
I’m
#hiring
for full-time Generative AI roles in
@awscloud
(Amazon Web Services). We're seeking exceptional talent skilled in deep learning, eager to work on cutting-edge AI, and interested in building the next generation of AI-powered solutions (powered by LLMs, VLMs, Diffusion
I’m
#hiring
for full-time Generative AI roles in
@awscloud
(Amazon Web Services). We're seeking exceptional talent skilled in deep learning, eager to work on cutting-edge AI, and interested in building the next generation of AI-powered solutions (powered by LLMs, VLMs, Diffusion
📝Announcing our survey paper covering 30+ Prompt Engineering techniques
🔹"A Systematic Survey of Prompt Engineering in Large Language Models: Techniques and Applications"
🔹In collaboration with IIT Patna
➡️
@IITPAT
#ArtificialIntelligence
#LLM
#ChatGPT
🎨 Diffusion Models Primer •
- Diffusion models are all the rage right now owing to their extraordinary applications in text-to-image applications, which has given AI wings to generate art from text prompts.
- Here’s my primer on diffusion models that
📈 Mixture-of-Experts (MoE) Primer •
- MoE enhances model performance by dynamically selecting specialized subnetworks for different inputs, improving efficiency and scalability. This approach reduces computational cost while maintaining high accuracy,
📝 Announcing our paper on Energy-Based World Models (EBWM), an architecture that emulates facets of human cognition to improve the scalability and performance of autoregressive world models across Computer Vision and Natural Language Processing.
➡️ 𝐍𝐨𝐯𝐞𝐥
📝 All you need to know about LLM Prompt Engineering •
⚡️ This primer presents 30+ flavors of Prompt Engineering along with their use-cases.
#ArtificialIntelligence
#GenAI
#LLMs
📝 Announcing our new paper that proposes a framework to enhance causal reasoning & explainability in LLMs
🔹"Cause and Effect: Can Large Language Models Truly Understand Causality?"
🔹With
@CarnegieMellon
,
@UNTsocial
,
@rpi
, and
@UMassAmherst
🔗
#AI
#LLM
📈 Mixture-of-Experts (MoE) Primer •
- MoE enhances model performance by dynamically selecting specialized subnetworks for different inputs, improving efficiency and scalability. This approach reduces computational cost while maintaining high accuracy,
📖 The A-to-Z of Large Language Models (LLMs) •
- What are Embeddings?
- Contextualized vs. Non-Contextualized Embeddings
- How do LLMs work?
- LLM Training Steps
- Computing Similarity between Embeddings (Dot Product Similarity, Geometric Intuition,
📚 All the Math fundamentals for AI (including Backpropagation primers)
- Don’t let the math behind AI concepts hinder you from understanding what goes on under-the-hood as you train your neural network. Think in terms of first principles to develop an intuition of what’s going
📝 Looking to learn about Hallucination in LLMs?
✅ Our tutorial at
@LrecColing
2024 will offer an introduction to the issue of hallucination, present a taxonomy of categories, and cover hallucination detection and mitigation methods.
🔗 Tutorial Schedule:
🎉 Happy to share that the Wikipedia article on AI Hallucination has adopted the definition in our paper!
📖 AI Hallucination Wiki: )
📝 Our paper:
👍🏼 Shoutout to my collaborators!
#GenAI
🧠 A Detailed Overview of Vision-Language Models (VLMs) •
✅An overview of VLMs and ~40 popular VLM models for Generation and Understanding such as GPT-4V, LLaVA, Frozen, Flamingo, PaLM-E, MiniGPT, etc.
#LLMs
#ArtificialIntelligence
#MachineLearning
#AI
📝 Top papers in Computer Vision, NLP, Speech, Multimodal AI, and Core ML
👉🏼
I’ve put together a summary of key papers in Computer Vision, NLP, and Speech and segregated them into 1️⃣ need-to-know and 2️⃣ good-to-know.
What’s…
🎖️Recommender Systems' Architectures:
➡️ This primer explores the most popular neural architectures used in recommender systems.
🔷 Wide and Deep, DeepFM, NCF, DCN, AutoInt, DCN V2, DHEN, GDCN
#RecSys
#ArtificialIntelligence
#MachineLearning
#AI
📝 Recommender Systems in Production: Case Studies
🔹Synopses of the inner-workings of some of the most popular recommendation system platforms:
-
@tiktok_us
's Monolith Recommender System by Zhuoran Liu et al.
-
@netflix
's Recommender System by Carlos
📝Announcing our
@eaclmeeting
2024 papers
🔹On the Relationship between Sentence Analogy Identification and Sentence Structure Encoding in LLMs
➡️
🔹Generative Data Augmentation using LLMs Improves Distributional Robustness in QA
➡️
📈 Graph Neural Networks (GNNs) Primers: Overview, Applications in RecSys and NLP, and Case Studies
➡️ Overview of GNNs in RecSys and NLP:
➡️ GNN-based RecSys with Case-Studies from Snap, Meta, Google, Uber, Pinterest:
#AI
🧠 A Detailed Overview of Vision-Language Models (VLMs) |
✅ An overview of VLMs and ~40 popular VLM models for Generation and Understanding such as GPT-4V, LLaVA, Frozen, Flamingo, PaLM-E, MiniGPT, etc.
#LLMs
#ArtificialIntelligence
#MachineLearning
#AI
📝 Announcing our new paper surveying 32 methods for mitigating hallucinations in LLMs
🔷 A Comprehensive Survey of Hallucination Mitigation Techniques in Large Language Models
➡️
#LLMs
#ArtificialIntelligence
#MachineLearning
#AI
📖All you need to know about LLMs:
➡️Embeddings
➡️LLM Training Steps
➡️Context Length Scaling
➡️Vector DBs
➡️Knowledge Augmenting LLMs (RAG, Prompting, Finetuning)
➡️50+ Popular LLMs (+Code / Indic / Medical LLMs)
➡️Leaderboards
➡️Frameworks
#GenAI
#LLMs
📝 Announcing our paper that investigates biases in LLMs used for hate speech detection (focusing on gender, race, religion, and disability biases) and proposes mitigation strategies
➡️ 𝐃𝐞𝐭𝐞𝐜𝐭𝐢𝐨𝐧 𝐨𝐟 𝐀𝐧𝐧𝐨𝐭𝐚𝐭𝐨𝐫 𝐁𝐢𝐚𝐬: We demonstrate the presence of gender,
🧠 All you need to know about Vision-Language Models (VLMs) |
➡️ A primer that offers an overview of VLMs and ~25 popular VLM architectures such as GPT-4V, LLaVA, Frozen, Flamingo, PaLM-E, MiniGPT, etc.
#ArtificialIntelligence
#MachineLearning
#AI
📝 Announcing our paper that unveils the lack of cultural awareness in Vision-Language Models (VLMs)!
➡️ 𝐂𝐮𝐥𝐭𝐮𝐫𝐚𝐥 𝐀𝐰𝐚𝐫𝐞𝐧𝐞𝐬𝐬 𝐒𝐜𝐨𝐫𝐞 (𝐂𝐀𝐒): A novel metric proposed to measure the inclusion of culturally relevant information in image captions generated by
📚 Notes and Videos on Multimodal Machine Learning
- The world surrounding us involves multiple modalities – we see objects, hear sounds, feel texture, smell odors, and so on.
- In order for AI to make progress in understanding the world around us, it…
🔹All you need to know about Large Language Models (LLMs):
➡️ Embeddings
➡️ LLM Training Steps
➡️ Context Length Scaling
➡️ Traditional v/s Vector DBs
➡️ Knowledge-Augmenting LLMs (RAG, Prompting, Finetuning)
➡️ Popular LLMs
➡️ Leaderboards
➡️ Frameworks
📝 Announcing our
@ECIR2024
paper with IIT Patna:
🔹MedSumm: A Multimodal Approach to Summarizing Code-Mixed Hindi-English Clinical Queries
➡️
➡️ We introduce the novel task of Multimodal Medical Codemixed Question Summarization (+dataset,models)
@IITPAT
📝 Announcing our paper that offers a comprehensive overview (surveying 100+ papers) of the Indic AI landscape
➡️ 𝐓𝐚𝐱𝐨𝐧𝐨𝐦𝐲 𝐚𝐧𝐝 𝐎𝐯𝐞𝐫𝐯𝐢𝐞𝐰: We survey the current state-of-the-art in Indic AI by covering LLMs, Corpora, Benchmarks and Evaluation, Techniques, Tools,
📝 A Curated Set of NLP Primers: Attention, Autoregressive vs. Autoencoder Models, Transformers, BERT
🔹 Attention:
- Overview (The Attention Mechanism, The Bottleneck Problem, The Context Vector Bottleneck, How Attention Solves the Bottleneck Problem)
-
🔤 The A-to-Z of Word Embeddings •
• Word embeddings are a crucial aspect of computational linguistics in NLP, providing a means for computers to interpret and process human language via nuanced language representation.
• Grounded in distributional
I’m
#hiring
for both full-time and intern roles for Amazon Alexa AI. We're looking for exceptional AI scientists skilled at deep learning, who are eager to work on cutting-edge AI, and build the next generation of Amazon
#alexa
. If this is you, get in tou…
📝 All you need to know about Transformers, GPT, and BERT
🔹 Transformers:
- Mathematical background (Vectors, Matrix Multiplication, Dot Product, Masking, Sampling)
- Attention (Additive/Multiplicative/Dot Product Attention, Self…
I’m
#hiring
for both full-time and intern roles for Amazon Alexa AI. We're looking for exceptional AI scientists skilled at deep learning, folks who are eager to work on cutting-edge AI and build the next generation of Amazon
#alexa
. If this is you, get i…
📝 Announcing our new paper that reviews the impact of LLMs in Recommender Systems
🔹 "Exploring the Impact of Large Language Models on Recommender Systems: An Extensive Review"
🔹 In collaboration with
@SantaClaraUniv
and
@CarnegieMellon
➡️
#AI
#LLM
📝 Announcing our new paper proposing a new debiasing technique for LLMs
🔹 "From Prejudice to Parity: A New Approach to Debiasing Large Language Model Word Embeddings"
🔹 In collaboration with
@IITGuwahati
,
@UTAustin
, and
@UofIllinois
🔹 PDF:
#AI
#LLM
📚All the Math fundamentals for
#AI
(+Backprop primers)
➡️Math primer:
- Linear Algebra
- Differential Calculus
- Probability Theory & Distributions
➡️Backprop primer:
- Chain Rule
- Derivatives of Standard Layers/Loss Functions
📝 Top Papers in Computer Vision, NLP, Speech, Multimodal AI, Core ML, RecSys, & Graph ML
🔗
👉🏼 I’ve put together a summary of key papers in
#AI
and segregated them into (i) need-to-know and (ii) good-to-know.
🔹 Vision
- Image Classification (CNN
Our work at ACL 2023! Great to see our work piqued
@chrmanning
's interest :)
FACTIFY-5WQA: 5W Aspect-based Fact Verification through Question Answering
Paper:
#NLP
#NLProc
#ACL
#ACL2023NLP
Presented factify 5WQA, a small step towards building automatic and explainable fact verification system at ACL. Thanks
@chrmanning
for stopping by. It was great interacting with you.
cc-
@amit_p
P.S. - I am open for PhD positions for fall’24. DM to chat.
#NLProc
#ACL2023
📝 Recommender Systems in Production
🔹 Synopses and case studies of the inner-workings of some of the most popular recommendation system platforms:
- TikTok's Monolith Recommender System by Zhuoran Liu et al.
- Netflix's Recomme…
📝Announcing our new paper with
@NUSingapore
and
@Tsinghua_Uni
that explores how inducing personalities in LLMs affects their Theory-of-Mind reasoning
🔹"PHAnToM: Personality Has An Effect on Theory-of-Mind Reasoning in Large Language Models"
🔗
#AI
#LLMs
📝Announcing our new paper with
@UW
on Claim Verification using LLMs + Knowledge Graphs (KGs)
🔹Framework that verifies claims against a trusted KG with explanations & attribution
✅Allows for pinpointing inaccuracies like
#LLM
hallucinations
🔗
#GenAI
📚 Curated list of Books, Blogs, and YouTube Channels for AI and Data Science
- I’ve curated a list of books, blogs and YouTube channels based on what has been meaningful for my career growth over the past several years.
- These contain a broad range of domains in AI including
(2/2)🔹 EMNLP 2023 Paper: "The Troubling Emergence of Hallucination in Large Language Models - An Extensive Definition, Quantification, and Prescriptive Remediations"
➡️
#GPT
#LLMs
#hallucinate
#AI
📚 Curated list of Books, Blogs, Course Notes, Newsletters, and YouTube Channels for AI and Data Science
👉🏼 I’ve curated a list of books, blogs and YouTube channels based on what has contributed to my career growth over the past few years.
These contai…
📖 Encoder vs. Decoder vs. Encoder-Decoder Models •
- Among self-supervised representation learning objectives, encoder-based and decoder-based (i.e., autoregressive) language modeling have been the two most successful pretraining objectives. These two
👷🏽 Machine Learning Infrastructure •
- Infrastructure is the backbone of modern data-driven applications.
- Here's my primer that covers the essential components that make up a successful ML project, from data ingestion to model deployment, and the major
📖 The A-to-Z of Parameter Efficient Fine-Tuning (PEFT) •
- PEFT methods are crucial for foundation models such as LLMs because they allow for adapting these large models to specific tasks without needing to update all the parameters. This reduces
📖 The A-to-Z of Distributed Training Parallelism •
- Distributed training parallelism is crucial for efficiently training large-scale deep learning models that require extensive computational resources. This approach leverages multiple GPUs or machines
🤖 LLM Alignment Primer •
✅ RLHF, RLAIF, DPO, KTO, GPO, CPO, IPO, ICDPO, ORPO, sDPO, RS-DPO, Diffusion-DPO
- Reinforcement Learning with Human Feedback (RLHF), used to align LLM behavior with human preferences, has been pivotal for accurate and
📚 All the Math fundamentals for AI
🔹 Math primer:
- Linear Algebra & Differential Calculus
- Probability Theory
🔹 Backprop primer:
- Chain Rule
- Derivatives of Layers and Loss Functions
#ArtificialInteligence
📝 Top papers in Computer Vision, NLP, Speech, and Core ML
👉🏼
I’ve put together a summary of key papers in Computer Vision, NLP, and Speech and segregated them into 1️⃣ need-to-know, and 2️⃣ good-to-know.
What’s included (*not…
📚 All the Math fundamentals for AI
👉🏼 Don’t let the math behind AI concepts hinder you from understanding what goes on under-the-hood as you train your neural network. Think in terms of first principles to develop an intuition of what’s going on behin…
📝 Announcing our ACL 2024 (
@aclmeeting
) paper on MemeGuard, a novel framework that leverages LLMs and VLMs for generating interventions to counteract the toxicity in cyberbullying memes
➡️ 𝐍𝐨𝐯𝐞𝐥 𝐓𝐚𝐬𝐤 𝐚𝐧𝐝 𝐃𝐚𝐭𝐚𝐬𝐞𝐭: We introduce the task of meme intervention and
📝 Announcing our new paper at ICASSP 2023 next month!
🔹 Paper: “I See What You Hear: A Vision-inspired Method To Localize Words”
with Mohammad Samragh Razlighi, Arnav Kundu, Hu Ting-Yao, Minsik Cho, Ashish Shrivastava,
@OncelTuzel
, and Devang Naik
🔹…
📝 Tutorial: Hallucination in LLMs
✅ Our tutorial at
@LrecColing
2024 last week offered an introduction to the issue of hallucination, presented a taxonomy of categories, and covered the top hallucination detection and mitigation methods.
✅ We also covered the next big
📝 Announcing our paper that (i) demonstrates that SLMs can effectively compete with – and sometimes outperform – frontier LLMs such as GPT-4 when appropriately selected and prompted, and (ii) proposes a framework for selecting the best model and prompt style based on the
🧠 Curated set of NLP Primers: A one-stop shop!
👉🏼 Primers for Attention, Autoregressive vs. Autoencoder Models, Transformers, BERT, BigBird
🔹 Attention:
- Classic Sequence-to-Sequence Model
- Sequence-to-Sequence Model with A…
🤓 Primers for AI and Data Science: Python, PyTorch, TensorFlow, NumPy, Pandas, Matplotlib
👉🏼 Looking to kickstart a career in AI? Here are 6 primers covering some of the pillars of AI and Data Science: Python, PyTorch, TensorFlow, NumPy, Pandas, and M…
🤖 Recommender Systems (RecSys) Primer: The why (motivation), what (concepts under-the-hood), and how (code deep-dive)
👉🏼 TL;DR:
- Recommender models are an integral part of product offerings from every major company, with ofte…
🤖 Large Language Model (LLM) Primers | ChatGPT, Prompt Engineering, RLHF
With the advent of ChatGPT, LLMs have been the talk of the town! We’ve recently seen a bunch of extraordinary advancements in the NLP space including GPT-4, LLaMA, Toolformer, RLHF…