logo

Programme Starts: 15th February, 2025

Duration: 06 Months

Programme Overview

IIT Delhi's Certificate Programme in Generative AI offers a deep dive into advanced AI techniques, focusing on Large Language Models (LLMs) like GPT, BERT, and T5. Starting with foundational concepts like Linear Algebra and progressing to Machine Learning, participants will gain hands-on experience with model optimisation techniques such as fine-tuning and Parameter-Efficient Fine-Tuning (PEFT). The programme also covers cutting-edge topics like Reinforcement Learning with Human Feedback (RLHF) and Vision-Language Models (VLMs). Participants will be equipped to apply LLMs in real-world scenarios and ensure responsible AI use.

Programme Highlights

Practical learning with tutorials and latest tools

Advanced curriculum for cutting-edge AI expertise

6-month, online programme for working professionals

60 hours of online live sessions by IIT Delhi faculty and industry experts

Peer-learning and networking opportunities

IIT Delhi Continuing Education Programme (CEP) Certificate

Programme Content

Module 1: Mathematical Foundations for ML


  • Linear Algebra: Vector and Matrix Dot Product, Matrix Vector Multiplication, Matrix Decomposition (SVD)s

Learning Outcomes:

  • Building mathematical foundations, an essential prerequisite for the course
  • Probability Theory: Random Variables, Bayes Theorem, Conditional Probability

Learning Outcomes:

  • Understanding the fundamentals of Probability Theory in Machine Learning
  • Optimisation: Gradient Descent, First/Second Order Condition, Convex Optimisation, KL-Divergence

Learning Outcomes:

  • Concepts of Optimisation and its application in Machine Learning algorithms

Module 2: Machine Learning


  • Intro to ML: Linear Regression, Logistic Regression

Learning Outcomes:

  • Fundamental ML concepts, understanding regression methods
  • Optimisation Continued, SVM, Decision Tree, Ensemble Methods

Learning Outcomes:

  • Learning supervised ML methods
  • Unsupervised Learning: Clustering, Dimensionality Reduction (PCA, LDA, t-SNE)

Learning Outcomes:

  • Introduction to unsupervised methods
  • Artificial Neural Networks: Perceptron, Multilayer Network, Backpropagation

Learning Outcomes:

  • Understanding the basics of neural networks and their training

Read More >

Module 3: NLP


  • Basic Text Processing (NLTK, spaCy), Morphology, Stemming, Edit Distance

Learning Outcomes:

  • Use of NLP tools for text processing
  • Language Modelling: N-gram Modelling, Smoothing Techniques, Perplexity

Learning Outcomes:

  • Learning different language modelling approaches
  • POS Tagging: Sequential Learning, HMM, Viterbi Algorithm

Learning Outcomes:

  • Introduction to sequence learning for NLP
  • Parsing: Constituency vs Dependency Parsing, CKY Algorithm, CFG, PCFG

Learning Outcomes:

  • Learning techniques for syntactic analysis in NLP
  • Text Classification: Naive Bayes Algorithm, Lexical Similarity (Word Embeddings, TF-IDF, Word2Vec, GloVe)

Learning Outcomes:

  • Understanding classification in NLP using traditional and modern approaches

Module 4: Generative AI for Text


  • Neural Language Models (CNN, RNN, LSTM, GRU, Seq2Seq)

Learning Outcomes:

  • Introduction to neural models and attention mechanisms for text generation
  • Attention Mechanism: Self-Attention, Transformer Architecture

Learning Outcomes:

  • Understanding self-attention and Transformer architectures for language models
  • Pre-trained Models: BERT, GPT, T5

Learning Outcomes:

  • Architecture and training of different pre-trained large language models
  • Fine-tuning Strategies: Task-specific Fine-tuning, Instruction Fine-tuning, Preference Tuning (RLHF, PPO)

Learning Outcomes:

  • Learning various fine-tuning approaches to improve model performance
  • Prompting Strategies: In-context Learning, Chain of Thought, Knowledge Probing, Text Generation

Learning Outcomes:

  • Techniques for optimising LLM performance through effective prompting
  • Augmented LLMs: Retrieval-Augmented Generation (RAG), Tool Augmented LLM

Learning Outcomes:

  • Leveraging RAG and tool augmentation to improve LLM efficiency and reasoning capabilities

Module 5: Generative AI for Vision


  • Vision Language Models (VLM)

Learning Outcomes:

  • Understanding how VLMs enable combined text and image generation for multimodal applications

Module 6: Responsible AI


  • Misinformation, Bias, Toxicity, Security and Fairness

Learning Outcomes:

  • Learning strategies to mitigate bias, toxicity, and hallucinations in AI outputs

Assignments/Case Studies/Projects


  • Implement a fully connected neural network using PyTorch or TensorFlow for some task. Train the model and visualise its accuracy and loss over epochs.
  • Fine-tune a pre-trained transformer model on a text classification task. Evaluate the model on a test set and report its accuracy, precision, and recall.
  • Implement a simplified version of the transformer architecture focusing on self-attention and positional encoding. Train the model on a small dataset for translation or another task. Compare the results with a pre-trained transformer model.
  • Implement PEFT on a large LLM for a specific task. Compare the training time and resource usage between standard fine-tuning and PEFT.
  • Create a simple reward model using human-labelled data for a text generation task (e.g., summarisation or sentiment correction). Implement a reinforcement learning loop to improve the LLM's responses based on the reward model.
  • Compare the performance of different open-source Large Language Models (LLMs) on reasoning tasks using various prompting techniques—zero-shot, few-shot, chain-of-thought, self-consistency prompting.

Tutorials


  • Compute matrix operations using NumPy.
  • Preprocess a text dataset and generate embeddings using spaCy or Hugging Face.
  • Build a simple neural network using TensorFlow/PyTorch.
  • Train the model on basic NLP tasks like text classification problem.
  • Implement a simplified self-attention mechanism.
  • Use a pre-trained transformer models for various NLP tasks like translation or summarization.
  • Fine-tune a GPT or BERT model on a custom dataset for various NLP problems
  • Use Hugging Face's pre-trained transformer model to generate text.
  • Apply PEFT to a large model for fine-tuning on a small dataset.
  • Apply quantization and pruning on a pre-trained model to optimize for lower resource usage.
  • Implement a RAG model using Hugging Face transformers and document retrieval from a custom corpus.
  • Use Hugging Face multilingual models for problems in cross-lingual scenarios.

Read Less <

Tools Covered

CERTIFICATION

  • Candidates who score at least 50% marks overall and have a minimum attendance of 70%, will receive a 'Certificate of Completion'.
  • Candidates who score less than 40% marks overall and have a minimum attendance of 60%, will receive a 'Certificate of Participation'.
  • The organising department of this programme is the Department of Electrical Engineering, IIT Delhi.

Note: For more details download brochure.

ELIGIBILITY CRITERIA

  • Educational Background:
    - Graduates or Postgraduates in Science, Technology, Engineering or Mathematical Sciences
    - Participants with prior experience or knowledge of coding or programming will be preferred

Class Schedule

Sunday 09:00 AM to 12:00 PM (IST)

MEET OUR PROGRAMME Coordinator

Prof. Tanmoy Chakraborty
Rajiv Khemani Young Faculty Chair Professor in AI
Associate Professor, Dept. of Electrical Engineering
Associate Faculty Member, Yardi School of Artificial Intelligence
Indian Institute of Technology Delhi, New Delhi, India

Prof. Tanmoy is an Associate Professor of Electrical Engineering and the Yardi School of AI at the Indian Institute of Technology (IIT) Delhi. He leads the Laboratory for Computational Social Systems (LCS2), a research group specialising in Natural Language Processing (NLP) and Computational Social Science. His current research primarily focuses on empowering small language models for improved reasoning, grounding, and prompting and applying them specifically to two applications -- mental health counselling and Cyber-informatics.

Read More >

Tanmoy obtained his PhD in 2015 from IIT Kharagpur as a Google PhD scholar. Subsequently, he worked as a postdoctoral researcher at the University of Maryland, College Park, USA. Tanmoy has received numerous awards, including the Ramanujan Fellowship, the PAKDD Early Career Award, ACL'23 Outstanding Paper Award, IJCAI'23 AI for Good Award, and several faculty awards/gifts from companies like Facebook, Microsoft, Google, LinkedIn, JP Morgan, and Adobe. He has authored two textbooks – "Social Network Analysis" and “Introduction to Large Language Models.

Read Less <

MEET OUR PROGRAMME Faculty

Prof. Rachit Chhaya
Assistant Professor
Dhirubhai Ambani Institute of Information and Communication Technology (DA-IICT) Gandhinagar

Prof. Rachit is currently an Assistant Professor at Dhirubhai Ambani Institute of Information and Communication Technology (DA-IICT), Gandhinagar. He completed his PhD in Computer Science and Engineering at IIT Gandhinagar in 2022. His research focuses on scalable algorithms for machine learning problems with provable guarantees, specifically by creating small summaries of data called ‘coresets’. He has worked on machine learning problems involving regularisation and/or fairness constraints. He has published in prestigious venues like ICML, AAAI, AISTATS, and TMLR. Currently he teaches courses like machine learning and approximation algorithms. He has also been involved in various training programs on AI/ML.

Prof. Rahul Mishra
Assistant Professor
Language Technology Research Centre (LTRC),
International Institute of Information Technology Hyderabad

Prof. Rahul is an Assistant Professor at IIIT Hyderabad's Language Technology Research Centre (LTRC), where his research focuses on Controllable Text Summarisation, Misinformation Detection, Model Explainability, Graph Representation Learning, and Natural Language Generation. Previously, he served as a senior postdoctoral researcher at the University of Geneva, Switzerland, specialising in biomedical NLP. Prior to that, as a Senior Staff Engineer/Researcher, he contributed to research projects at Samsung Research Lab in Bangalore, optimising and benchmarking Large Language Models on Process in Memory (PIM) enabled GPUs. He holds a PhD from the University of Stavanger, Norway and an M.Tech from IIIT Delhi. During his doctoral studies, he also worked as a visiting researcher at the Computer Science Department of ETH, Zurich, Switzerland, and the University of Hannover, Germany. Before pursuing his PhD, he worked as an NLP data scientist in the automatic vehicle diagnostic department at KPIT Technologies, Pune, focusing on automatic fact extraction from car service manuals. Prior to that, he also held roles as a consultant researcher at Tata Research Development and Design Centre (TRDDC) and a research intern at IBM Research Bangalore.

Prof. Sourish Dasgupta
Assistant Professor
Dhirubhai Ambani Institute of Information and Communication Technology

Prof. Sourish is an ever-curious researcher and educator. His deep interest in the role of AI in research methodologies led him to take a break from academia and set on an entrepreneurial journey for five years, which resulted in the founding of RAx Labs Inc., Delaware, USA. With some bright ex-students of DA-IICT, Prof. Dasgupta built RAx (https://raxter.io) - an AI-powered online assistant for making literature-review faster and more enriching for young researchers. Prof. Dasgupta is currently actively engaged in the analysis of less explored but important aspects of "intelligence" in LLMs, such as their personalization capabilities, and also designing personalized models that are a lot smaller and more eco-friendly than contemporary LLMs. In his pastime, Prof. Dasgupta loves to cook and debate with students. Prof. Dasgupta did his Ph.D. in Computer Science from the University of Missouri – Kansas City, USA.

Programme Fees

₹1,69,000 + GST

(Installment available)

Admission Criteria

Admission to the program is based on a comprehensive review of your application.