0% found this document useful (0 votes)
55 views9 pages

AI - ML Resource Sheet

Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
55 views9 pages

AI - ML Resource Sheet

Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
You are on page 1/ 9

Introduction to the Mastering Fundamentals Of AI/ML

through Bottom-Up Approach.


The Bottom-Up Approach is perfect for those who want to understand how AI works
under the hood. It starts from the basics of math, programming, and algorithms, then
gradually builds up to machine learning, deep learning, and AI model development.

This path is ideal for:

● Curious students

● Researchers

● Developers who want to build their own AI

● People who enjoy understanding logic and theory

Goal: By the end of this journey, you’ll be able to understand, build, and train your own AI
models.

Step 1: Math for AI/ML


Why learn this? Math is the muscle behind AI — especially for model training and
understanding how learning happens.

Core Concepts:

● Linear Algebra: Vectors, matrices, dot product, matrix multiplication

● Probability & Statistics: Mean, variance, distributions, Bayes’ Theorem

● Calculus: Derivatives, gradients (backpropagation in neural networks)

Resources:

● Khan Academy – Linear Algebra

● Linear Algebra by FreeCodeCamp

● StatQuest with Josh Starmer (YouTube)


● Calculus by FreeCodeCamp
● Probability must knows for AL/ML
● Probability and Probability Distributions for Machine Learning

Benefit:
● You'll understand how a model adjusts itself (gradient descent).

● Crucial for understanding inner workings of algorithms.

Step 2: Programming Foundations


Why learn this? You need a language to talk to AI. Python is the most widely used for AI
and data science.

Topics to Learn:

● Python syntax, variables, data types

● Loops, conditionals, functions

● Lists, dictionaries, sets, tuples

● Object-oriented programming (OOP)

Resources:

● W3Schools Python Tutorial

● freeCodeCamp Python Course (YouTube)


● Python Basics – RealPython

Tools to use:

● Jupyter Notebook

● Google Colab (free cloud-based coding)

● VS Code

Step 3: Data Handling and Visualization


Why learn this? AI is nothing without data. You need to know how to manipulate, clean,
and visualize data before training any model.

Topics to Learn:

● NumPy: Arrays, matrices, broadcasting, vectorized operations

● Pandas: DataFrames, filtering, grouping, reading CSV/Excel

● Matplotlib & Seaborn: Visualize trends, correlations, charts

Resources:
● Kaggle Pandas Course

● NumPy Tutorial – DataCamp

● Matplotlib/Seaborn – Official Docs + YouTube series


● DataAnalysis and Visualisation

Outcome: You’ll know how to load real-world datasets, clean them, and visualize key
insights.

Step 4: Machine Learning (ML)


Why learn this? This is where the AI journey truly begins. You’ll teach machines to learn
from data and make predictions.

Core Topics:

● Supervised Learning: Linear Regression, Logistic Regression, Decision Trees, SVM

● Unsupervised Learning: K-Means, Hierarchical Clustering, PCA

● Model evaluation: Accuracy, precision, recall, F1 score

Resources:

● Google’s Machine Learning Crash Course

● DataCamp ML Fundamentals with Python

● Kaggle Courses – Intro to ML

Tools:

● Scikit-learn

● Google Colab

● Real datasets from UCI or Kaggle

Projects to Try:

● Predict house prices (regression)

● Classify emails as spam (classification)

● Customer segmentation (clustering)


Step 5: Deep Learning (DL)
Why learn this? Deep learning is what powers modern AI — including ChatGPT, image
recognition, and voice assistants.

Core Concepts:

● Neural Networks: Perceptron, layers, weights, biases

● Activation Functions: ReLU, Sigmoid, Tanh

● Loss Functions: MSE, Cross-Entropy

● Optimizers: SGD, Adam

● Advanced Models:

○ CNN (Image tasks)

○ RNN/LSTM (Text, sequence)

○ Autoencoders (Compression)

○ Transformers (Language Models)

Tools & Frameworks:

● TensorFlow

● PyTorch

● Keras (simplified deep learning)

Resources:

● DeepLearning.AI’s DL Specialization (Coursera)

● CS231n (Stanford) – CNNs for Visual Recognition


● PyTorch YouTube Tutorials

Projects:

● Image classifier using CNN

● Sentiment analysis using RNN

● Build your own Chatbot with Seq2Seq

Absolutely! Here's a step-by-step guide for Step 6: Projects & Applications with what
projects to build, what you’ll learn from each, and why it’s important. This is designed
like a practical roadmap so that you move from basic to advanced, and extract maximum
learning from every project. Let's go!

Step 6: Projects & Applications


Why this step is important:
After learning theory, coding, and math—this is where your AI knowledge becomes real.
Projects help you:

● Solidify concepts

● Solve real-world problems

● Build your portfolio

● Prepare for internships or job roles

Project 1: Titanic Survival Prediction (Basic ML –


Classification)
Objective: Predict if a passenger survived the Titanic based on features like age, gender,
ticket class, etc.
Dataset: Kaggle Titanic Dataset
Tools: Pandas, Scikit-learn, Matplotlib

What you will learn:


● Data cleaning (handle null values, encode categorical data)

● Feature engineering (create new columns)

● Classification models (Logistic Regression, Decision Tree)

● Evaluation metrics (Accuracy, Confusion Matrix)

Why important: It's a great entry point. You get comfortable with end-to-end ML flow —
from data preprocessing to prediction.

Youtube Tutorial: https://youtu.be/fATVVQfFyU0?si=yNipprgrhoLj7Y9H

Github link: https://github.com/Ruban2205/titanic-classification

Project 2: Handwritten Digit Recognition (Basic DL –


Computer Vision)
Objective: Classify handwritten digits (0–9) using image data.
Dataset: [MNIST Dataset – built into TensorFlow/Keras or from Kaggle]
Tools: TensorFlow/Keras or PyTorch, Matplotlib

What you will learn:


● How image data is represented (28x28 grayscale matrix)

● Neural network architecture (input, hidden, output)

● Activation functions (ReLU, Softmax)

● Loss functions (Categorical Crossentropy)

● Accuracy improvements with CNNs

Why important: You understand how to handle image data, and you build your first neural
network for vision tasks.

Youtube Tutorial: https://youtube.com/playlist?list=PLiWNvnK7PSPE--


36RIdeHg8Sgg02w9chE&si=Bmd6cbwZWSuQqigN

Github link: https://github.com/aakashjhawar/handwritten-digit-recognition

Project 3: Movie Recommendation System


(Intermediate – Recommender Systems)
Objective: Recommend movies to users based on their preferences or viewing history.
Dataset: MovieLens Dataset
Tools: Pandas, Scikit-learn, Surprise, or TensorFlow Recommenders

What you will learn:


● Collaborative filtering

● Cosine similarity, matrix factorization

● User-item rating matrix

● Evaluation using RMSE, Precision@K

Why important: You’ll learn how platforms like Netflix or Spotify suggest content using
machine learning.

Youtube Tutorial: https://youtu.be/kuC38ZCcbZI?si=6SrO1XcHcTlwUeqC

Github Link: https://github.com/NJ1219/Project-3-Movie-Recommender-System

Project 4: Stock Price Prediction (Intermediate – Time


Series Forecasting)
Objective: Predict future stock prices based on historical data.
Dataset: [Yahoo Finance API using yfinance]
Tools: Pandas, NumPy, LSTM with Keras or PyTorch

What you will learn:


● Time-series data preparation (windowing, lag features)

● Recurrent Neural Networks (RNNs), LSTMs

● Overfitting and regularization in temporal models

● Visualizing trends and confidence intervals

Why important: You learn how to work with sequential data, crucial for many real-world
tasks like weather forecasting, sales prediction, etc.

Youtube Tutorial: https://youtu.be/s3CnE2tqQdo?si=qTT90-sEIVF9OExb

Github Link: https://github.com/Vatshayan/Final-Year-Machine-Learning-Stock-Price-


Prediction-Project

Project 5: Sentiment Analysis on Tweets (Intermediate


– NLP)
Objective: Analyze tweets and classify them as positive, negative, or neutral.
Dataset: Twitter API (via Tweepy) or open sentiment datasets from Kaggle
Tools: NLTK / SpaCy / Hugging Face Transformers

What you will learn:


● Tokenization, Stopword removal, Lemmatization

● Text vectorization (TF-IDF, Word2Vec)

● Basic NLP pipelines

● Fine-tuning pre-trained models (like BERT)

Why important: You'll gain real skills in Natural Language Processing, one of the most in-
demand AI fields.

Youtube Tutorial: https://youtu.be/4YGkfAd2iXM?si=lJkSPyWah2ER97L3

Github Link: https://github.com/chilas/tweet-analyza

Project 6: Build Your Own Chatbot (Advanced – End-to-


End AI)
Objective: Build a chatbot that can respond to questions in a domain (e.g., FAQ bot).
Dataset: Custom intent-response dataset or use a QA dataset
Tools: TensorFlow, Flask, DialogFlow, Rasa, LangChain (optional)

What you will learn:


● NLP + DL integration

● Building conversation flow

● Intent recognition

● Deploying models as APIs

Why important: You combine everything learned—ML, DL, APIs, deployment—to build a
real AI product.

Youtube Tutorial: https://youtu.be/2e5pQqBvGco?si=vCPnikXfhu-P5Tn3

Github Link: https://github.com/vnk8071/E2E-AI-Chatbot

Platforms to Practice

Platform Use

Kaggle Datasets + Competitions + Practice

Google Colab Free GPU for AI project coding

Hugging Face Datasets Latest NLP datasets

PapersWithCode See how SOTA models are built

Final Notes:
● Always start simple, then improve the model

● Document every project (README, Jupyter Notebook)

● Add your projects to GitHub + create a portfolio


● Don’t chase perfection — chase progress

Bonus: Community & Practice Platform

Practice Daily:
● Kaggle

● HackerRank – AI/ML Domain

● LeetCode – SQL, Python, ML challenges

Conclusion
By completing this Bottom-Up journey, you will:

● Understand how AI learns

● Know the math, code, and models behind AI

● Be able to build your own AI projects

● Be ready to even build or fine-tune your own models

From here, you can move into LLMs, LangChain, RAG, and AI agents with confidence —
knowing the core concepts solidly.

Bottom-Up = Deep Understanding + Control. You’re not just a user of AI — you become a
builder, a creator.
Now, the world of AI is in your hands.

You might also like