Module 1 Notes
Module 1 Notes
These notes are strictly prepared in alignment with the textbook “Artificial Intelligence: Beyond
Classical AI” by Reema Thareja and the prescribed syllabus.
You must first refer to the textbook for every topic.
If you still find any concept difficult to understand, then use these notes as your supporting
reference.
If the topic remains unclear even after that, please come to me directly for further explanation.
MODULE 1 – INTRODUCTION TO
ARTIFICIAL INTELLIGENCE
1.1 What is Artificial Intelligence
Artificial Intelligence (AI) is the scientific and engineering discipline dedicated to building
intelligent machines capable of performing tasks that normally require human reasoning,
perception, learning, and problem-solving. It focuses on designing systems that can perceive
their environment, interpret situations, learn from experiences, and act towards achieving
specific goals with minimal human intervention.
The fundamental idea behind AI is that human thought processes can be replicated by
machines using computational models. AI integrates concepts from computer science,
cognitive psychology, neuroscience, linguistics, mathematics, and engineering. Hence, it is
not limited to algorithms alone but represents a multi-disciplinary attempt to simulate human
intelligence in a digital environment.
AI has evolved to encompass both theoretical and applied domains. The theoretical
dimension involves developing logical and mathematical models of reasoning and learning,
while the applied side deals with implementing these models in real-world systems such as
robots, chatbots, autonomous vehicles, and decision-support systems.
Thus, AI strives to enable computers to think, learn, and act intelligently—not merely to
execute instructions, but to adapt and evolve with changing circumstances.
AI systems rely on large quantities of data. The data may be structured (tables, databases) or
unstructured (text, audio, video). Data cleaning, normalization, and transformation are crucial
because poor data quality leads to unreliable models.
For instance, a voice assistant such as Alexa is trained using millions of speech samples,
covering different accents and tones, to ensure it understands diverse speakers.
At this stage, the AI system uses machine learning algorithms to identify patterns within
the data. The algorithms iteratively adjust their internal parameters to minimize prediction
errors. This training process enables the AI to “learn” relationships and make informed
decisions in new situations.
A hallmark of AI is its ability to improve with feedback. New data constantly refines the
existing model, leading to better performance over time. This property distinguishes AI from
traditional programming, which remains static once coded.
Hence, AI systems function as adaptive entities: they sense their surroundings, interpret data,
learn from patterns, and respond intelligently.
Disadvantages
The conceptual seeds of AI were planted when scientists like Alan Turing proposed that
machines could simulate any human computation. His Turing Test (1950) remains a
fundamental benchmark: if a machine can converse indistinguishably from a human, it may
be considered intelligent.
Early experiments in symbolic logic and mechanical reasoning inspired computer scientists to
believe that machines could replicate aspects of human thought.
The term Artificial Intelligence was formally introduced at the Dartmouth Conference,
organized by John McCarthy, Marvin Minsky, Allen Newell, and Herbert Simon. This
event marked AI’s recognition as an independent field of study.
However, due to limited computing power and data, AI progress was slow. This period also
saw the first wave of AI Winter—a time when interest and funding declined due to unrealistic
expectations.
The 1980s witnessed the emergence of expert systems, which encoded human expertise into
rule-based frameworks. MYCIN, for example, assisted doctors in diagnosing bacterial
infections. Such systems demonstrated practical applications of AI in medicine, business, and
engineering.
With the rise of machine learning, AI moved from symbolic reasoning to data-driven
approaches. Algorithms could now learn automatically from examples. IBM’s Deep Blue
defeating Garry Kasparov in 1997 symbolized the power of computational intelligence.
The combination of massive datasets, cloud computing, and neural networks triggered
explosive AI growth. AI systems today power voice assistants, image recognition,
autonomous vehicles, and generative tools like ChatGPT and DALL·E. The focus has shifted
toward explainability, ethics, and human-AI collaboration.
Weak AI is designed for a specific function. It excels in the particular domain for which it
was programmed but cannot adapt beyond it. For instance, Google Translate cannot play
chess, and a chess AI cannot diagnose diseases.
It operates under fixed parameters without consciousness or understanding.
Examples: Voice assistants (Siri, Alexa), spam filters, and recommendation engines.
These are the most fundamental AI systems that respond only to current inputs. They have no
memory and cannot use past experiences.
Example: IBM’s Deep Blue could evaluate thousands of chess positions per second but had
no concept of learning or improvement.
This category represents most AI systems today. They can analyze past data to improve
performance. For example, self-driving cars use sensory data from previous trips to make
driving decisions.
Still in the research phase, Theory of Mind AI would possess the ability to understand human
emotions, beliefs, and intentions. This advancement would allow AI to engage in genuinely
empathetic interactions.
1.3.6 Self-Aware AI
Thus, while Artificial Intelligence aspires to replace human effort, Augmented Intelligence
seeks to enhance it, and Cognitive Computing attempts to simulate it.
1. Supervised Learning: The algorithm learns from labeled data. (e.g., spam detection)
2. Unsupervised Learning: The system identifies patterns in unlabeled data. (e.g.,
clustering customers)
3. Reinforcement Learning: The system learns by interacting with the environment and
receiving feedback. (e.g., game-playing AI)
Deep Learning (DL), a subset of ML, utilizes Artificial Neural Networks (ANNs) with
multiple layers to model complex non-linear relationships. DL systems mimic the structure of
the human brain and are capable of understanding images, audio, and natural language.
Applications include:
Deep Learning represents the core of today’s AI revolution, transforming how machines
interpret and engage with the world.
Reema Thareja defines intelligence as “the ability of an entity to perceive its environment,
process information, and act in a goal-directed manner.”
Hence, intelligence is not merely problem solving; it involves perception, reasoning,
learning, memory, and adaptation.
Human intelligence arises from biological neural activity; machine intelligence results from
algorithmic computation.
The goal of AI is not to copy the human brain physically but to replicate the functions that
make human thinking effective.
1. Perception – The process of collecting data from the environment through sensory
mechanisms (eyes, ears, sensors, cameras, microphones).
Machines perceive through sensors; for instance, a robot uses ultrasonic sensors to
detect obstacles.
2. Learning – The acquisition of knowledge or skills through study, experience, or
teaching.
In humans, this occurs via repetition and memory; in machines, through algorithms
such as supervised and unsupervised learning.
Learning forms the foundation of adaptability.
3. Reasoning – The ability to derive conclusions, inferences, or judgments from existing
knowledge.
Machine reasoning often follows logical inference rules such as modus ponens (if A
→ B and A is true, then B is true).
Example: An expert system diagnosing diseases reasons through symptom–disease
relationships.
4. Problem Solving – The act of determining a sequence of actions that transforms an
initial state into a desired goal state.
AI uses algorithms (search, constraint satisfaction, optimization) to achieve this
transformation efficiently.
5. Memory – Storage and retrieval of information.
Machines maintain memory in structured databases or knowledge bases.
Effective memory enables re-use of previous experiences.
6. Adaptation – Adjustment of behaviour based on changes in the environment.
Adaptive systems recalibrate parameters when they detect new conditions; e.g.,
thermostats adjusting temperature automatically.
7. Language Understanding and Communication – The ability to comprehend and
express information symbolically.
Natural-Language Processing (NLP) allows machines to interpret human language for
dialogue systems, translation, or summarization.
While machines excel in speed, accuracy, and data volume, humans surpass in contextual
understanding, empathy, and moral reasoning.
The most effective systems therefore combine both — an approach often termed Augmented
Intelligence.
An agent is an entity that perceives its environment through sensors and acts upon that
environment through actuators to achieve specific objectives.
Formally, an agent implements an agent function that maps a sequence of percepts (inputs)
to an action:
where P* is the set of all percept sequences and A is the set of possible actions.
Example:
A self-driving car (agent) perceives its surroundings through cameras, radar, and LiDAR
(sensors) and acts through steering, acceleration, and braking (actuators).
3.4.2 Environment
1. Simple Reflex Agents: Act solely on the current percept; implement condition-action
rules (“if traffic light is red → stop”).
No memory of past states.
2. Model-Based Reflex Agents: Maintain an internal model representing aspects of the
world not currently observed, enabling reasoning in partially observable
environments.
3. Goal-Based Agents: Incorporate explicit goals and choose actions that bring them
closer to those goals, using search or planning algorithms.
4. Utility-Based Agents: Beyond goals, they evaluate outcomes in terms of a utility
function measuring performance or satisfaction, selecting actions that maximize
expected utility.
A state space is the set of all possible configurations reachable from the initial state.
A solution is a path through this state space satisfying the goal condition.
1. Defining the Problem: initial state, goal state, actions, transition model, and cost.
2. Searching the Space: exploring possible sequences systematically.
3. Evaluating Paths: selecting the most efficient route based on cost or heuristic.
4. Executing the Plan: applying the derived sequence of actions.
• Expands the node with the lowest cumulative path cost g(n).
• Uses a priority queue ordered by path cost.
• Finds the optimal (least-cost) solution for positive costs.
Example: Dijkstra’s shortest-path algorithm.
A heuristic function h(n) estimates the cost from node n to the goal.
The smaller the estimated cost, the more promising the node.
Also called Greedy Search, it selects the node with the minimum heuristic h(n)**, ignoring
the path cost already incurred.
It “looks ahead” optimistically, assuming the best-appearing node leads directly to the goal.
Advantages:
The algorithm repeatedly selects the node with the lowest f(n).
Implementation uses a priority queue ordered by h(n).
Example:
Suppose a map-based agent must reach city G from city A.
Each node represents a city; h(n) = straight-line distance to G.
The algorithm always chooses the city appearing geographically closest to G until the
destination is reached.
While efficient for small problems, it can become sub-optimal when heuristic estimates are
inaccurate.
4 KNOWLEDGE REPRESENTATION
(KR)
4.1 Introduction
Knowledge Representation is the method of encoding information about the world in a
format that a computer can utilize to solve complex tasks such as reasoning,
understanding, and decision making.
Architecture:
Operation Cycle:
Example: In a medical-diagnosis system, symptoms are told to the KB, it infers possible
diseases, and suggests appropriate actions.
1. Declarative Knowledge:
Facts and assertions about objects or relationships.
Example – “Paris is the capital of France.”
2. Procedural Knowledge:
Describes how to perform actions or procedures.
Example – “Steps for solving a linear equation.”
3. Heuristic Knowledge:
Rule-of-thumb strategies based on experience rather than strict logic.
Example – “In chess, control the center of the board early.”
4. Meta-Knowledge:
Knowledge about the use and structure of other knowledge — knowing how to reason
with knowledge.
Used in expert systems for adaptive reasoning.
5. Structural Knowledge:
Describes the relationships among concepts, forming semantic networks or frames.
Example – hierarchies in ontology: Animal → Mammal → Dog.