0% found this document useful (0 votes)
13 views2 pages

Machine Learning

The document discusses the candidate elimination algorithm in machine learning, detailing the process of deriving specific and general hypotheses from training examples. It concludes with a final hypothesis and version space, indicating that new examples can be classified based on their attributes. Additionally, it explains the inductive bias of the algorithm, which favors simpler hypotheses and assumes concepts can be represented as conjunctions of attributes.

Uploaded by

Lovemore kabby
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
13 views2 pages

Machine Learning

The document discusses the candidate elimination algorithm in machine learning, detailing the process of deriving specific and general hypotheses from training examples. It concludes with a final hypothesis and version space, indicating that new examples can be classified based on their attributes. Additionally, it explains the inductive bias of the algorithm, which favors simpler hypotheses and assumes concepts can be represented as conjunctions of attributes.

Uploaded by

Lovemore kabby
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd

Yolanda Matengambiri

R216892J
HAI
Machine learning

15. Final version space using candidate elimination algorithm.

Initial Specific Hypothesis S0 = <null, null, null, null>

Initial General Hypothesis G0 = <{Japan, USA}, {1980, 1970, 1990}, {Economy, Sports},
{Honda, Toyota, Chrysler, null}>

Training Examples:
1. <Japan, ColorDecade1980, Economy, ManufacturerHonda> : S1 = <Japan, 1980,
Economy, Honda> and G1 = <{Japan, USA}, {1980}, {Economy, Sports}, {Honda,
Toyota, Chrysler, null}>
2. <Japan, ColorDecade1970, Sports, ManufacturerToyota> : S2 = <Japan, null, Sports,
null> and G2 = <{Japan, USA}, {1970, 1980, 1990}, {Economy, Sports}, {Toyota, null,
null, null}>
3. <Japan, ColorDecade1990, Economy, ManufacturerToyota> : S3 = <Japan, 1990,
Economy, null> and G3 = <{Japan, USA}, {1990}, {Economy, Sports}, {Toyota, null, null,
null}>
4. <USA, ColorDecade1980, Economy, ManufacturerChrysler> : G4 = <{Japan, USA},
{1980}, {Economy, Sports}, {null, null, Chrysler, null}>
5. <Japan, ColorDecade1980, Economy, ManufacturerHonda> : S5 = <Japan, 1980,
Economy, Honda> and G5 = <{Japan}, {1980}, {Economy}, {Honda}>
6. <Japan, ColorDecade1990, Economy, ManufacturerToyota> : S6 = <Japan, 1990,
Economy, null> and G6 = <{Japan}, {1990}, {Economy}, {Toyota}>
7. <Japan, ColorDecade1980, Sports, ManufacturerHonda> : G7 = <{Japan}, {1980},
{Sports}, {null}>

Final hypothesis after all examples: <Japan, 1980, Economy, Honda>


15b) Write the final version space for mentioned training examples using candidate
elimination algorithm.
[Link] bound on hypothesis space: H = {<Size, Colour, Shape>}
2. Initialize version space: S0 = {<, , >}
3. Initialize the set of all hypotheses: G0 = {<Big, , >, <, Red, Circle>, <, Blue, Circle>,
<Small, , >}
4. Iterate over each training example. For example 1 (Size=Big, Colour=No,
Shape=Circle, Class=No):
- Remove from G any hypothesis that is not consistent with the example - G1 =
{<Big, , >, <, Red, Circle>, <, Blue, Circle>, <Small, , >}
- Find the subset of hypotheses in S that are consistent with the example - S1 =
{<Big, , >}
5. Iterate over the remaining examples and update G and S for each example. After
all examples are processed, we get the following final version space:
- S = {<Big, , >, <, Red, Circle>}
- G = {<Big, , >, <, Red, Circle>}

This final version space suggests that for any new example, if it is a circle or if it is red,
it can be classified as "No". However, if it is not red or a circle and is big, it can also
be classified as "No".

16) Explain in detail the inductive Bias of candidate elimination algorithm.

The candidate elimination algorithm is a machine learning algorithm used for


learning concept descriptions from data. Inductive bias refers to a set of assumptions
or expectations that a learning algorithm makes about the data it will be working
with.
The candidate elimination algorithm is an example of a type of machine learning
algorithm called a lazy learning algorithm. This means that the algorithm does not
create a general model of the data, but instead makes decisions based on the
specific examples it encounters during training.

The inductive bias of the candidate elimination algorithm reflects the


assumptions it makes about the nature of the data it will be working with. The
algorithm assumes that the concept it is trying to learn can be represented as a
conjunction of attributes, and that there is a finite set of hypotheses that can be
used to represent the concept.
The algorithm also has a bias towards simpler hypotheses, meaning it will
choose the most specific and least general hypothesis to represent the concept. This
is known as the minimum description length principle.

You might also like