The document discusses the FP-Growth algorithm for frequent pattern mining. It improves upon the Apriori algorithm by not requiring candidate generation and only requiring two scans of the database. FP-Growth works by first building a compact FP-tree structure using two passes over the data, then extracting frequent itemsets directly from the FP-tree. An example is provided where an FP-tree is constructed from a sample transaction database and frequent patterns are generated from the tree. Advantages of FP-Growth include only needing two scans of data and faster runtime than Apriori, while a disadvantage is the FP-tree may not fit in memory.
FP GrowthStands for frequent pattern growth
It is a scalable technique for mining frequent pattern
in a database
3.
FP growthimproves Apriority to a big extent
Frequent Item set Mining is possible without
candidate generation
Only “two scan” to the database is needed
BUT HOW?
4.
Simply atwo step procedure
– Step 1: Build a compact data structure called the FP-tree
• Built using 2 passes over the data-set.
– Step 2: Extracts frequent item sets directly from the FP-
tree
5.
Now LetsConsider the following transaction table
TID List of item IDs
T100 I1,I2,I3
T200 I2,I4
T300 I2,I3
T400 I1,I2,I4
T500 I1,I3
T600 I2,I3
T700 I1,I3
T800 I1,I2,I3,I5
T900 I1,I2,I3
6.
Now wewill build a FP tree of that database
Item sets are considered in order of their descending
value of support count.
I2 7
I1 6
I36
I4 2
I5 2
null
I2:7
I1:4
I3:2
I3:2
I5:1
I4:1
I4:1
To facilitate tree traversal, an
item header table is built so
that each item points to its
occurrences in the tree via a
chain of node-links.
I1:2
I3:2
I5:1
FP Tree Construction Over!!
Now we need to find conditional pattern base
and Conditional FP Tree for each item