Impact Factor:6.549
 Scopus Suggested Journal: UNDER REVIEW for TITLE INCLUSSION

International Journal
of Computer Engineering in Research Trends (IJCERT)

Scholarly, Peer-Reviewed, Open Access and Multidisciplinary


Welcome to IJCERT

International Journal of Computer Engineering in Research Trends. Scholarly, Peer-Reviewed,Open Access and Multidisciplinary

ISSN(Online):2349-7084                 Submit Paper    Check Paper Status    Conference Proposal

Back to Current Issues

Review of Rule Quality Measurement: Metrics and Rule Evaluation Models

Munirah Muslim, E. Winarko, , ,
Affiliations
Computer Science Department, Gadjah Mada University of Yogyakarta, Indonesia
:10.22362/ijcert/2018/v5/i1/v5i102


Abstract
A rule-based system is a system based on the set of rules used to make inference knowledge. The system gathers knowledge into the representation of knowledge in the form of a rule. However, the knowledge in the form of the rule is inductive, meaning that the algorithm can construct the rule by studying a limited number of cases and then the induced rule of a limited number of cases and then generalize it to the general reality from time to time. This, of course, has the degree of inaccuracy in expressing reality into knowledge, or an experienced expert builds it but it is not absolute that the knowledge it possesses is 100% accurate or always consistently true from one time-space location to another time-space location. Therefore, the need for a formula that can measure the quality of the resulting rule and assess the consistency of the rule. In this study, we did a review of the ideas of people trying to measure knowledge built inductively by either the algorithm or the experts. These measurements are based on several parameters defined by them according to the underlying assumptions. This review seeks to partially present how ideas to measure the rule as knowledge representation from a varied viewpoint and how people construct evaluation models to assess the resulting regulations either from the experts or human experts as well as those resulting from the induction rule algorithm much developed.


Citation
Munirah Muslim,E. Winarko(2018). Review of Rule Quality Measurement: Metrics and Rule Evaluation Models. International Journal of Computer Engineering In Research Trends, 5(1), 4-11. Retrieved from http://ijcert.org/ems/ijcert_papers/V5I102.pdf


Keywords : Review, Quality, Measurement, Metric, Rule, Evaluation, Model.

References
[1] P. Smyth and R. M. Goodman, “An Information Theoretic Approach to Rule Induction from Databases,” vol. 4, no. 4, pp. 301–316, 1992.
[2] R. Shinghal, “Evaluating the Interestingness of Characteristic Rules,” pp. 263–266, 1996.
[3] S. Dreiseitl, M. Oil, C. Baumgartner, and S. Viterbo, "An evaluation of heuristics for rule ranking," Artif. Intell. Med., vol. 50, no. 3, pp. 175–180, 2010.
[4] J. F. Roddick and M. Spiliopoulou, “A survey of temporal knowledge discovery paradigms and methods,” IEEE Trans. Knowl. Data Eng., vol. 14, no. 4, pp. 750–767, 2002.
[5] F. Provost, “A Survey of Methods for Scaling Up Inductive Algorithms,” vol. 169, pp. 131–169, 1999.
[6] I. Inza, P. Larrañaga, R. Etxeberria, and B. Sierra, “Feature Subset Selection by Bayesian network-based optimization,” vol. 123, pp. 157–184, 2000.
[7] F. Provost, “Tree Induction for Probability-Based Ranking,” vol. 5, pp. 199–215, 2003.
[8] J. Sulzmann and F. Johannes, “An Empirical Comparison of Probability Estimation Techniques for Probabilistic Rules,” no. 2003, pp. 317–331, 2009.
[9] A. A. Freitas, “On rule interestingness measures,” Knowledge-Based Syst., vol. 12, no. March, pp. 309–315, 1999.
[10] K. E. N. Mcgarry, “A survey of interestingness measures for knowledge discovery,” pp. 39–61, 2005.
[11] X. Huynh, F. Guillet, J. Blanchard, and P. Kuntz, “A Graph-based Clustering Approach to Evaluate Interestingness Measures : A Tool and a Comparative Study,” vol. 50, pp. 25–50, 2007.
[12] B. Vaillant, S. Lallich, and P. Lenca, “On the behavior of the generalizations of the intensity of implication : A data-driven comparative study,” vol. 447, pp. 421–447, 2008.
[13] J. Hills, L. M. Davis, and A. Bagnall, “Interestingness Measures for Fixed Consequent Rules,” pp. 68–75, 2012.
[14] P. Flach, N. Lavrac, and B. Zupan, “Rule Evaluation Measures: A Unifying View,” Proc. 9th Int. Work. Inductive Log. Program., pp. 174–185, 1999.
[15] F. Johannes and P. A. Flach, “An Analysis of Rule Evaluation Metrics,” 2003.
[16] D. Christensen, “David Christensen - Measuring Confirmation.pdf.” pp. 437–461, 1999.
[17] S. Greco, R. Słowi, and I. Szcz, “Measures of rule interestingness in various perspectives of confirmation,” vol. 347, pp. 216–235, 2016.
[18] M. Michalak, M. Sikora, and Ł. Wróbel, “Rule Quality Measures Settings in a Sequential Covering Rule Induction Algorithm - an Empirical Approach,” vol. 5, pp. 109–118, 2015.
[19] P. F. Nada Lavrac, Bojan Cestnik, Dragan Gamberger, “Decision Support Through Subgroup Discovery : Three Case Studies and the Lessons Learned,” no. 1994, pp. 115–143, 2004.
[20] D. M. W. Powers, “ROC-ConCert,” pp. 12–15, 2012.
[21] P. Salgado, “Relevance as a new measure of relative importance: of sets of rules,” no. 3, pp. 3770–3777, 2000.
[22] [22] F. Coenen and P. Leng, “An Evaluation of Approaches to Classification Rule Selection,” IEEE Int. Conf. Data Min., pp. 2–5, 2004.
[23] Y. Yao and B. Zhou, “Micro and Macro Evaluation of Classification Rules,” Proc. Seventh IEEE Int. Conf. Cogn. Informatics, ICCI 2008, Stanford Univ. California, USA, 2008.
[24] D. M. W. Powers, “Evaluation: From Precision, Recall And F-Measure To Roc, Informedness, Markedness & Correlation,” vol. 2, no. 1, pp. 37–63, 2011.
[25] H. Abe, S. Tsumoto, M. Ohsaki, and T. Yamaguchi, “Evaluating Learning Algorithms to Construct Rule Evaluation Models Based on Objective Rule Evaluation Indices,” 2007.
[26] H. Abe, S. Tsumoto, M. Ohsaki, and T. Yamaguchi, “Evaluating Learning Algorithms to Support Human Rule Evaluation with
Predicting Interestingness Based on Objective Rule Evaluation Indices,” vol. 282, no. 2008, pp. 269–282, 2008.
[27] H. Abe and S. Tsumoto, “Comparing Accuracies of Rule Evaluation Models to Determine Human Criteria on Evaluated Rule Sets,” pp. 1–7, 2008.
[28] H. Abe and S. Tsumoto, “Rule Evaluation Model as Behavioral Modeling,” pp. 8–15, 2009.
[29] A. Gruca and M. Sikora, "Rule-based functional description of genes – Estimation of the multicriteria rule interestingness measure by the UTA method," Integr. Med. Res., vol. 33, no. 4, pp. 222–234, 2013.
[30] A. Gruca and M. Sikora, “Data- and expert-driven rule induction and filtering framework for functional interpretation and description of gene sets,” pp. 1–14, 2017.
[31] U. Stanczyk, “Weighting and Pruning of Decision Rules,” pp. 106–114, 2016.
[32] K. K. Sethi, D. K. Mishra, and B. Mishra, “Novel Algorithm to Measure Consistency between Extracted Models from Big Dataset and Predicting Applicability of Rule Extraction,” IEEE Trans. Knowl. Data Eng., 2014.
[33] H. Mutluri and P. Sujatha, “Challenges in Big Data using Data Mining Techniques,” Int. J. Comput. Eng. Res. Trends, vol. 2, no. 12, pp. 924–930, 2015.



DOI Link : http://dx.doi.org/10.22362/ijcert/2018/v5/i1/v5i102

Download :
  V5I102.pdf


Refbacks : Currently there are no refbacks

Announcements


Authors are not required to pay any article-processing charges (APC) for their article to be published open access in Journal IJCERT. No charge is involved in any stage of the publication process, from administrating peer review to copy editing and hosting the final article on dedicated servers. This is free for all authors. 

News & Events


Latest issue :Volume 10 Issue 1 Articles In press

A plagiarism check will be implemented for all the articles using world-renowned software. Turnitin.


Digital Object Identifier will be assigned for all the articles being published in the Journal from September 2016 issue, i.e. Volume 3, Issue 9, 2016.


IJCERT is a member of the prestigious.Each of the IJCERT articles has its unique DOI reference.
DOI Prefix : 10.22362/ijcert


IJCERT is member of The Publishers International Linking Association, Inc. (“PILA”)


Emerging Sources Citation Index (in process)


IJCERT title is under evaluation by Scopus.


Key Dates


☞   INVITING SUBMISSIONS FOR THE NEXT ISSUE :
☞   LAST DATE OF SUBMISSION : 31st March 2023
☞  SUBMISSION TO FIRST DECISION :
In 7 Days
☞  FINAL DECISION :
IN 3 WEEKS FROM THE DAY OF SUBMISSION

Important Announcements


All the authors, conference coordinators, conveners, and guest editors kindly check their articles' originality before submitting them to IJCERT. If any material is found to be duplicate submission or sent to other journals when the content is in the process with IJCERT, fabricated data, cut and paste (plagiarized), at any stage of processing of material, IJCERT is bound to take the following actions.
1. Rejection of the article.
2. The author will be blocked for future communication with IJCERT if duplicate articles are submitted.
3. A letter regarding this will be posted to the Principal/Director of the Institution where the study was conducted.
4. A List of blacklisted authors will be shared among the Chief Editors of other prestigious Journals
We have been screening articles for plagiarism with a world-renowned tool: Turnitin However, it is only rejected if found plagiarized. This more stern action is being taken because of the illegal behavior of a handful of authors who have been involved in ethical misconduct. The Screening and making a decision on such articles costs colossal time and resources for the journal. It directly delays the process of genuine materials.

Citation Index


Citations Indices All
Citations 1026
h-index 14
i10-index 20
Source: Google Scholar

Acceptance Rate (By Year)


Acceptance Rate (By Year)
Year Rate
2021 10.8%
2020 13.6%
2019 15.9%
2018 14.5%
2017 16.6%
2016 15.8%
2015 18.2%
2014 20.6%

Important Links



Conference Proposal




DOI:10.22362/ijcert