[Typetext][Typetext][Typetext]2014TradeScienceInc.
ISSN:0974-7435Volume10Issue24BioTechnologyAnIndianJournalFULLPAPERBTAIJ,10(24),2014[16338-16346]ApplicationresearchofdecisiontreealgorithminenglishgradeanalysisZhaoKunBeihuaUniversity,Teacher'scollege,Jilin,(CHINA)ABSTRACTThispaperintroducesandanalysesthedatamininginthemanagementofstudents'grades.
Weusethedecisiontreeinanalysisofgradesandinvestigateattributeselectionmeasureincludingdatacleaning.
WetakecoursescoreofinstituteofEnglishlanguageforexampleandproducedecisiontreeusingID3algorithmwhichgivesthedetailedcalculationprocess.
Becausetheoriginalalgorithmlacksterminationcondition,weproposeanimprovedalgorithmwhichcanhelpustofindthelatencyfactorwhichimpactsthegrades.
KEYWORDSDecisiontreealgorithm;Englishgradeanalysis;ID3algorithm;Classification.
BTAIJ,10(24)2014ZhaoKun16339INTRODUCTIONWiththerapiddevelopmentofhighereducation,EnglishgradeanalysisasanimportantguaranteeforthescientificmanagementconstitutesthemainpartoftheEnglisheducationalassessment.
Theresearchonapplicationofdatamininginmanagementofstudents'gradeswantstotalkhowtogettheusefuluncoveredinformationfromthelargeamountsofdatawiththedataminingandgrademanagement[1-5].
Itintroducesandanalysesthedatamininginthemanagementofstudents'grades.
Itusesthedecisiontreeinanalysisofgrades.
Itdescribesthefunction,statusanddeficiencyofthemanagementofstudents'grades.
Ittellsushowtoemploythedecisiontreeinmanagementofstudents'grades.
ItimprovestheID3arithmetictoanalyzethestudents'gradessothatwecouldfindthelatencyfactorwhichimpactsthegrades.
Ifwefindoutthefactors,wecanofferthedecision-makinginformationtoteachers.
Italsoadvancesthequalityofteaching[6-10].
TheEnglishgradeanalysishelpsteacherstoimprovetheteachingqualityandprovidesdecisionsforschoolleaders.
Thedecisiontree-basedclassificationmodeliswidelyusedasitsuniqueadvantage.
Firstly,thestructureofthedecisiontreemethodissimpleanditgeneratesruleseasytounderstand.
Secondly,thehighefficiencyofthedecisiontreemodelismoreappropriateforthecaseofalargeamountofdatainthetrainingset.
Furthermorethecomputationofthedecisiontreealgorithmisrelativelynotlarge.
Thedecisiontreemethodusuallydoesnotrequireknowledgeofthetrainingdata,andspecializesinthetreatmentofnon-numericdata.
Finally,thedecisiontreemethodhashighclassificationaccuracy,anditistoidentifycommoncharacteristicsoflibraryobjects,andclassifytheminaccordancewiththeclassificationmodel.
Theoriginaldecisiontreealgorithmusesthetop-downrecursiveway[11-12].
Comparisonofpropertyvaluesisdoneintheinternalnodesofthedecisiontreeandaccordingtothedifferentpropertyvaluesjudgedownbranchesfromthenode.
Wegetconclusionfromthedecisiontreeleafnode.
Therefore,apathfromtheroottotheleafnodecorrespondstoaconjunctiverules,theentiredecisiontreecorrespondstoasetofdisjunctiveexpressionsrules.
Thedecisiontreegenerationalgorithmisdividedintotwosteps[13-15].
Thefirststepisthegenerationofthetree,andatthebeginningallthedataisintherootnode,thendotherecursivedataslice.
Treepruningistoremovesomeofthenoiseorabnormaldata.
Conditionsofdecisiontreetostopsplittingisthatanodedatabelongstothesamecategoryandtherearenotattributesusedtosplitthedata.
Inthenextsection,weintroduceconstructionofdecisiontree.
InSection3weintroduceattributeselectionmeasure.
InSection4,wedoempiricalresearchbasedonID3algorithmandproposeanimprovedalgorithm.
InSection5weconcludethepaperandgivesomeremarks.
CONSTRUCTIONOFDECISIONTREEUSINGID3ThegrowingstepofthedecisiontreeisshowninFigure1.
Decisiontreegenerationalgorithmisdescribedasfollows.
Thenameofthealgorithmis__Generatedecisiontreewhichproduceadecisiontreebygiventrainingdata.
Theinputistrainingsampleswhichisrepresentedwithdiscretevalues.
Candidateattributesetisattribute.
Theoutputisadecisiontree.
Step1.
SetupnodeN.
IfsamplesisinasameclassCthenreturnNasleadnodeandlabelitwithC.
Step2.
Ifattribute_listisempty,thenreturnNasleafnodeandlabelitwiththemostcommonclassinthesamples.
Step3.
Choose_testattributewithinformationgainintheattribute_list,andlabelNas_testattribute.
Step4.
Whileeachiainevery_testattributedothefollowingoperation.
Step5.
NodeNproducesabranchwhichmeetstheconditionof_itestattributeaStep6.
Supposeisissamplesetof_itestattributeainthesamples.
Ifisisempty,thenplusaleafandlabelitasthemostcommonclass.
OtherwiseplusanodewhichwasreturnedbyiGeneratedecisiontreesattributelisttestattribute.
16340ApplicationresearchofdecisiontreealgorithminenglishgradeanalysisBTAIJ,10(24)2014Figure1:GrowingstepofthedecisiontreeANIMPROVEDALGORITHMAttributeselectionmeasureSupposeSisdatasamplesetofsnumberandclasslabelattributehasmdifferentvalues(1,2,,)iCim.
SupposeiSisthenumberofsampleofclassiCinS.
Foragivensampleclassificationthedemandedexpectationinformationisgivenbyformula1[11-12].
1221log(1,2,,,)mjjmjijijiIssKsppiKn(1)12121()VjjmjjjmjjSSSEAISSKSS(2)ipisprobabilitythatrandomsamplebelongstoiCandisestimatedby/iss.
SupposeattributeAhasVdifferentvalues12Vaaa.
WecanuseattributeAtoclassifySintoVnumberofsubset12(,,)VSSS.
SupposeijSisthenumberofclassiCinsubsetjS.
Theexpectedinformationofsubsetisshowninformula2.
12()jjmjSSSSistheweightofthej-thsubset.
ForagivensubsetjSformula3setsup[13].
1221log(1,2,,,)mjjmjijijiIssKsppiKn(3)BTAIJ,10(24)2014ZhaoKun16341ijijjspsistheprobabilitythatsamplesofjsbelongstoclassiC.
IfwebranchinA,theinformationgainisshowninformula4[14].
12mGainAIsssEA(4)TheimprovedalgorithmTheimprovedalgorithmisasfollows.
Function__Generatedecisiontree(trainingsamples,candidateattributeattribute_list){SetupnodeN;IfsamplesareinthesameclassCthenReturnNasleafnodeandlabelitwithC;Recordstatisticaldatameetingtheconditionsontheleafnode;Ifattribute_listisemptythenReturnNastheleafnodeandlabelitasthemostcommonclassofsamples;Recordstatisticaldatameetingtheconditionsontheleafnode;SupposeGainMax=max(Gain1,Gain2,…,Gainn)IfGainMax='85'Updatekssetci_pi='medium'whereci_pj>='75'andci_pj='60'andci_pj<'75'Updatekssetsjnd='high'wheresjnd='1'Updatekssetsjnd='medium'wheresjnd='2'Updatekssetsjnd='low'wheresjnd='3'ResultofID3algorithmTABLE2istrainingsetofstudenttestscoressituationinformationafterdatacleaning.
Weclassifythesamplesintothreecategories.
1"outstanding"C,2"medium"C,3"general"C,1300,s21950s,3880s,3130s.
Accordingtoformula1,weobtain123300,1950,880)(300/3130)Isss2/log(300/3130).
22(1950/3130)log(1950/3130)(880/3130)log(880/3130)1.
256003.
Entropyofeveryattributeiscalculatedasfollows.
Firstlycalculatewhetherre-learning.
Foryes,11210s,21950s,31580s.
112131210,950,580)Isss222(210/1740)log(210/1740)(950/1740)log(950/1740)(580/1740)log(580/1740)1.
074901Forno,1290s,221000s,32300s.
12223290,1000,300)Isss222(90/1390)log(90/1390)(1000/1390)log(1000/1390)(300/1390)log(300/1390)1.
373186.
IFsamplesareclassifiedaccordingtowhetherre-learning,theexpectedinformationis1121311222321740/3130)1390/3130)EwhetherrelearningIsssIsss0.
5559111.
0749010.
4440891.
3731861.
240721.
Sotheinformationgainis1230.
015282GainwhetherrelearningIsssEwhetherrelearning.
Secondlycalculatecoursetype,whenitisA,112131110,200,580sss.
112131222110,200,580)(110/890)log(110/890)(200/890)log(200/890)(580/890)log(580/890)Isss1.
259382.
ForcoursetypeB,122232100,400,0sss.
BTAIJ,10(24)2014ZhaoKun1634312223222100,400,0)(100/500)log(100/500)(400/500)log(400/500)0Isss0.
721928.
ForcoursetypeC,1323330,550,0sss.
132333220,550,0)(0/550)log(0/550)(550/500)log(550/500)0Isss1.
168009.
ForcoursetypeD,14243490,800,300sss.
14243422290,800,300)(90/1190)log(90/1190)(800/1190)log(800/1190)(300/1190)log(300/1190)Isss1.
168009.
112131122232("")(890/3130)500/3130)EcoursetypeIsssIsss132333142434(550/3130)1190/3130)0.
91749.
IsssIsss("")1.
2560030.
917490.
338513Gaincoursetype.
Thirdlycalculatepaperdifficulty.
Forhigh,112131110,900,280sss.
112131222110,900,280)(110/1290)log(110/1290)(900/1290)log(900/1290)(280/1290)log(280/1290)Isss1.
14385.
Formedium,122232190,700,300sss.
122232222190,700,300)(190/1190)log(190/1190)(700/1190)log(700/1190)(300/1190)log(300/1190)Isss1.
374086Forlow,1323330,350,300sss.
1323332220,350,300)(0/650)log(0/650)(350/650)log(350/650)(300/650)log(300/650)0.
995727.
Isss112131122232("")(1290/3130)1190/3130)EpaperdifficultyIsssIsss132333(650/3130)1.
200512.
Isss("")1.
2560031.
2005120.
55497.
GainpaperdifficultyFourthlycalculatewhetherrequiredcourse.
Foryes,112131210,850,600sss16344ApplicationresearchofdecisiontreealgorithminenglishgradeanalysisBTAIJ,10(24)2014112131222210,850,600)(210/1660)log(210/1660)(850/1660)log(850/1660)(600/1660)log(600/1660)Isss1.
220681.
Forno,12223290,1100,280sss12223222290,1100,280)(90/1470)log(90/1470)(1100/1470)log(1100/1470)(280/1470)log(280/1470)Isss1.
015442.
112131122232("")(1660/3130)1470/3130)1.
220681.
EwhetherrequiredIsssIsss("")1.
2560031.
2206810.
035322.
GainwhetherrequiredTABLE2:TrainingsetofstudenttestscoresCoursetypeWhetherre-learningPaperdifficultyWhetherrequiredScoreStatisticaldataDnomediumnooutstanding90Byesmediumyesoutstanding100Ayeshighyesmedium200Dnolownomedium350Cyesmediumyesgeneral300Ayeshighnomedium250Bnohighnomedium300Ayeshighyesoutstanding110Dyesmediumyesmedium500Dnolowyesgeneral300Ayeshighnogeneral280Bnohighyesmedium150Cnomediumnomedium200ResultofimprovedalgorithmTheoriginalalgorithmlacksterminationcondition.
ThereareonlytworecordsforasubtreetobeclassifiedwhichisshowninTABLE3.
TABLE3:SpecialcaseforclassificationofthesubtreeCoursetypeWhetherre-learningPaperdifficultyWhetherrequiredScoreStatisticaldataAnohighyesmedium15Anohighyesgeneral20BTAIJ,10(24)2014ZhaoKun16345Figure2:DecisiontreeusingimprovedalgorithmAllGainscalculatedare0.
00,andGainMax=0.
00whichdoesnotconformtorecursiveterminationconditionoftheoriginalalgorithminTABLE3.
Thetreeobtainedisnotreasonable,soweadopttheimprovedalgorithmanddecisiontreeusingimprovedalgorithmisshowninFigure2.
CONCLUSIONSInthispaperwestudyconstructionofdecisiontreeandattributeselectionmeasure.
Becausetheoriginalalgorithmlacksterminationcondition,weproposeanimprovedalgorithm.
WetakecoursescoreofinstituteofEnglishlanguageforexampleandwecouldfindthelatencyfactorwhichimpactsthegrades.
REFERENCES[1]XueleiXu,ChunweiLou;"ApplyingDecisionTreeAlgorithmsinEnglishVocabularyTestItemSelection",IJACT:InternationalJournalofAdvancementsinComputingTechnology,4(4),165-173(2012).
[2]HuaweiZhang;"LazyDecisionTreeMethodforDistributedPrivacyPreservingDataMining",IJACT:InternationalJournalofAdvancementsinComputingTechnology,4(14),458-465(2012).
[3]Xin-huaZhu,Jin-lingZhang,Jiang-taoLu;"AnEducationDecisionSupportSystemBasedonDataMiningTechnology",JDCTA:InternationalJournalofDigitalContentTechnologyanditsApplications,6(23),354-363(2012).
[4]ZhenLiu,XianFengYang;"Anapplicationmodeloffuzzyclusteringanalysisanddecisiontreealgorithmsinbuildingwebmining",JDCTA:InternationalJournalofDigitalContentTechnologyanditsApplications,6(23),492-500(2012).
[5]Guang-xianJi;"Theresearchofdecisiontreelearningalgorithmintechnologyofdataminingclassification",JCIT:JournalofConvergenceInformationTechnology,7(10),216-223(2012).
[6]FuxianHuang;"ResearchofanAlgorithmforGeneratingCost-SensitiveDecisionTreeBasedonAttributeSignificance",JDCTA:InternationalJournalofDigitalContentTechnologyanditsApplications,6(12),308-316(2012).
[7]M.
SudheepElayidom,SumamMaryIdikkula,JosephAlexander;"DesignandPerformanceanalysisofDataminingtechniquesBasedonDecisiontreesandNaiveBayesclassifierFor",JCIT:JournalofConvergenceInformationTechnology,6(5),89-98(2011).
[8]MarjanBahrololum,ElhamSalahi,MahmoudKhaleghi;"AnImprovedIntrusionDetectionTechniquebasedontwoStrategiesUsingDecisionTreeandNeuralNetwork",JCIT:JournalofConvergenceInformationTechnology,4(4),96-101(2009).
[9]Bor-tyngWang,Tian-WeiSheu,Jung-ChinLiang,Jian-WeiTzeng,NagaiMasatake;"TheStudyofSoftComputingontheFieldofEnglishEducation:ApplyingGreyS-PChartinEnglishWritingAssessment",JDCTA:InternationalJournalofDigitalContentTechnologyanditsApplications,5(9),379-388(2011).
[10]MohamadFarhanMohamadMohsin,MohdHelmyAbdWahab,MohdFairuzZaiyadi,CikFazilahHibadullah;"AnInvestigationintoInfluenceFactorofStudentProgrammingGradeUsingAssociationRuleMining",AISS:AdvancesinInformationSciencesandServiceSciences,2(2),19-27(2010).
16346ApplicationresearchofdecisiontreealgorithminenglishgradeanalysisBTAIJ,10(24)2014[11]HaoXin;"AssessmentandAnalysisofHierarchicalandProgressiveBilingualEnglishEducationBasedonNeuro-Fuzzyapproach",AISS:AdvancesinInformationSciencesandServiceSciences,5(1),269-276(2013).
[12]Hong-chaoChen,Jin-lingZhang,Ya-qiongDeng;"ApplicationofMixed-Weighted-Association-Rules-BasedDataMiningTechnologyinCollegeExaminationgradesAnalysis",JDCTA:InternationalJournalofDigitalContentTechnologyanditsApplications,6(10),336-344(2012).
[13]YuanWang,LanZheng;"EndocrineHormonesAssociationRulesMiningBasedonImprovedAprioriAlgorithm",JCIT:JournalofConvergenceInformationTechnology,7(7),72-82(2012).
[14]TianBai,JinchaoJi,ZheWang,ChunguangZhou;"ApplicationofaGlobalCategoricalDataClusteringMethodinMedicalDataAnalysis",AISS:AdvancesinInformationSciencesandServiceSciences,4(7),182-190(2012).
[15]HongYanMei,YanWang,JunZhou;"DecisionRulesExtractionBasedonNecessaryandSufficientStrengthandClassificationAlgorithm",AISS:AdvancesinInformationSciencesandServiceSciences,4(14),441-449(2012).
[16]LiuYong;"TheBuildingofDataMiningSystemsbasedonTransactionDataMiningLanguageusingJava",JDCTA:InternationalJournalofDigitalContentTechnologyanditsApplications,6(14),298-305(2012).
TMThosting发布了一个2021 Summer Sale活动,针对西雅图VPS主机提供月付7折优惠码,年付65折优惠码,独立服务器提供95折优惠码,本轮促销活动到7月25日。这是一家成立于2018年的国外主机商,主要提供VPS和独立服务器租用业务,数据中心包括美国西雅图和达拉斯,其中VPS基于KVM架构,都有提供免费的DDoS保护,支持选择Windows或者Linux操作系统。Budget ...
ProfitServer已开启了黑色星期五的促销活动,一直到本月底,商家新加坡、荷兰、德国和西班牙机房VPS直接5折,无码直购最低每月2.88美元起,不限制流量,提供IPv4+IPv6。这是一家始于2003年的俄罗斯主机商,提供虚拟主机、VPS、独立服务器、SSL证书、域名等产品,可选数据中心包括俄罗斯、法国、荷兰、美国、新加坡、拉脱维亚、捷克、保加利亚等多个国家和地区。我们随便以一个数据中心为例...
官方网站:点击访问青云互联官网优惠码:五折优惠码:5LHbEhaS (一次性五折,可月付、季付、半年付、年付)活动方案:的套餐分为大带宽限流和小带宽不限流两种套餐,全部为KVM虚拟架构,而且配置都可以弹性设置1、洛杉矶cera机房三网回程cn2gia 洛杉矶cera机房  ...
1100lu.com为你推荐
微信回应封杀钉钉微信违规操作被封了,嘉兴商标注册个人如何申请商标注册比肩工场命比肩多 是什么意思啊?789se.com莫非现在的789mmm珍的com不管了qq530.com求教:如何下载http://www.qq530.com/ 上的音乐ww.66bobo.com这个WWW ̄7222hh ̄com是不是真的不太易开了,换了吗?www.ca800.com西门子plc仿真软件有什么功能javlibrary.comImage Library Sell Photos Digital Photos Photo Sharing Photo Restoration Digital Photos Photo Albums59ddd.comarmada m300什么装系统4399宠物连连看2.5我怎么找不到QQ里面的宠物连连看呢
新加坡虚拟主机 美国主机评测 香港vps99idc locvps 美国主机网 韩国电信 谷歌香港 鲜果阅读 evssl 阿里云代金券 搜狗12306抢票助手 电子邮件服务器 腾讯实名认证中心 服务器合租 四核服务器 vul 注册阿里云邮箱 成都主机托管 国内空间 winds 更多