portabilitylinuxcp
linuxcp 时间:2021-04-10 阅读:(
)
ArstlookatscalableI/OinLinuxcommandsKenMatney1,ShaneCanon1,andSarpOral1CenterforComputationalSciencesOakRidgeNationalLaboratoryOakRidge,TN,37831Abstract.
Datacreatedfromandusedbyterascaleandpetascaleapplicationscontinuestoincrease,butourabilitytohandleandmanagetheselesisstilllimitedbythecapabilitiesofthestandardserializedLinuxcommandset.
ThispaperintroducestheCenterforComputationalSciences(NCCS)atOakRidgeNationalLaboratory(ORNL)eortstowardsprovidingparallelizedandmoreecientversionsofthecommonlyusedLinuxcommands.
Thedesignandimplementationdetailsaswellasperformanceanalysisofanin-housedevelopeddistributedparallelizedversionofthecptool,spdcpispresented.
Testsshowthatourspdcputilitycanachieve73timesmoreperformancethanitsserializedcounterpart.
Inaddition,weintroducecurrentworktoextendthisapproachtoothertools.
1IntroductionUsersofHPCsystemswithparallellesystemsstillrelyonlegacyserialtoolstoperformmanyday-to-dayoperations.
ParallellesystemssuchasLustreandGPFSaretodaycapableofdeliveringhundredsofGigabytespersecond(GB/s)inaggregatebandwidth,butstandardserial-basedLinuxutilitiescannotharnessthiscapability.
Forexample,makingabackupcopyofcheckpointles,compressingoutput,orcreatingatarleofresultstypicallyiscarriedoutwithstandardLinuxtools.
Consequently,usersarelimitedtotheperformancethatcanbesustainedbyasinglenodeforthesetasks.
Thus,theuserisnotabletotakeadvantageoftheextensivecapabilitiesoftheparallellesystem.
TheCenterforComputationalSciencesatOakRidgeNationalLaboratoryhasbegunworkingontoolstoaddressthisissues.
Inthispaperwewilldescribetheapproachusedindevelopingthesetoolsandpresentsomeearlyperformanceresults.
Wewillalsodiscussworkinprogressandfutureplans.
2MotivationTheNationalCenterforComputationalSciences(NCCS)atOakRidgeNationalLaboratoryoperatesanumberofthemostpowerfulcomputersystemsusedforopenresearch[1][2].
Theagshipsystem,Jaguar,isaCrayXT4withover20,000coresand40TBofmemory.
Itisconguredwithaparallellesystemwithnearly1PBofdiskcapacityandover40GB/soflesystembandwidth.
ThesystemusestheLustrelesystem[3].
TheLustrelesystemaggregatesdistributedstorageunitsintoonelogicallesystem.
Filesarestripedtransparentlybythelesystemacrossmultiplestoragetargetstoaggregatebothcapacityandbandwidth.
Asaresult,userscanachievehighthroughputtostorageforcriticalI/Ooperationssuchaswritingorreadingacheckpointle.
ApplicationssuchastheGryokineticTokamakCode(GTC)havedemonstratedover10GB/sofaggregatebandwidth.
However,manydaytodayoperationsfailtoachieveevenasmallfractionofthiscapabilitybecausetheunderlyingutilitiessuchascp,bzip2,andtarmustbeconnedtoasinglenode.
Afullystripedle(asinglelestripedacrossallstoragetargets)canbewrittenatover20GB/sonaJaguarlesystem.
However,usingcptocopythislebetweentwolocalLustrelesystemsmightonlysustain200MB/s.
Asaresult,whileitmighthavetakenaround50secondstocreatea1TBcheckpointle,itwouldtakemorethan80minutestomakeacopyofthele.
Theuserwouldlikelyencountersimilarproblemswhencompressinganduncompressingles,creatingatarle,orotheroperationsthatrelyonserial-basedtools.
Fromdiscussionswithourusers,ithasbecomeevidentthatthesebottlenecksinday-to-dayoperationsarethesourceofsomeveryrealbarrierstoproductivityandthattherewasaclearandgrowingneedforparallelversionsofthesecommontools.
Furthermore,ifageneralizedframeworkcouldbecreatedforparallelizingmanyofthesecommontasks,itcouldbeextendedtootherusecases.
Fortunately,manyofthesetoolslendthemselvestoparallelizationwithveryclearwaystodecomposethetheinputdomain.
Wechosetofocusonthoseutilitiesthatwouldquicklyprovidethemostbenettoourusercommunity.
23ApproachTherearesomelimitingfactorsinparallelizingLinuxcommands.
First,thesourcedatamustberandomlyaccessible.
Datafromacheckpointleinalesystemisanexample,whiledatafromasocketorpipeisnot.
Second,thedatasetmustresideonmultipleindependentphysicaldevices.
SinceperformanceimprovementisbasedonparallelI/O,accessingmultipleindependentphysicaldevicesconcurrentlyincreasestheachievableaggregatebandwidth.
Therearetwotypesofparallelizationthatcanbeexploited.
First,thereistheparallelismassociatedwithprocessingmultiplelessimultaneously.
Second,thereistheparallelismassociatedwithusingmultipleprocessorstomapcooperativelythedataofasinglele.
Obviously,thegainfromtheuseofthelatterisdependentonhowwellthelehasbeendistributedacrossmultipleserversandiftheworkcanbeeasilydecomposed.
Anothercriticalfactortoperformanceisthesizeofthedatabuersthatareemployed.
Likemostlesystems,parallellesystemspreferlargebuers.
Forexample,Lustrelesystemachievesbestperformancewith1MBbuers.
Parallellesystemsaretypicallymoresensitivetobuersizessincetheselesystemsrelyonnetworkstotransportdatafromthestorageserverstotheclients.
Furthermore,byterangelockingistypicallyusedtoinsureconsistency.
Largerbuersrequirelessoverheadinmanagingtheselocks,resultinginbetterperformance.
Sincethedetailsofhowtodecomposetheworkdependsonthespeciccommandtargeted,eachcommandhastobeexaminedindividually.
However,thebasisofalgorithmsforperformingI/Oinparallelremainsthesame.
Inaddition,amethodforcommunicatingbetweenthevariousparticipatingprocessorsmustbeestablished.
Whilesystemspeciclow-levelprotocolssuchasPortalsonaCrayXTorVerbsonanInniBandclustermightprovidethebestperformance,theylackportability.
Therefore,MPIisusedtoensureportabilitywhilesacricingsomedegreeofperformance.
Ourparallelizedutilitiescaneasilybeportedandcompiledformostparallelsystems.
WhileaLustrelesystemwasusedinthedevelopmentandtestingoftheinitialimplementation,thesetechniquescanbeappliedtootherparallellesystems.
Incertaincases,Lustre-speciccallstoquerythelayoutofthedataareusedtoimproveeciency.
However,goodperformanceandeciencycanstillbeachievedwithouttheseLustrespeciccalls.
LustreisaPOSIXcompliant,object-basedlesystemcomposedofthreecomponents:MetaDataServerAsingleMetaDataServer(MDS)perlesystemthatstoresandmanagesLustrelemetadata,suchaslenames,directories,permissions,stripingpattern,andlelayout.
ObjectStorageTargetOneormoreObjectStorageTargets(OSTs)areblockdevicesthatactuallystoretheledata.
OSTsaremanagedbytheObjectStorageServers(OSSs).
AtanygivencongurationtherecanbeoneormoreOSTscontrolledbyagivenOSS.
ClientClient(s)accessandusethedata.
LustreprovidesallclientswithstandardPOSIXsemanticsandconcurrentreadandwriteaccesstothelesinthelesystem.
Currently,Lustreusesanenhanchedversionofext3lesystemonMDSandOSTstostoreLustreledata.
LustreachieveshighreadandwriteperformancebydistributingtheledataovermultipleOSTs.
Thisisknownasstriping.
ThenumberofOSTsthataleisstripedacrossisknownasstripecount.
Withstriping,themaximumlesizeisnotlimitedbythesizeofasingleblockdevice,andtheaggregateI/ObandwidthscaleswiththenumberofOSSs.
AmoredetaileddescriptionofLustrelesystemisbeyondthescopeofthisdocument.
Interestedreadersareencouragedtoread[3].
TheLinuxcputilitywasselectedasthersttoolforparallelization,asitisacommonlyusedfunction,andthedecompositionissimplesincethemappingofinputdatatooutputdataisdirect.
Consequently,therearealmostnodependenciesbetweentheindividualthreadscarryingoutthecopy.
Theparallelversionofcpistermedspdcpforstreamingparalleldistributedcp.
Currently,spdcponlyworksonLustrelesystem,butourfutureplansinvolveextendingittootherlesystems,suchasGPFS.
Weareintheprocessofpubliclyreleasingthespdcpsourcecodeunderanopensourcelicense.
34PrototypeforaParallelDistributedCopyInpreparingtheprototype,therearetwopossiblewaysinwhichtoproceed.
TherstistotakethesourceforGNUcpandmodifyit.
Thesecondistowritethefunctionfromscratch.
Itisunlikelythatapatchtoreworkcpcouldmakeitintothemainstreamgiventheamountofchangesthatareneededtoparallelizeit.
Thereforewechosetoimplementanewcopycommandstartingfromscratch.
However,wetriedtopreservemanyofthecommand-lineoptionsandgeneralbehaviourofcp.
Theoveralldesignconsistsofseveralcomponents.
AdiagramofthecomponentsisshowninFig.
1.
Thebasecomponentisthe"launchprocess"whichinvokestheMPI-basedcomponents.
InadditiontolaunchingtheMPIjob,italsoperformsanumberofotheroperations,asdescribedbelow.
The"rank0process"intheMPIjobisdesignatedasamaster.
Itisresponsibleformanagingthework.
Anumberofslaveprocessesareresponsibleforcopyingtheledatafromsourcetotarget.
Howthisworkisdistributedacrosstheslavenodesisdescribedbelow.
Thereareanumberofdesignconsiderationstobemade.
First,theprototypeneedstobeawareoftheparallelcharacteristicsofsourcele(s).
Itneedstobeabletoacquiretheseattributesforsourcele(s)andsettheseontargetle(s).
Next,itneedstobeawareoftheavailableresources.
Thatistosay,iftheLinuxcommandisnotrunwithinthecontextofabatchjob,itneedstospawnabatchjobandrequestappropriateresources.
Anotherdesignchoicewastodecidehowmeta-dataoperationswouldbedecomposed.
Currently,LustreemploysasingleMetadataServer(MDS)foralesystem.
Consequently,havingmultipleclientsinteractwiththeMDSmaynotimproveperformanceandmayevenreduceit.
Therefore,theprototypeperformsmanyofthemeta-dataspecictasksinthelaunchprocess.
Forexample,theLinuxcommandthatlaunchestheMPIjob,performsthesearchforsourcele(s),acquiresbothLinuxmeta-dataandLustremeta-dataforthese,andsendsallofthisinformationtoMPImasterviaapipe.
Furthermore,thisprocesscreatesthetargetdirectoryhierarchybeforesendingthelistoflestotheMPIbasedcomponents.
Thisavoidsduplicationofeortandraceconditions,e.
g.
,multipleprocessesrequestingcreationofthesametargetdirectory.
Finally,thelaunchprocesshandlescorrectlysettingtimestampsontargetdirectorieswhenneeded.
Theadvantagetothisstrategymaynotbeobvious.
Sincethelaunchprocesshasalreadyhastraversedthesourcehierarchy,itonlyneedstoretainalistofthedirectoriesandtheirmeta-data.
ThelaunchprocessmustallowtheMPIjobtocompletesothatitcanensureanyupdatestotheaccesstimearenotoverwrittenbyanyoftheslaveprocesses.
Theprototypeemploysavariablestrategyfordecomposingworktodeterminethenumberofclientstoemployincopyingeachle.
Itmakesthisdeterminationbasedonaperformancepredictionmodelofthedataset.
ForsmalllesorleswithonlyasingleLustrestripe,theentireoperationiscarriedoutbyasingleslavenode.
Forlesthataredistributedovermultiplestripes,theworkisdistributedacrossasubsetofprocesses.
Themasterprocesswaitsuntiltheappropriatenumberofslaveprocessesareavailableandthenschedulesthecopyoperationacrossthesubset.
A"teamleader"isselectedwithinthesubset.
TheteamleaderensuresthatthetargetlehasbeencreatedwiththeappropriateLustremeta-dataparameters,suchasthestripecountandstripewidth.
Ifthetypicallemeta-data(modicationdate,etc.
)istobeanexactcopyoftheoriginal,thenalloftheteammembersreporttotheteamleaderthattheyhavecompletedalloftheirI/Orequests.
Otherwise,theteammembersreportdirectlybacktothemasternodefortheirnextassignment.
Likewise,aftertheteammembersreportbacktotheirteamleaderforcompletionnotication,theyawaitfurtherinstructionsfromthemasternode.
Theteamleaderreportstothemasternodetoindicatethatthecopyhascompletedandtheteammembersarearereadyforthenextassignment.
ThetechniquesdescribedaboveallowtheloadonthetargetOSTstobemanaged.
Byinstructingtheprototypecommandtouseonlyaspeciednumberofprocessorsfortheparallelpart,inconjunctionwithspecifyingthebuerwidth,wecanensurethattheidealnumberofclientsareparticipatinginthecopyoperationforagivenle.
ContentioncanstillarisefromothercopythreadshavingstripesthatoverlaponthesameOST.
However,preventingthiswouldincreasethecomplexityandlikelyprovideonlymarginalimprovementsinperformance.
Theprototypeimplementationofspdcpstrivestomimicthestandardcpcommandthatusersarefamiliarwith.
Theintentistocreateadropinreplacementforcpthatuserscaneasilyemployintheirexistingscripts.
4Fig.
1.
Diagramofthecomponentsusedintheparalleldistributedcopy.
Allcomputenodesaccessthelesystem.
ThenumberofteammembersusedforasourceinputledependsonthesourceleLustrestripepattern.
However,someadditionalcommand-lineoptionshavebeenaddedtocontrolaspectsoftheparallelexecutionoftheutility.
Forexample,thereareoptionstocontrolthenumberoftasksandbuersizes.
Furthermore,sinceourenvironmentrequiressubmittingabatchjobtorunaparalleljob,theutilitycantransparentlysubmititselftothebatchqueue.
Consequentlythereareoptionsrelatedtothebatchsubmissionaswell.
AsampleexecutionisshowninFig.
2.
5PerformanceAseriesofperformancemeasurementswerecarriedoutonspdcptool.
Threereferencedatasetswerecreatedinordertomeasuretheperformanceofthespdcptool.
Therstdataset(workload1)consistedof2400les,eachofsize100MB.
Thisisrepresentativeoflestypicallycreatedbyamodelingapplicationwhichlaterareanalyzedorvisualized.
Theseconddataset(workload2)consistedof10les,eachofsize24000MB.
Thisisrepresentativeofacheckpointwhichisdonetoasharedle.
Thethirddataset(workload3)consistedof1200lesofsize100MBand5lesofsize24,000MB.
Thiswasdonetodemonstratetheabilitytoecientlycopyanon-uniformdataset.
TheLinuxcpcommandwasusedtoestablishbaselineperformance.
Then,weevaluatedtheperformanceatvariousscalesinordertounderstandthescalingbehaviorfortheprototype.
Thesemeasurementswereperformedona3500socketCrayXT3systemagainstitslocalLustrelesystem.
TheLustrelesystemconsistedof80OSTsservedby20ObjectStorageServers(OSSs).
Thebackendstoragewasprovidedby10coupletsofDDN8500[10].
ThislesystemhasbeenmeasuredusingtheIOR[11]benchmarktosustainover10GB/sonale-per-processrun.
5spdcp-s16-r/source/directory//target/directory/spdcp-hUsage:spdcp[options]SRCDESTorspdcp[options]SRC.
.
.
DIRECTORYCopyfileSRCtofileDESTorlistoffilesSRC.
.
.
todirectoryDIRECTORY,replicatingLustrestripeinformationwherepossible.
CopyisperformedinparallelbydistributedclientsusingMPImessagepassingforsynchronizationandcontrol.
Whencomputenoderesourcesareaccessibleonlyinbatchmode,commandwillstagejobandretaincontroluntiljobfinishes.
Thefollowingoptionsoffercontrolovercommand:-hPrintthismessage(disablescopy)-VPrintcommandandagentversions(disablescopy)-dUsedummyform(disablescopy,printstargets)-vIncreaseverbositylevel(maximum2)-pPreservemode,ownership,andtimestamps-r,-RCopyrecursively-cReduceOSTcountatdestinationtosourceusage-nDonotoffsetinitialOSTatdestination-b{F}IncreaseI/OrequestsizebyafactorofF-s{M}EmployMparallelclients-A{P}Ifspawningbatchjob,chargeruntoproject,P-w{T}Ifspawningbatchjob,limitwalltimetoTseconds-q{Q}Ifspawningbatchjob,directtobatchqueue,QFig.
2.
Sampleexecutionofspdcp(Top).
Thetotalnumberofclientsrequestedisidentiedbythe-sswitch.
Notethat,thisnumberalsoincludesthe"master(orrank0)node.
"Thespdcphelpmenu(Bottom).
AscanbeseeninFig.
3,spdcpachievesgoodparallelspeedup.
ThedataexhibitacertainamountofvariationbecausetheywereobtainedduringthecourseofnormalproductionoperationoftheCrayXT3.
ItshouldbenotedthatthestockLinuxcputilityachieved324MB/s,126MB/s,and177MB/sforworkload1,workload2,andworkload3,respectively.
Intermsofpeakperformance,ascanbeseeninFig.
3,theworkload2achievesthebestperformancewithspdcp,ataround9300MB/s.
Thisisa73xperformanceincreasecomparedtotheLinuxcputility.
Thepeakperformanceis7300MB/sforworkload1.
Thisis22xspeedupcomparedtotheLinuxcputility.
Forworkload3thepeakisatapproximately9100MB/s;a51xspeedupovertheLinuxcputility.
Also,ascanbeseeninFigure3,thepeakperformanceisobtainedat160to256clients.
However,fromapracticalpointofview,thescalingofperformancelevelsoataround100clients.
ThismakessensegiventhatthenumberofclientsandOSTsareroughlyequivalent.
Consequently,theOSTshavenearlyreachedtheirpeakbandwidth.
Thisisfurtherdemonstratedbythefactthattheaggregatebandwidthis73%to93%ofthepeakbandwidthasmeasuredbyIOR.
6On-goingworkTheparallelimplementationofthecopyutilityisjusttherststepinabroaderinitiativetocreateasuiteofparallelizedtools.
Towardsthisend,wehavestartedtocreateaframeworktogeneralizetheapproachesused61001000100001101001000AggregateBandwidth(MB/s)NumberofClientsWorkload1Workload2Workload3Fig.
3.
spdcpperformanceforclientsupto512.
Theworkload1iscomposedoflargeles,workload2iscomposedofsmallles,andworkload3isamixoflargeandsmallles.
ThestockLinuxcputilityachieved324MB/s,126MB/s,and177MB/sforworkload1,workload2,andworkload3,respectively(notshownonthegure).
inspdcpsothattheycaneasilybeappliedtoothercommonutilities.
Thespdcputilitydoesnotcurrentlyusetheframework,butmaybere-implementedusingtheframeworkinthenearfuture.
Thisframework,whichiscalledspdframe,hasalreadybeenusedforcompressionanddecompressionofbzip2les[12].
Thispresentsslightlymoredicultythanthecopytool,asthedecompositionfordecompressionismoredicult.
Preliminarytestsshowthatourbzip2implementationispromisingandunderrightcongurations(e.
g.
64processorswitha20MBle)itcanachieve15timesmoreperformanceforcompressioncomparedtoitsserializedversionon.
Futureworkwillfocusonapplyingtheframeworktotarandothercommonlebasedutilities.
Whilewearefocusingonapplyingtheframeworktocommontools,theframeworklendsitselftootherusesaswell.
Theframeworkprovidesaneasywayforuserstoapplyafunctionovermultiplelesinparallel.
So,forexample,ausercouldeasilyapplytheframeworktoperformaparallelgreponasetofles.
7RelatedworkIncreasingtheperformanceofcommonLinuxutilitiesgatheredsomeattentionfromtheresearchcommunityovertheyears.
WilliamGroppandEwingLusk[4]haverstrealizedthelimitationsoflegacyserialUNIXutilitiesinparallelenvironments.
TheyintroducedseveralparallelversionsofcommonlyusedUNIXutilitieswithparallelrshastheunderlyingparallelsynchronizationandcommunicationmechanism.
Asafollowuptotheirwork,EmilOng,EwingLusk,andWilliamGroppdevelopedtheMPI-basedversionoftheirparallelizedUNIXutilities[5].
However,thereisacleardistinctionbetweenourgoalandtheirs.
ThetargetforGroppandLuskwasincreaseeciencybyexecutingthesamecommandwiththesameargumentlistandparametersinparallelovermultipleindependentnodeswithindependentoperatingsystemsandlesystems.
Inmanyaspects,theyhaveimplementedSIMD-likeversionsofthecommonUNIXtools.
However,ourapproachdepartsfromtheirsasourgoalwastoincreasetheeciencyofasingleexecutionagivenLinuxutilitybyparallelizinganddistributingitsworkloadovermultipleworker/computenodes,allsharingacommonlesystem,butindependentOSes.
JeGilchristandAysegulCuhadar[7]introducedtwoparallelizedversionsofBWT-basedbzip2nblock-sortinglecompressor,namelypbzip2andmpibzip2.
Thepbzip2isathread-parallelversionofbzip2foruse7onsharedmemorymachines.
Itproducescompatiblebutlargerarchivescomparedtotheoriginalbzip2.
Thempibzip2isanMPI-basedparallelimplementationofthebzip2block-sortinglecompressorforclusters.
Thebzip2smpprogramisanotherparallelizedversionofthebzip2compressor[8].
ItisspecicallytargetedforSMPsystems.
Itisverycache-dependantanddoesnotperformwellwithhyperthreadedsystems.
Itissimilartopbzip2innature,butunlikepbzip2,bzip2smpsupportscompressionfromstdin.
ConclusionIncreasingparallelisminlesystemspavethewayforprocessinglargerdatasetsinshortertimes.
However,whilecapabilitiesforgeneratinglargerdatasetsareconstantlyincreasing,ourtoolsforhandlingandmanagingsuchles,stillremainserialandlimitedinperformance.
TheCenterforComputationalSciences(NCCS)atOakRidgeNationalLaboratory(ORNL)hasstartedaninitiativeforprovidinghigh-performance,parallelversionsofcommonlyusedLinuxcommands.
Thecpcommandwasourstartingpoint.
WehavedevelopedandimplementedaMPI-basedbatch-processingcapableparallelversionofthestandardcpcommand.
Testsshowthat,ourversioncanachieve73timesmoreperformanceoveritsstandardserializedcounterpart.
Also,thispaperintroducesoureortstowardsdevelopingaparallelizeddistributedversionofthebzip2command.
Theimplementationfollowsaframework,whichifsuccessful,willbeusedfordevelopingandparallelizingotherLinuxcommands.
AcknowledgmentsTheauthorswouldliketothankthestaandcolleagueswhohavecontributedmaterialtothispaper.
ResearchsponsoredbytheMathematical,Information,andComputationalSciencesDivision,OceofAd-vancedScienticComputingResearch,U.
S.
DepartmentofEnergy,underContractNo.
DE-AC05-00OR22725withUT-Battelle,LLC.
AbouttheAuthorsKenMatneyisaresearcherintheTechnologyIntegrationGroupwhichispartoftheNationalCenterforCom-putationalSciencesatOakRidgeNationalLab.
HecanbereachedbyE-Mail:matneykdsr@ornl.
gov.
ShaneCanonistheGroupLeaderforTechnologyIntegrationTeam.
HecanbereachedbyE-Mail:canonrs@ornl.
gov.
SarpOralisaresearcherintheTechnologyIntegrationGroupwhichispartoftheNationalCenterforCom-putationalSciencesatOakRidgeNationalLab.
HecanbereachedbyE-Mail:oralhs@ornl.
gov.
References1.
NationalCenterforComputationalSciences.
WebPagehttp://nccs.
gov.
2.
Top500Supercomputersites-November2007list.
WebPagehttp://www.
top500.
org/list/2007/11.
3.
ClusterFileSystems,Inc.
Lustremanual.
Webpage.
http://www.
lustre.
org/manual.
html.
4.
WilliamGroppandEwingL.
Lusk.
ScalableUnixtoolsonparallelprocessorsInProceedingsoftheScalableHigh-PerformanceComputingConference,pp.
56-62,1994.
5.
EmilOng,EwingL.
Lusk,andWilliamGropp.
ScalableUnixCommandsforParallelProcessors:AHigh-PerformanceImplementationInProceedingsofthe8thEuropeanPVM/MPIUsers'GroupMeetingonRecentAdvancesinParallelVirtualMachineandMessagePassing,pp.
410-418,2001.
6.
M.
BurrowsandD.
J.
Wheeler.
Ablock-sortinglosslessdatacompressionalgorithmTechnicalReport124,DigitalSystemsResearchCenter,1994.
7.
JeGilchristandAysegulCuhadar.
ParallelLosslessDataCompressionBasedontheBurrows-WheelerTransformIn21stInternationalConferenceonAdvancedNetworkingandApplications(AINA'07),pp.
877-884,2007.
8.
WebPagehttp://bzip2smp.
sourceforge.
net/9.
R.
S.
CanonandH.
SarpOral.
ACenter-wideFileSystemusingLustre.
InCUGProceedings,2006.
10.
DataDirectNetworks.
WebPagehttp://datadirectnetworks.
com/11.
HedgesetalParallellesystemtestingforthelunaticfringe:thecareandfeedingofrestlessI/OpowerusersInIEEEMassStorageSystemsandTechnologiesProceedings,200512.
JulianSeward.
Thebzip2andlibbzip2ocialhomepage.
WebPagehttp://sources.
redhat.
com/bzip2
diyvm怎么样?diyvm是一家国内成立时间比较久的主机商家了,大约在6年前站长曾经用过他家的美国机房的套餐,非常稳定,适合做站,目前商家正在针对香港沙田机房的VPS进行促销,给的是五折优惠,续费同价,香港沙田机房走的是CN2直连的线路,到大陆地区的速度非常好,DiyVM商家采用小带宽不限流量的形式,带宽2Mbps起步,做站完全够用,有需要的朋友可以入手。diyvm优惠码:五折优惠码:OFF50...
hostkvm怎么样?hostkvm是一家国内老牌主机商家,商家主要销售KVM架构的VPS,目前有美国、日本、韩国、中国香港等地的服务,站长目前还持有他家香港CN2线路的套餐,已经用了一年多了,除了前段时间香港被整段攻击以外,一直非常稳定,是做站的不二选择,目前商家针对香港云地和韩国机房的套餐进行7折优惠,其他套餐为8折,商家支持paypal和支付宝付款。点击进入:hostkvm官方网站地址hos...
轻云互联成立于2018年的国人商家,广州轻云互联网络科技有限公司旗下品牌,主要从事VPS、虚拟主机等云计算产品业务,适合建站、新手上车的值得选择,香港三网直连(电信CN2GIA联通移动CN2直连);美国圣何塞(回程三网CN2GIA)线路,所有产品均采用KVM虚拟技术架构,高效售后保障,稳定多年,高性能可用,网络优质,为您的业务保驾护航。官方网站:点击进入广州轻云网络科技有限公司活动规则:用户购买任...
linuxcp为你推荐
哈利波特罗恩升级当爸哈利波特中的赫敏为什么要喜欢罗恩,不喜欢哈利哈利波特罗恩升级当爸哈利波特的爸爸妈妈身份云爆发云联惠是什么来的嘉兴商标注册我在濮院想注册一个羊毛衫商标?该怎么做?地陷裂口地陷是由什么原因引起的seo优化工具想找一个效果好的SEO优化软件使用,在网上找了几款不知道哪款好,想请大家帮忙出主意,用浙江哪款软件效果好www.622hh.comwww.710av.com怎么不可以看了www.gegeshe.com有什么好听的流行歌曲5xoy.comhttp://www.5yau.com (舞与伦比),以前是这个地址,后来更新了,很长时间没玩了,谁知道现在的地址? 谢谢,www.niuav.com在那能找到免费高清电影网站呢 ?
vps优惠码 西部数码vps kddi patcha 天猫双十一抢红包 新站长网 骨干网络 免费个人空间 河南移动网 免费申请网站 免费测手机号 免费phpmysql空间 常州联通宽带 厦门电信 万网主机 如何登陆阿里云邮箱 万网服务器 websitepanel ncp是什么 globalsign 更多