• Tidak ada hasil yang ditemukan

fall effect is known to naturally dilute ARGs; however, ARGs are not sufficientlymanagedinmarine environments because ofcur- rentglobalwastewater management practices(Bedri etal., 2015; LawandTang,2016)

N/A
N/A
Protected

Academic year: 2023

Membagikan "fall effect is known to naturally dilute ARGs; however, ARGs are not sufficientlymanagedinmarine environments because ofcur- rentglobalwastewater management practices(Bedri etal., 2015; LawandTang,2016)"

Copied!
14
0
0

Teks penuh

(1)

aDepartment of Urban and Environmental Engineering, Ulsan National Institute of Science and Technology (UNIST), 50, UNIST-gil, Eonyang-eup, Ulju-gun, Ulsan, 44919 South Korea

bDivision of Radioactive Waste Disposal Research, Korea Atomic Energy Research Institute (KAERI), 989-111, Daedeok-daero, Yuseong-gu, Daejeon, 34057, South Korea

cDepartment of Civil and Environmental Engineering, Hanyang University, 222, Wangsimni-ro, Seongdong-gu, Seoul, 04763, South Korea

a rt i c l e i nf o

Article history:

Received 6 November 2020 Revised 27 February 2021 Accepted 1 March 2021 Available online 3 March 2021 Keywords:

Antibiotic-resistance genes (ARGs) prediction model

deep neural network

long short-term memory (LSTM) input attention

recreational beach

a b s t r a c t

Antibioticresistancegenes(ARGs)havebeenreportedtothreatenthepublichealthofbeachgoersworld- wide. Although ARG monitoring and beach guidelines are necessary, substantial efforts are required forARG samplingand analysis.Accordingly,inthisstudy,wepredicted ARGsoccurrencethatarepri- marily foundonthe coastafter rainfallusingaconventional longshort-term memory(LSTM), LSTM- convolutionalneuralnetwork(CNN)hybridmodel,andinputattention(IA)-LSTM.Todevelopthemod- els,10categoriesofenvironmentaldatacollectedat30-minintervalsandconcentrationdataof4types ofmajorARGs (i.e.,aac(6-Ib-cr), blaTEM,sul1,and tetX)obtainedatthe GwangalliBeachinSouthKo- rea,between2018and2019wereused.WhenindividuallypredictingARGsoccurrence,theconventional LSTMandIA-LSTMexhibitedpoorR2valuesduringtrainingandtesting.Incontrast,theLSTM-CNNexhib- iteda2–6-timesimprovementinaccuracyoverthoseoftheconventionalLSTMandIA-LSTM.However, when predictingall ARGsoccurrencesimultaneously, the IA-LSTMmodel exhibiteda superiorperfor- manceoverallcomparedtothatofLSTM-CNN.Additionally,theinfluenceofenvironmentalvariableson predictionwasinvestigatedusingtheIA-LSTMmodel,andtherangesofinputvariablesthataffecteach ARGwereidentified.Consequently,thisstudydemonstratedthepossibilityofpredictingtheoccurrence anddistributionofmajorARGsatthebeachbasedonvariousenvironmentalvariables,andtheresults areexpectedtocontributetomanagementofARGoccurrenceatarecreationalbeach.

© 2021ElsevierLtd.Allrightsreserved.

1. Introduction

Theemergence ofantibioticresistancegenes(ARGs)asaquatic environmentcontaminants(Prudenetal.,2006)hasbecomeasig- nificant global threat to human public health.ARGs are released from landfills or sludge through runoff, and they can flow into recreationalareasalongthecoast(Zhangetal.,2016b).Specifically, recreationalbeachesaresusceptibletoARGcontaminationthrough various sourcessuch aswastewatertreatmentplants(Proiaetal., 2018), animal feed mills (Fang et al., 2018), and storm runoff (Joyetal.,2013).Hence,inapreviousstudy,surferswerefoundto be 4.2timesmorelikely tobe exposedto ARGsthannon-surfers intheswimmingareasinEngland(Leonardetal.,2018).Therain-

Corresponding author.

E-mail address: khcho@unist.ac.kr (K.H. Cho).

1 These authors contributed equally to this study.

fall effect is known to naturally dilute ARGs; however, ARGs are not sufficientlymanagedinmarine environments because ofcur- rentglobalwastewater management practices(Bedri etal., 2015; LawandTang,2016).

Monitoring of ARGs at recreational beaches is required for beach user safety. However, ARG monitoring has the following limitations.Conventionalanalysismethods aretime-consuming;it takes5.2donaveragetoverifyincubationresults(McAdametal., 2012). Current molecular biological techniques such as quantita- tive polymerasechainreaction(qPCR) havebeenused toidentify andquantifycertain ARGs (de Castroetal., 2014;Schmiederand Edwards, 2012). Although qPCR is simpler and faster compared to conventional techniques such as the culture method or tradi- tionalPCR(KralikandRicchi,2017;SmithandOsborn,2009),reg- ularmonitoringisrestrictedduetothehighcostofqPCRanalysis (Sakthivel et al., 2012). Although multiplex PCR has been devel- oped tosave time andeffortbyreacting multiplesinglePCRssi-

https://doi.org/10.1016/j.watres.2021.117001 0043-1354/© 2021 Elsevier Ltd. All rights reserved.

(2)

multaneously,itislessaccuratebecauseitrespondstononspecific amplification products (Jansen etal.,2011; Sakthivel etal., 2012).

Therefore,forpreemptiveresponseswithinalimitedtimeframefor ARG occurrences atbeaches,prediction through modelingcan be moreefficientthanthroughmonitoring.

Long short-term memory (LSTM), a type of recurrent neural network (RNN),hasbeenwidelyused asanefficienttool tosim- ulate and predict water quality due to an ability to extract fea- tures from time-series data (lin Hsu et al., 1997). For example, Barzegar et al. (2020) recently utilized LSTM and LSTM hybrid models to predict water quality variables in a lake. An advan- tage of LSTM is that it can use memory to learn features over time. Accordingly,it isconsidereda suitableneural network(NN) forpredictingpollutantdistributions andwater qualityover time (Wangetal.,2019;Wangetal.,2017).Ontheother hand,hydro- logicalmodelssufferfromhigheruncertaintiesbecauseoftheirin- abilitytosimulatecomplexmechanisticrelationshipsamongenvi- ronmentalvariables(Abimbolaetal.,2020).Althoughdeeplearn- ing models are black boxmodels, they can improveperformance by trainingfromobservationdata(Andrychowiczetal.,2016)and simulate nonlinear phenomena occurringin the environment. In particular,deeplearningmodelshavebeenwidelyusedtoenhance the predictionperformance ofhydrologicalmodels(Parmaretal., 2017; Sumi et al, 2012). Therefore, hypothetically, it is expected that theaccuracyofdeeplearningwillbe higherthanthat ofhy- drologicalmodelstopredictARGsatarecreationalbeachofKorea affectedbyraininashortperiod.

Based on the collected literature, however, the potential of LSTM has yet to be utilized to estimate ARGs released into the environment. We previously observed the occurrence of ARGs at a combined sewer overflow (CSO) site in Gwangalli Beach over time, which varied in relation to rainfall and tides (Jang et al., 2021).Recreationalactivities atthebeach areconcentratedinthe summer andthe beachisannuallyaffected bymonsoon weather.

Therefore, ARG prediction is significant for preservingthe health of beachgoers, and the application of LSTM would be promising inpredictingtheoccurrenceofARGsovertime. Therefore,inthis study,weproposeanapproachbasedonNNtechniquestopredict ARGsoccurrencequicklyandaccuratelyformanagingandmonitor- ing their occurrenceinbeach environments.Thisstudycompared conventional LSTM, LSTM-convolutional NN (CNN), andinput at- tention(IA)-LSTMmodels(Fig.1)withthefollowingobjectives:1) to proposeapplicable modelsforpredictingfourmajor ARGs(i.e., aac(6-Ib-cr), blaTEM,sul1, andtetX) at a recreationalbeach, 2) to comparemodelaccuracieswhenpredictingsingleARGindividually andmultipleARGssimultaneously,and3)todeterminecriticalen- vironmentalfeaturesforpredictingARGoccurrences.

2. Materialandmethods 2.1. Samplinglocationandperiod

GwangalliBeach,a popularbeachinSouthKorea,wasselected as the study area. The eastern coast of Gwangalli Beachis adja- cent to theSuyeongRiver estuary, which consistsof urban areas and has a wastewater treatment plantand several seweroutlets along theriver(Fig.2).The totalarea ofthebeach is82000m2; the beach is 1.4 kmin length and25–110 m in widthalong the coastline (Choi etal., 2016). Seawater samplingswere conducted at a CSO outfall on the right side of the beach (Fig. 2). Surface seawater samples (within 1 m in depth)were automatically col- lected usingan ISCO6712 portablewatersampler (TeledyneISCO Inc.,USA)duringthedryseasonandrainfalleventsfromJune2018 toSeptember 2019(Table1).Intotal,218sampleswerecollected, andtheywerekeptinthedarkat4°Cincontainersuntiltheywere pretreatedforARGanalysis.

Table1 Samplingintervalsandamountandmaximumintensityofrainfallduringsamplingperiods(mean±standarddeviation). EventDate&TimeARGsSamplinginterval(No.ofARGdata)Rainfall(No.ofrainfalldata)Maximumrainfallintensity (a)Dryseasonsampling DryeventIJune19,2018(12:00)–June22,2018(00:00)3hrs(5)0mm(145)- DryeventIIAug20,2019(00:00)–Aug21,2019(19:00)1~3hrs(22)0mm(87)- DryeventIIISep4,2019(15:30)Sep8,2019(00:00)1~15hrs(33)0mm(186)- (b)Wetseasonsampling RainfalleventIJune26,2018(00:00)–July1,2018(00:00)3~5hrs(14)138.5mm(241)32.5mm/h RainfalleventIIMay17,2019(00:00)–May21,2019(00:00)1~31hrs(42)46.6mm(241)5.3mm/h RainfalleventIIIAug21,2019(19:30)–Aug24,2019(00:00)1hr(15)4.0mm(106)2.5mm/h RainfalleventIVAug25,2019(00:00)–Aug31,2019(00:00)1~22hrs(40)5.0mm(289)1.0mm/h RainfalleventVSep1,2019(00:00)Sep4,2019(15:00)1~12hrs(47)68.0mm(151)17.0mm/h Totalnumberofdata2181446-

(3)

Fig. 1. Summary diagram for predicting ARGs at a recreational beach.

Fig. 2. Map of study area and sampling site in Gwangalli Beach in Busan, South Korea. CSO, combined sewer overflows.

2.2. Dataacquisition

2.2.1. Environmentalvariables

Atotaloftencategoriesofenvironmentaldataincludingseven categories ofmeteorological variables (i.e., cumulative rainfall for 1hand3h,winddirection, windspeed,airpressure,airtemper- ature,andrelativehumidity) andthree categoriesofaquatic vari- ables(i.e.,watertemperature,tides,andsalinity)wereusedtode- velop the models. A total of 1,446data points were collected at 30-minintervalsformodeltraining,andthenumberofdatapoints foreacheventandtheirtrendsaresummarizedinTable1andFig.

S1,respectively.Rainfalldataforallsamplingeventswereobtained

fromtheKoreaMeteorologicalAdministration(KMA)(KMA,2015).

Becauseantecedentrainfallhaspreviouslybeenfoundtobeanim- portantenvironmentalfactorforbacteria occurrenceinGwangalli Beach(Park etal., 2018), cumulative rainfall data for1 hourand 3hours wereusedasinputvariables forthe modeldevelopment.

The total rainfall foreach wet season and maximum rainfall in- tensity in an hour are presented (Table 1). Other meteorological datawerealsocollectedfromtheKMAwhileaquaticenvironmen- tal data were obtained from the South Korea Hydrographic and OceanographicAgency(KHOA,2020).

2.2.2. EnumerationofARGs

Seawatersamples(50–500mL) werefilteredthrough0.45-

μ

m

membrane filters (Nylon, Whatman) in a vacuum to determine relative abundance ofARG. The membranes were storedin 2-mL tubes,andthetubeswerekeptat−20°Cuntilanalysis.Samplede- oxyribonucleic acid(DNA) wasextractedfrom themembrane fil- tersusinganExgeneTM soilkit(GeneAllBiotechnology, SouthKo- rea)accordingtomanufacturerinstructions.TheARGswere quan- tified using a CFX96 TouchTM Real-Time PCR Detection System (Bio-Rad,USA)and20

μ

Lofamastermixconsistingofthefollow-

ingmaterials:10

μ

LofKAPASYBR® FASTqPCRkits(KAPABiosys-

tems,USA),8.2

μ

L ofPCR-gradewater, 0.8

μ

L ofprimer (0.4

μ

L

offorward+0.4

μ

Lofreverse), and1

μ

LofDNAtemplate.Each

mastermixcontainingthetargetDNAwasamplifiedfor39cycles using the following three steps:3 min at 95°C for denaturation, 3minattheoptimaltemperatureofeach targetgene foranneal- ing,and10s at72°Cforelongation(Jangetal., 2017;Shin etal., 2019).ThetargetARGsinthisstudywereaac(6-Ib-cr),blaTEM,sul1, andtetX,whichwerethemostfrequentlyfoundonthebeachina previous study(Jangetal., 2021). Foreach target gene, theopti- mal annealing temperatures were 66.0°C for aac(6-Ib-cr), 44.0°C forblaTEM,56.0°Cfor sul1,and 61.0°C for tetX (Janget al., 2017).

The ARG abundances were quantified based on standard curves generatedusinga10-foldserialdilutionoftheplasmidDNA.Four

(4)

Fig. 3. Structure of LSTM-CNN hybrid model for time step 1 to t , which is suggested in this study. (a) and (b) indicate the input and LSTM layers of the conventional LSTM, respectively. On the LSTM layer, X tand Y tare the input and output at time t , respectively. Cell state ( C t) records past information through three gates ( f, i , and o ), sigmoid ( σ), and the activate function ( F Act). (c) LSTM output is loaded into the 1D CNN and convoluted based on the optimized filter size and activation function type. (d) ARG prediction values are derived from the fully connected layer.

ARGsweredetectedin218samples,andtheirtemporaltrendsare displayedinFig.S2.

2.3. Data-drivenmodeling

Three models were used to predict singleand multi-ARGs as described in Fig. 1. Each model was developed with the meteo- rological variables (i.e., cumulative rainfallfor1 h and3 h,wind direction, wind speed, air pressure, air temperature, and relative humidity)andaquaticvariables(i.e.,watertemperature,tides,and salinity) as input data andfour ARG abundances asoutput data.

Moreover, the hyperparameters such as the batch size, lookback, learningrate, andweightmatrixsizesinLSTM andCNN,andac- tivation functionsfor LSTMand CNNforeach modelwere deter- mined by a Bayesian optimization algorithm as describedin the supplementaryinformation.

2.3.1. ConventionalLSTM

LSTMwasdevelopedbyHochreiterandSchmidhuber(1997)to process long sequential dataandextract its temporal features.In LSTM, theflow ofinformationiscontrolled by the“gate” mecha- nism(Fig.3aandb).Thesegatescontrol theincomingandoutgo- ing information inan LSTM unit. LSTM employsthree gate types that arenamedbasedonthefunction theyperform(i.e.,theout- putgate(o),forget gate(f),andinputgate(i)).Inadditionto gates, LSTMalsohas“memory”,whichisresponsiblefortheflow of information across time. The gates interactwith the memory toeitherremoveoraddinformationtothememory.Thememory is alsocalledthe“state.” There aretwo typesofstateswithin an LSTM unit: the cell state (ct) andhiddenstate (ht). Ateach time step, an LSTMunit generates two outputs that arerepresentative of thecell andhiddenstates, whichare givenin theFig. 3band equationsbelow(HochreiterandSchmidhuber,1997).

C˜t=tanh

(

Wc[ht−1,Xt]+ bc

)

, (1)

f=

σ

Wf[ct−1,Xt]+bf

, (2)

o=

σ (

Wo[ct1,Xt]+bo

)

, (3)

i=

σ (

Wi[ct−1,Xt]+bi

)

, (4) ct=

iC˜t +

fct1, (5)

ht=

otanhct (6)

Attime stept,an LSTMmodel generatesacandidatecell (C˜t), whichisusedtoupdatethecellstate(ct)Eqs.(1)and(5).Avector oftheoutputofthepreviouscell(ht−1)andthecurrentinput(Xt) aremultipliedbytheweightofthecandidatecell(Wc),andgener- atesC˜t withbias(bc) usingthehyperbolic tangent(tanh) function Eq.(1).InEqs.(2)–(4),Wandb aretheweights andbiasesofthe gates,respectively,whicharelearnableparameters.Avectorofthe previouscellstate(ct1)andXtismultipliedbytheweightofeach gate(i.e.,Wf,Wo,andWi),andthegates are calculatedusingthe sigmoidfunction (

σ

).The outputs areproduced asct andhidden

states(ht) usingthecandidatecellandthegates Eqs.(5)and(6). ht is then used in a fully connected network to obtain the final modelpredictions.

Inthis study,we used a“many-to-many” scenario, which cal- culate theoutputs ateach time step,so that thenumberof out- puts represent the numberof input. This differs from“many-to- one” scenariowhich produces anoutput onlyat the lasttime step.

Furthermore,thenumberofinformation(i.e.,cellstateandhidden state)inthepastiscalledthe“lookback”,andinthepresentstudy, lookback1representstheinformationof30minutesago.

2.3.2. LSTM-CNNhybridmodel

Weapplied aone-dimensional(1D) CNN(Fukushima, 1979) to theoutputsequenceoftheLSTMmodeltoenhancetheprediction performance byfurtherextractingfeatures fromtheLSTM output

(5)

Fig. 4. Structure of IA-LSTM model suggested in this study. (a) IA model layer was adopted from the DA-RNN proposed by Qin et al. (2017) . x nindicates the original inputs of n number of input features, and a kt represents the attention weight for the k -th input at time t . Each attention weight is given to the original input by tensor product ( ).

(b) Driving inputs ( X ˜ t) calculated from the IA layer are used instead of the original LSTM inputs for prediction.

Table 2

Optimized hyperparameters for ARG prediction with conventional LSTM and LSTM-CNN.

ARGs LSTM CNN

Batch size Lookback Learning rate LSTM units Activation function

Filter size Activation function (a) Single ARGs prediction with conventional LSTM

bla TEM 24 8 1.00E-07 128 ReLU N/A N/A

(b) Single ARGs prediction with LSTM-CNN

aac(6’-Ib-cr) 12 16 9.59E-07 256 ReLU 128 ReLU

bla TEM 4 16 3.04E-07 64 ReLU 256 ReLU

sul 1 4 14 9.30E-07 64 Tanh 128 Tanh

tet X 24 16 9.70E-07 256 ReLU 128 Tanh

(c) Multi-ARGs prediction with LSTM-CNN

aac(6’-Ib-cr) + bla TEM+ sul 1 + tet X 4 16 3.82E-07 64 ReLU 128 ReLU

ReLU and Tanh indicate Rectified Linear Unit and hyperbolic tangent, respectively. N/A: not available

(Fig.3).Ateachtimestep,theoutputoftheLSTMmodelisthein- putintothenextmodel,the1DCNN(Fig.3c).1DCNNcanextract features by convolving a weight matrix with the inputs (Eq. 7).

The sizeoftheweightmatrixdeterminesthetotalnumberofpa- rameters in theCNN whilethe size ofthe convolution operation is based on the kernel size. The convolution output wasfurther modifiedby applyinga non-linear activationfunction. The values ofthesehyperparameterswereobtainedusingBayesianoptimiza- tion, and they are provided in Table 2. The CNN layer was fol- lowed by a maximumpoolinglayer (Eq.8) to reduce theoutput size. Finally, we applied a fullyconnected layer to obtain the fi- nal prediction from the modelaccording to the below equations (Fukushima,1979).

y=f

xw+b

(7)

z=max

(

0, x

)

(8)

where xandyindicatethe inputandoutput oftheCNN,respec- tively. w is the weight matrix, and b is the bias. The activation function(f)usedineachmodelisindicatedinTable2.

2.3.3. IA-LSTMmodel

The LSTM model can extract temporal features from inputs.

However,notalloftheinputsintoanLSTMareequallyimportant.

Modelperformancecanbeenhancedifitfocusesonmorerelevant inputs.Toaccountforthis,manyresearchershaveproposedan“at- tention” mechanism that forces the NNto focus on a certain part

oftheinput(Luongetal.,2015).Oneexampleisthedual-stageat- tentionmechanism(DA-RNN),whichwasdesignedfortimeseries predictionproblems(Qinetal.,2017).InDA-RNN,the“inputatten- tion(IA)” isfirstapplied,whichallowsthemodeltofocusonrel- evantinputs. Inthesecond stage,“temporal attention” isapplied, whichassiststhemodelinselectingrelevantdatafromthehistory.

Inthisstudy,weusedanIAmechanismincombinationwithLSTM asillustratedinFig.4.ThepurposeofIAistoadaptivelyselectrel- evantdrivinginputdata.IA-LSTMisnotablackboxmodelunlike otherdeeplearningmodelsinthatitcancalculatetheimportance ofinput variable onthe modelaccordingto the belowequations (Qinetal.,2017).

ekt =

v

Tetanh

(

We

)

ht−1;st−1+Uexk

(9)

α

tk= exp

etk

n i=1exp

eit

(10)

X˜t=

a1txt1, a2txt2, ...ant xnt

(11) Givennnumberofinputfeatures,theattentionweightforthe k-thinputatkiscalculatedusingthehiddenandcellstatesofLSTM accordingtoEq.(10).TheparametersUe,

v

Te,andWearecalibrated duringthetrainingprocess.Eq.(10)isasoftmaxfunctionthatre- turnsanarrayofattentionweightsthatsumtooneChollet,2018).

Theattentionweights foreach inputare thenusedto extractthe drivinginputsX˜t accordingtoEq.(11).Themorerelevant driving inputs are amplified by their higher weights while the less im- portantonesarereducedtosmallercorrespondingweights.These

(6)

Fig. 5. Conventional LSTM results for bla TEM. (a) Loss curves during training and testing. (b) Scatter plot and (c) time series plot comparing observations and model predic- tions.

inputs are then used as inputs into LSTM to extract the tempo- ral features.Thus,insteadofusingtheoriginalinputsXt Eqs.(1)– (4),theIAmechanismusesX˜t withinLSTM.Theremainderofthis modelissimilar tothatoftheconventionalLSTMasdescribedin Section2.3.1,whichproducesthefinalmodelprediction.

2.3.4. Modeltraining

The predictions from each model were used to calculate the model “loss” values by comparing them with the observed ARG values.Toreducethisvalue,agradient-basedlearningmechanism called“backpropagation” wasused(Rumelhartetal.,1986).During backpropagation,themodelweightsareupdatedtoreducetheloss value. Theweight changesarecontrolled bythe learningratepa- rameter.Weusedtheadaptivemotionestimation,Adamoptimizer (Kingma andBa, 2014), an adaptive learningrate algorithm. This algorithmadjuststhelearningrateduringthetrainingprocessand has beenused inmany deeplearningbenchmarks(Gregor etal., 2015;Xuetal.,2015).TheNNwasdesignedtoproducecontinuous predictions even though the observed ARG values were sparsely distributed.BecausetheobservedARGwasnotavailableatthe30- minute time step, we calculated the loss value using predictions that correspond to available observations. Predictions for which no corresponding observations were available were not used for model training. Thiswasachievedusinga maskinglayer that in- formedthemodelabouttheavailabilityorunavailabilityofobser- vations.In thismanner, itwaspossibletotrain themodels using 30-minute inputdataandsparselyobservedtargetdata.Thetotal

datasetsweredividedinto70%fortrainingand30%forvalidation.

Toimprovetheperformance,trainingandvalidationdatasetswere randomlyselectedfortrainingontheentiredataset.

The RNN-based models we usedin this studyrequirehistori- calinputdataforsingle-stepprediction.Thenumberofhistorical stepsusedbythemodelsformakingpredictionsisknownaslook- backsteps(Chollet,2018).Byincreasingthelengthofthelookback steps,wecanfeedmorehistoricaldatatotheNNtomakepredic- tions.Inthisstudy,weoptimizedthevalueoftheselookbacksteps duringhyperparameteroptimization.

ThevariationintheobservedARGatthebeachwasverysignif- icant.This isevident fromthevery highstandard deviationsand variances (Fig. S3). Toaccount for thislarge variation, we trans- formedtheARGvaluesonlogarithmicscalewithabaseof10.This isbecauselogtransformationcanreducetheeffectofoutliersfrom thedata(SinghandKingsbury,2017).Ithasbeenreportedthatlog transformationcanimprovetheperformance ofdata-drivenmod- elswhenthedatacontainoutliers(ZhengandCasari,2018).

WeusedTensorflowAPIversion 1.15(Abadietal.,2016)inthe Python programming language to build the models. The models were trained using an Intel® CoreTM i7-8700 processor with an NVIDIAGeForce GTX1060graphic card,6Gigabytes ofdedicated GPUmemory,and32Gigabytesofrandomaccessmemory.

2.3.5. Hyperparameteroptimization

WeusedaBayesianoptimizationalgorithmwithGaussianpro- cesses as a surrogate function to optimize the hyperparameters,

(7)

Fig. 6. (a) Scatter plots with trend lines (b) timeseries plots of observations and model predictions for single ARG LSTM-CNN predictions.

including thebatch size, lookback,learningrate, andweight ma- trixsizesinLSTMandCNN,andactivationfunctionsforLSTMand CNN.Theconvergence,evaluationandobjectiveplotsforeachop- timization are exhibited withmore information on the hyperpa- rameteroptimizationinthesupplementaryinformation.

2.3.6. Performanceevaluation

Outofthe218ARG samples,weused150totraintheNN,and the remaining 68 were used to evaluate themodel performance.

We used the meanabsolute error(MSE) to train the model.We alsoevaluatedthemodelperformanceusingthecoefficientofde- termination(R2), rootMSE(RMSE), andpercentbias(PBIAS).The equationstocalculatethesemetricsareasfollows:

RMSE=

n

i=1

(

oipi

)

2

n , (12)

PBIAS=

n i=1oipi

n

i=1oi × 100 (13)

whereoiand piaretheobservedandpredicteddata,respectively, and n indicates the number ofdatasets. Because the neural net- work wastrainedwithlogarithmically transformeddata,the per- formancemetricsreportedinthisstudywerecalculatedandinter- pretedonlogarithmicscale.

3. Resultsanddiscussion 3.1. Hyperparameteroptimization

3.1.1. HyperparameteroptimizationforsingleARGprediction

Bycomparingthepartialdependenceplotsinobjectiveplotsof allthemodels(Figs.S4andS5),wecaninferthatthelearningrate wasthemostsensitiveparameterduringoptimization.Incontrast, theactivationfunctionintheCNNlayerwastheleastsensitivepa- rameter forthe NNacross models. The optimization results also demonstrated that a higher lookback value resulted in a greater reduction in the model test MSEs in all cases except for blaTEM. No uniformtrendfor batch sizecan be inferred forall themod- els exceptforblaTEM, wherea smallerbatch sizeof fourresulted in the maximum improvement in performance. For LSTM units, a highervalue resultedinan improved performance forsul1and tetX, which exhibited a maximum improvement in performance with256 LSTM units.For blaTEM, the optimum number of LSTM unitswas64,anda smallernumberof LSTMunits improvedthe modelperformance.Theoptimizedvaluesofallparametersforthe LSTM-CNN andIA-LSTM models are listed inTables 2 and 3,re- spectively.Inthisanalysis,thebatchsizewasadjustedto4,8,12, and24.Theresultsdemonstratethatvariousbatch sizeswerese-

(8)

Table 3

Optimized hyperparameters for ARG prediction with IA-LSTM.

ARGs Batch size Lookback Learning rate LSTM units Activation function input attention prediction (a) Single ARGs prediction with input attention-LSTM

aac(6 -Ib-cr) 8 20 9.94E-04 16 ReLU ELU

bla TEM 24 14 5.39E-04 64 ELU Tanh

sul 1 12 19 3.39E-04 16 ReLU Tanh

tet X 24 19 9.90E-04 32 none ELU

(b) Multi-ARGs prediction with input attention-LSTM

aac(6 -Ib-cr) 12 6 8.55E-04 32 ELU ELU

bla TEM 64 LeakyReLU LeakyReLU

sul 1 16 ReLU ReLU

tet X 64 ReLU ReLU

ReLU, ELU, and Tanh indicate Rectified Linear Unit, Exponential Linear Unit, and hyperbolic tangent, respectively.

lected to test modelefficiencyinsingleARGsprediction,andthe batch sizes in the LSTM-CNN model were smaller compared to those of the conventional LSTM orIA-LSTM models (Table 2). In deeplearning,thesmallerthebatchsize,thegreaterthevariation (Keskaretal.,2016);thereby,smallerbatchsizesappeartoprovide sufficient variances forcalculating thegradient descent inLSTM- CNN.

3.1.2. Hyperparameteroptimizationformulti-ARGprediction Formulti-ARGNNpredictionwithLSTM-CNN,thehyperparam- eters wereobtainedusingtheBayesianoptimizationmethod,and they are provided in Table 2. The top rowin Fig. S4e show the variation inobjectivefunctionwithchangeineach hyperparame- ter. The steepest slope forlearningrate showsits change caused highest variations in objective functions. Thus, the learning rate can be considered as the mostsensitive hyperparameter for this model. In theLSTM-CNN model, theRectified Linear Unit (ReLU) activation functionwasthe mosteffectiveforboth LSTM andthe CNN layer. Asmaller batch size of4and 64LSTM units resulted inimprovedmodelperformancecomparedtothatwithhigherpa- rametervalues.DuetostructuraldifferenceswithintheLSTM-CNN model, the LSTM unit and activation function were assigned dif- ferently to each ARG in the LSTM-CNN. The optimized hyperpa- rameters fortheLSTM-CNNare presentedinTable 3withoutthe convergence,evaluation,andobjectiveplots.

3.2. ARGpredictionresults

3.2.1. SingleARGpredictionwithconventionalLSTM

The conventional LSTMwasexpectedto beusefulforpredict- ingARGsatarecreationalbeachbecauseARGshavebeenreported to occur duringrainfall events insteadof immediatelyafter rain- fall(Fig.S2).Inthisstudy,singleARG(blaTEM)predictionwasfirst testedusingconventionalLSTMafteroptimization.Themodelwas trained for 25,000 epochs with optimized hyperparameters, and thelosscurvesofthetrainingandtestingaredisplayedinFig.5a.

Toselectthefinalmodel,wechosethemodelwheretestlossbe- ganincreasingwithfurthertrainingsincetheincreaseintestloss indicatesoverfitting.Afterthispoint,weignoredfurtherdecreases in trainingloss. Fig.5b andcdisplay the scatter andtime series plotsoftheobservedandpredictedvalues,respectively.Theirper- formances were evaluated using R2, RMSE, andPBIAS (Table 4a).

The R2 valueswere0.09and0.16fortrainingandtesting,respec- tively, demonstrating that the conventional LSTM model did not adequately capturethefeaturesofthe blaTEM occurrenceevenaf- ter optimization.Inthiscase, increasing trainingoraddinglayers did not improvethe model performance supporting previous re- portsthatincreasing NNcomplexitydoesnot necessarilyimprove theperformancecomparedtosimpleNNs(Makridakisetal.,2018).

Table 4

Model performances for single and multi-ARG prediction.

aac(6 -Ib-cr) bla TEM sul 1 tet X train test train test train test train test (a) Single ARGs prediction with conventional LSTM

R 2 N/A N/A 0.09 0.16 N/A N/A N/A N/A

RMSE N/A N/A 0.68 0.61 N/A N/A N/A N/A

PBIAS N/A N/A -0.18 1.00 N/A N/A N/A N/A (b) Single ARGs prediction with LSTM-CNN

R 2 0.69 0.57 0.50 0.32 0.63 0.38 0.87 0.65 RMSE 0.89 1.02 0.50 0.61 0.46 0.53 0.46 0.71 PBIAS 1.67 1.78 0.36 1.23 -0.27 0.97 0.81 1.11 (c) Multi-ARGs prediction with LSTM-CNN

R 2 0.65 0.55 0.20 0.15 0.52 0.35 0.67 0.56 RMSE 0.87 0.97 0.63 0.68 0.48 0.50 0.72 0.80 PBIAS -0.02 1.89 0.67 1.84 0.18 0.63 -0.06 0.78 (d) Single ARGs prediction with IA -LSTM

R 2 0.86 0.41 0.16 0.15 0.62 0.09 0.61 0.49 RMSE 0.06 0.14 0.98 0.96 0.09 0.13 0.14 0.16 PBIAS -1.21 1.44 10.51 9.71 1.47 -3.22 -2.26 3.03 (e) Multi-ARGs prediction with IA-LSTM

R 2 0.68 0.44 0.35 0.31 0.37 0.27 0.80 0.67 RMSE 0.10 0.14 0.12 0.13 0.12 0.12 0.09 0.13 PBIAS 1.22 -0.41 -0.32 -1.77 -1.19 -3.07 -2.67 -6.35

N/A: not available

Overall, conventional LSTM didnot effectivelycapturethe occur- rence patternsof ARGs.Therefore,we combined LSTM withCNN insteadofpredictingtheotherARGswithLSTM.

3.2.2. Singleandmulti-ARGpredictionwithLSTM-CNNhybridmodel Each singleARG prediction model wastrained with hyperpa- rametersobtainedfromoptimization.Eachmodel wastrainedfor 10,000 epochs, and then the performance was evaluated using a testdatasetbecauseallmodelsachievedtheiroptimalperformance before 10,000 epochs. The trainingandtest losses ofthe models aredisplayedinFig.S6a–S6d.Weobservedthatthetestlossesfor aac(6-Ib-cr)andtetXbeganincreasingmuchearlierthanthosefor blaTEM andsul1.Theminimumtestlossesforaac(6-Ib-cr)andtetX wereachievedaftertrainingthemodelfor2,843epochsand6,461 epochs, respectively. For all models, the MSE values sufficiently predictedeach singleARG;however,thetraininglossescontinued todecrease demonstratingsigns ofoverfitting,asisevident from therisingtestlosscurves.R2 andPBIASwerealsomeasured dur- ingtrainingandtestingtoevaluatemodelperformances(Table4b).

TheR2 valuesrangedfrom0.50to0.87fortrainingandfrom0.32 to0.65fortesting.The greaterNNtrainingaccuracy comparedto thatoftestingindicatesthattheNNswereoverfittedaspresented inthelosscurves.PBIASisarelativeerrorfromtheobservations;

thereby,a positive value indicates overprediction, anda negative valuedemonstratesunderprediction.Inourmodels,thePBIASpre- sentedvaluesbetween0.27%and1.78%inboththetrainingand

(9)

Fig. 7. (a) Scatter plots with trend lines (b) timeseries plots of observations and model predictions for multi-ARG LSTM-CNN predictions.

testing results. These values can be considered “very good” lev- els forNNs with no bias in the average tendencyof predictions forobservationsaccordingtothehydrologicalcriteriaproposedby Moriasi etal.(2015). Comparisonsofthe observed andpredicted ARGsarepresentedasscatterandtime seriesplotsinFig.6aand b,respectively.Overall,theoptimalperformancewasachievedfor tetXwhiletheworstperformancewasobtainedforblaTEM.

Compared to conventional LSTM, LSTM-CNN exhibits a 2–6 times improved accuracy in blaTEM prediction (Table 2a and b).

The convolution models for sequence data were more successful when combinedwithan RNNinpreviousstudies (Barzegaretal., 2020; Lee et al., 2017). However, those studies applied LSTM after the CNN for prediction. Our LSTM-CNN model differs in that the LSTM output was applied to the CNN. In LSTM-CNN, the CNN appears to have played a role in extracting local fea- tures from the LSTM outputs. It has already been reported by Barzegar etal. (2020) that LSTM can be trained to extract long- term dependencies, and the CNN can extract time-invariant fea- tures,therebyimprovingmodelpredictability.TheproposedLSTM- CNNmodelispresumedtoexhibittheadvantagesofbothmodels.

Overall,theproposedLSTM-CNNmodelcapturesinformationfrom erraticandcomplexchangesinARGsmoreeffectivelythanthatof theconventionalLSTMmodel.

Theperformancemetricsofmulti-ARGpredictionNNfortrain- ingandtestdatasetsareprovidedinTable4c.Duringmodeltrain- ing,wecalculatedthelossfunctionaftereachepochforbothtrain- ingandtestdatawhichisplottedinFig.S6e.Correlationsbetween the observed and predicted values are illustrated as scatter and timeseriesplotsinFig.7aandb,respectively.Modelperformance has been listed for each ARG, although the models predicted all fourARGssimultaneously. Theperformance metrics foreachARG declinedcomparedtothoseoftheircorrespondingsingleARGpre- dictionmodels.Thisisduetothehigherlearningcapacityofsin- gle ARG prediction models.In contrast,the multi-ARG prediction NNshadtolearnthecomplexanddifferingbehaviorsofallARGs.

Thus,theperformance deterioratedwhentheNNpredictedmulti- pleARGs.Nevertheless,wecanconcludethatthepredictivevalues agreedwiththeobservationvaluesofthemulti-predictionmodel indicating thatthis modelperformedwell insimultaneously pre- dictingmultipleARGs.Ourresultrevealsthepotentialofapplying theLSTM-CNN hybridmodelinpredictingmulti-ARGs atarecre- ationalbeach setting,whichcanprovide areliablefoundationfor waterpolicies.TheoptimaltestperformancewasobtainedfortetX while the poorest performance was obtained for blaTEM. This is similar tosingleARG predictionmodels, wheretheblaTEM model exhibited the highest RMSE compared to that of the other ARG

(10)

Fig. 8. (a) Scatter plots with trend lines (b) timeseries plots of observations and model predictions for single ARG IA-LSTM predictions.

predictionmodels.ThelowpredictiveperformanceforblaTEM may be attributedto improper variables such asthe skewness ofthe blaTEMobservations(TableS1).Ifthedatasetfollowsanormaldis- tribution, the prediction ability of the model can increase with- out ignoringthe tailparts inthe distribution.Accordingtoapre- viousstudy,datasetswithaskewnessrageof-2to+2areconsid- ered to follow a normaldistribution (George andMallery, 2010).

For blaTEM, the skewness value was within the range of a nor- mal distribution,butitwasclosetothethreshold indicatingthat the blaTEM dataset is slightly skewed to the right compared to the otherARG datasets.Thus, itappearsthat thisslightlyskewed datasetmayhaveaffectedthetrainingandpredictionofblaTEM.

3.2.3. Singleandmulti-ARGpredictionwithIA-LSTMmodel

Scatter and time series plots are presented in Figs. 8a and b, respectively, todemonstrate thecorrelations betweenobserva- tions andpredictions forsingle ARGsusing IA-LSTM.The perfor- mance metrics forsingleARG prediction obtainedusing IA-LSTM are displayed in Table 4d. The R2 values of single ARGs predic- tions using IA-LSTM were between 0.16 and 0.86 during train- ing and 0.09 and0.49 during testing. Additionally, for aac(6-Ib- cr)andsul1predictedusingIA-LSTM,overfittingwasexpectedbe- cause the R2 training values were larger than those of testing.

ForblaTEM predictedusingIA-LSTM,theperformance wasconsid-

erably worse, which canbe attributedto inappropriate inputsas mentioned above.The IA-LSTMpredictionsverified that allARGs, except for blaTEM, were “very good” based on the PBIAS values (Moriasietal.,2015).

The multi-ARGs predictions demonstrated that the perfor- mances of aac(6-Ib-cr) andsul1 with IA-LSTM were worse than those of the single ARG predictions (Table 4e). However, the performances of blaTEM and tetX in multi-ARG predictions us- ing IA-LSTM improved compared to those of the single ARGs predictions. It is uncommon to exhibit superior performance in multi-prediction compared to individual prediction.Higher accu- racy in multi-prediction may be because the weights can be al- located to significant inputswhen learning weights. IA-LSTM re- liesonanattentionmechanismtolearntheweights ofeachtime stage (Bahdanau etal., 2016). The attentionmechanismcan sup- press noisy or unnecessary inputs by using an attention weight (Qin etal.,2017). Inparticular,whenpredictingaac(6-Ib-cr)with IA-LSTM,themodelpredictedavalue twicethat observedduring thelargestrainfall(17.5mm)betweenJune26,2018,andJune30, 2018 (Fig. 9b). Inaddition, forother ARGssuch assul1 andtetX, theIA-LSTMalsotendedtopredictmuchlargervaluesthanobser- vationsduringthelargestrainfall.ThesesuggestthattheIA-LSTM modelpredictionsallocatedtheweightstorainfallasoneofsignif- icantinputswhenlearningweights.

(11)

Fig. 9. (a) Scatter plots with trend lines (b) timeseries plots of observations and model predictions for multi-ARG IA-LSTM predictions.

All R2 values, except for aac(6-Ib-cr) in training, were lower than those of LSTM-CNN for single ARG predictions (Table 4e).

Inthecaseofmulti-ARGprediction,IA-LSTMcouldpredict multi- pleARGssimultaneouslywithhigheraccuracythanthatofLSTM- CNN.Notably,theperformanceofNNforblaTEMwithIA-LSTMwas superior to that of LSTM-CNN during both training and testing.

WithIA-LSTM,the multi-predictionperformance washigherthan that of LSTM-CNN because the attention weight was focused on the mostrelevant information, thereby maximizing the informa- tion characteristics (Qin et al., 2017). When training NNs under complex conditions that consider the characteristics of all ARGs, attentionweights in IA-LSTMcan likelyextract features moreef- fectively than the weights of LSTM or CNN. Plots of single and multi-ARG predictions duringmodel training and testing for IA- LSTM aredisplayedinFigs.8and9.TableS2showstheShannon entropy(Cover,1999)valuesofthepredictedARGfromtheLSTM- CNNandIA-LSTMmodels.WeobservedthatforLSTM-CNNbased models, the entropy of the predicted arrays is larger than those of observed. Entropy is a measure ofuncertainty and chaos ina system (Sanei andChambers, 2013). In time-series data,it quan- tifies the degree of complexity (Kumar and Dewal, 2011). Based upon theseresultsweconcludethat,IA-LSTMisamorefavorable method than LSTM-CNN in predicting multi-ARGs, while LSTM- CNNismoresuitableforpredictingsingleARGs.

3.3. ImportantvariablesforARGprediction

Input variables are important factors in determining model accuracy. However, the conventional LSTM and LSTM-CNN hy- bridmodels are uninterpretable in the importance of input vari- ables due to the blackbox effect. One of the advantages of us- ing attention-based models isthat they allow ustointerpret the model.Theplotsofattentionweightswithrespecttolookbackfor differentinputsrepresent,whichinputwasmorerelevantindecid- ingmodel’soutput.Ahigherattentionweightforaninputfeature showsthatthemodelfocusedmoreonthisinputinpredictingthe output.Since theattentionweights alsovarywithlookback,they alsoshow which previous value ofan input wasmoreimportant inmodel’sprediction.Threeinputvariables(i.e.,rainfall,tides,and salinity)that affectARGpredictionwere selected.Thiswasbased onpreviousstudiesthatfoundthatARGswereaffectedbyrainfall andtidallevel(Jangetal.,2021)andrelativelyhigherabundances ofARGswerefoundinupstreamwithlowsalinity(Leeetal.,2017; Baronetal.,2018).Althoughthevariablesrelatedtowindandtem- peraturewerealsousedasinputtoincreasepredictionability,the significance of these variables was lower than that of the three variables.Thismaybebecausethisstudytargetedtherecreational seasonofthebeach(May-September).Hadweconsideredmonitor- inginotherseasons,thevariablesaffectingtheARGconcentration

(12)

Fig. 10. Importance of input variables such as (a) rainfall, (b) tide, and (c) salinity for prediction of ARGs. The color bars represent importance of each variable. Lookback time is the time of the historical data used by Neural Network to predict next value.

could have beendifferent. Fig.10 displaysthe importance ofthe input variableaccording tothe lookback time. The closer an im- portanceis totheyellowside,the morelikelyit isthatlookback timeaffectsthepredictionwhilethecloseranimportanceistothe blueside,thelessimpactlookbacktimehasonit.Basedonatten- tion weights, rainfall substantially affected predictions withvari- ous lookback timesdependingon theARG: rainfalldata of2.5–5 hago wasimportantforaac(6’-Ib-cr)andtetX,while7.5–10hago for blaTEM and both ~7.5 h and 10 h ago for sul1 (Fig. 10a). The

previouslyreportednetworkanalysisshowedthataac(6’-Ib-cr)and tetXwere associatedwithbacterialcommunitiesduringandafter rainfall,butblaTEM andsul1showedrelevanttobacterialcommu- nitiesonlyafterrainfall(Jangetal.,2021).Basedonthisresult,we assumethatblaTEM andsul1wereaffectedbyarelativelylonglag forrainfall,possiblyduetoslowgenetransferamongbacteria.

One ofthe main features ofARG prediction is thetime delay between rainfall and ARG outflow by CSOs. As illustrated in the time series results, both LSTM-CNN andIA-LSTM respond to the

(13)

tervalsof4hwhilesul1predictionwaspredominantlyaffectedby theoveralltidaldata.Lastly,salinitydatawithin5hpriorwerethe mostinfluentialforall ARGspredictions(Fig.10c).The NNswere sensitivetochangesinsalinitybecausetheseARGscommonlyoc- curinwaterbodieswithlowersalinitylevels;aac(6-Ib-cr),blaTEM, andsul1 arefrequently detected speciesinlivestockindustries in South Korea (Lee et al., 2017), and tetracycline-resistance genes (e.g.,tetX)arefrequentlyfoundinhumanfeces(Baronetal.,2018).

In addition, it is assumed that the inflow of freshwater by rain- fall affected the ARG predictions, as prediction sensitivity corre- spondedtorainfallperiods.

4. Conclusions

The goal of this study was to improve the accuracy of pre- dictions forARG occurrenceandtoidentifythe variablesthat af- fect thesepredictions.Thus,inthisstudy,theconventionalLSTM, LSTM-CNNhybrid,andIA-LSTMmodelswere comparedtopredict ARGs occurrence according to environmental variables. The pri- maryresultsofthisstudyareasfollows:

1) The sequential convergence of LSTM and CNN resulted in improved performance compared to that of conventional LSTMtopredict singleARGs.WeshowthattheuseofCNN ontheoutputofLSTMenhancesmodel’sperformancebyex- tractinglocalfeatures..

2) IA-LSTMwasnot abletooutperform LSTM-CNNinpredict- ingsingleARGs;however,itexhibitedsuperiorperformance inpredictingmulti-ARGs.Theattentionweights ofIA-LSTM helpsthemodeltofocusondecisivefeatureswhenitisnec- essarytolearnmorecomplexfeatures.

3) Unlike the LSTM andLSTM-CNN blackbox models, the im- portanceofinputvariablesthataffectthepredictioncanbe identifiedusingIA-LSTM.

4) ARGs occurrence predictions were sensitive to input vari- ables in different lookback ranges.Therefore, we identified thetime-dependencyrangeswithintheinputsequencesthat eachARGpredictionrelieson.

Throughout thisstudy,deeplearningmodelhasdemonstrated its performanceonpredictionofARGsoccurrence.Thisstudyalso provides usefulinformation onselection ofproper deep learning techniquesforpredictingmicrobialwaterqualityatabeach.

DeclarationofCompetingInterest

Theauthorsdeclarethattheyhavenoknowncompetingfinan- cialinterestsorpersonalrelationshipsthatcouldhaveappearedto influencetheworkreportedinthispaper.

Irving, G. , Isard, M. , 2016. Tensorflow: A system for large-scale machine learning 265–283 .

Abimbola, O.P. , Mittelstet, A.R. , Messer, T.L. , Berry, E.D. , Bartelt-Hunt, S.L. , Hansen, S.P. , 2020. Predicting Escherichia coli loads in cascading dams with ma- chine learning: An integration of hydrometeorology, animal density and grazing pattern. Science of The Total Environment 722, 137894 .

Zheng, A , Casari, A , 2018. Feature engineering for machine learning: principles and techniques for data scientists . O’Reilly Media, Inc. .

Andrychowicz, M., Denil, M., Gomez, S., Hoffman, M. W., Pfau, D., Schaul, T., … De Freitas, N. (2016). Learning to learn by gradient descent by gradient descent.

arXiv preprint arXiv:1606.04474.

Bahdanau, D. , Chorowski, J. , Serdyuk, D. , Brakel, P. , Bengio, Y. , 2016. End-to-end at- tention-based large vocabulary speech recognition. IEEE, pp. 4 945–4 94 9 . Baron, S.A. , Diene, S.M. , Rolain, J.-M. , 2018. Human microbiomes and antibiotic re-

sistance. Human Microbiome Journal 10, 43–52 .

Barzegar, R. , Aalami, M.T. , Adamowski, J. , 2020. Short-term water quality variable prediction using a hybrid CNN–LSTM deep learning model. Stochastic Environ- mental Research and Risk Assessment 1–19 .

Bedri, Z. , O’Sullivan, J.J. , Deering, L.A. , Demeter, K. , Masterson, B. , Meijer, W.G. , O’Hare, G , 2015. Assessing the water quality response to an alternative sewage disposal strategy at bathing sites on the east coast of Ireland. Marine pollution bulletin 91 (1), 330–346 .

Choi, S.-H. , Lee, S.-M. , Kim, G.-S. , Kim, M.-H. , Ji, H.-S. , Jeong, Y.-N. , Yoo, E.-C. , Cho, J.-G. , 2016. Effects of Rainfall on Microbial Water Quality on Haeundae and Gwangan Swimming Beach. Journal of Bacteriology and Virology 46 (2), 71–83 . Chollet, F. , 2018. Deep Learning mit Python und Keras: Das Praxis-Handbuch vom

Entwickler der Keras-Bibliothek. MITP-Verlags GmbH & Co. KG . Cover, T.M. , 1999. Elements of information theory. John Wiley & Sons .

de Castro, A.P. , Fernandes, G.d.R. , Franco, O.L , 2014. Insights into novel antimicrobial compounds and antibiotic resistance genes from soil metagenomes. Frontiers in Microbiology 5, 489 .

Fang, H. , Han, L. , Zhang, H. , Long, Z. , Cai, L. , Yu, Y. , 2018. Dissemination of antibi- otic resistance genes and human pathogenic bacteria from a pig feedlot to the surrounding stream and agricultural soils. Journal of hazardous materials 357, 53–62 .

Fukushima, K. , 1979. Neural network model for a mechanism of pattern recognition unaffected byshift in position. Neocognitron.Trans. IECE J62-A (10), 658–665 . George, D. and Mallery, M. (2010) SPSS for Windows Step BysStep: A Simple Guide

and Reference.

Gregor, K., Danihelka, I., Graves, A., Rezende, D.J. and Wierstra, D. (2015) Draw: A re- current neural network for image generation. arXiv preprint arXiv:1502.04623.

Hochreiter, S. , Schmidhuber, J. , 1997. Long short-term memory. Neural Computation 9 (8), 1735–1780 .

Jang, H.M. , Shin, J. , Choi, S. , Shin, S.G. , Park, K.Y. , Cho, J. , Kim, Y.M. , 2017. Fate of an- tibiotic resistance genes in mesophilic and thermophilic anaerobic digestion of chemically enhanced primary treatment (CEPT) sludge. Bioresource Technology 244 (1), 433–444 .

Jang, J. , Kim, M. , Baek, S. , Shin, J. , Shin, J. , Shin, S.G. , Kim, Y.M. , Cho, K.H. , 2021. Hy- drometeorological Influence on Antibiotic-Resistance Genes (ARGs) and Bacte- rial Community at a Recreational Beach in Korea. Journal of hazardous materials 403, 123599 .

Jansen, R.R. , Schinkel, J. , Koekkoek, S. , Pajkrt, D. , Beld, M. , de Jong, M.D. , Molenkamp, R. , 2011. Development and evaluation of a four-tube real time mul- tiplex PCR assay covering fourteen respiratory viruses, and comparison to its corresponding single target counterparts. Journal of Clinical Virology 51 (3), 179–185 .

Joy, S.R. , Bartelt-Hunt, S.L. , Snow, D.D. , Gilley, J.E. , Woodbury, B.L. , Parker, D.B. , Marx, D.B. , Li, X. , 2013. Fate and Transport of Antimicrobials and Antimicro- bial Resistance Genes in Soil and Runoff Following Land Application of Swine Manure Slurry. Environmental science & technology 47 (21), 12081–12088 . Keskar, N.S., Mudigere, D., Nocedal, J., Smelyanskiy, M. and Tang, P.T.P. (2016) On

large-batch training for deep learning: Generalization gap and sharp minima.

arXiv preprint arXiv:1609.04836.

KHOA (2020) http://khoa.go.kr .

Referensi

Dokumen terkait

REVIEWER 1 No Saran Komentar Reviewer & Editor Tanggapan Penulis 1 Gap analysis belum digambarkan dalam pendahuluan Gap analysis sudah ditambahkan dalam pendahuluan “Although

Monotilla, PhD National Institute of Molecular Biology and Biotechnology, College of Science UP Diliman R&D Partners Advanced Science and Technology Institute Bureau of Agricultural