• Tidak ada hasil yang ditemukan

Word2Vec Optimization on Bi-LSTM in Electric Car Sentiment Classification

N/A
N/A
Protected

Academic year: 2024

Membagikan "Word2Vec Optimization on Bi-LSTM in Electric Car Sentiment Classification"

Copied!
9
0
0

Teks penuh

(1)

Word2Vec Optimization on Bi-LSTM in Electric Car Sentiment Classification

Siti Uswah Hasanah*, Yuliant Sibaroni, Sri Suryani Prasetyowati School of Computing, Informatics, Telkom University, Bandung, Indonesia

Email: 1,*hsnnanaa@student.telkomuniversity.ac.id, 2yuliant@telkomuniversity.ac.id, 3srisuryani@telkomuniversity.ac.id Correspondence Author Email: hsnnanaa@student.telkomuniversity.ac.id

Abstract−The Indonesian government is actively promoting electric vehicles. This policy has generated many sentiments from the public, both positive and negative. Public sentiment can have a significant impact on the success of government policies.

Therefore, it is important to understand public sentiment towards these policies. This research develops a sentiment classification model to understand public sentiment towards electric vehicles in Indonesia. Sentiment classification is the process of identifying and measuring the positive or negative sentiment in a text. This research uses a Bi-LSTM model to perform classification on a dataset of tweets related to electric vehicles. To evaluate the performance, testing was conducted through two main scenarios. In Scenario I, the focus was on finding the optimal embedding size for two Word2Vec architectures, namely CBOW and Skip-gram. Model evaluation was performed using cross-validation to gain a deeper understanding of model performance. Scenario II focused on searching for the best dropout parameters for the Bi-LSTM model.

This step aimed to find the optimal configuration for the model to generate more accurate and consistent predictions in classifying tweets related to electric vehicles. The results showed that in the context of sentiment classification on tweets about electric vehicles, the combination of CBOW with an embedding size of 200 and the Bi-LSTM model with a Dropout value of 0.2 is the best choice and achieves an accuracy of 96.31%, precision of 92.57%, Recall of 98.61%, and F1-Score of 95.49%.

Keywords: Bi-LSTM; Word2Vec; Sentiment Classification; Electric Car; Tweet

1. INTRODUCTION

At present, the Indonesian automotive industry is planning a transition from internal fuel vehicles to electrification technology, often known as electric cars. This is explained in Presidential Regulation No. 55 of 2019, which accelerates the battery-based electric vehicle program to increase the number of electric vehicles and gradually replace conventional vehicles to reduce air pollution in urban centers [1]. In addition, customers' thoughts and feelings have an impact on the adoption rate of electric cars in Indonesia; some customers said they "felt good" or

"felt guilty" after driving an electric car [2]. Some people believe that electric cars can solve the oil and gas deficit, others argue that going to electric vehicles requires careful preparation, especially infrastructure plans [3]. In the context of the adoption of electric cars in Indonesia, the influence of social media in shaping public perception.

Therefore, social media is very suitable to be used as a source of data for this study. Social media, especially Twitter, became an important internet platform in Indonesia. This is because social media is a place for people to express their opinions. These opinions can be analyzed using opinion mining systems, which what can be called sentiment analysis [4]. Sentiment analysis with deep learning methods has become a trend in research. Deep learning methods are gaining popularity in sentiment analysis research due to their ability to handle large and complex datasets [5]. One of the deep learning methods is Long Short-Term Memory.

Research using Bidirectional Long Short-Term Memory (Bi-LSTM) has been conducted by several researchers in the field of sentiment [6]–[8] in several studies showing the advantages of Bi-LSTM compared to other models, not only that the use of Bi-LSTM and Paragraph Vector shows a significant increase in performance that can handle the problem of phrase position faced by Bi-LSTM. Other studies show that using Doppler Domain and Bi-LSTM data achieves an average accuracy of 90%, while coverage domain data reaches 79%. The research [9] used Word2Vec and Bi-LSTM Attention to classify sentiment on IMDB movie reviews. This model produces the best accuracy of 84.12%, higher than previous models.

Other research conducted [10] shows the implementation of Word2Vec and LSTM for sentiment classification in hotel reviews. The results obtained achieved an average accuracy of 85.96%, which was produced using the skip-gram parameter as the Word2Vec architecture. Word2Vec is a vector representation model developed by Google. This model can be used to represent each word as an N-dimensional vector. Word2Vec has two distinct but similar architectures: CBOW and skip-gram [11]. These two architectures work in opposite ways because CBOW is a specific word that is predicted based on its surrounding neighbors, while skip-gram predicts a neighbor word based on a given target word [12].

The use of LSTM combined with Word2Vec has been carried out [13], [14] in-text classification in the fields of security and disease detection from social media posts. According to the study, LSTM Word2Vec with CBOW outperforms LSTM using the Word2Vec skip-gram feature insertion method in terms of performance. The results demonstrate that, in comparison to current methods, the suggested method yields 94% accuracy. As a result, LSTM performs better than other leading methods.

In some of the research that has been discussed, there is an important note regarding the limitations of LSTM and Word2Vec methods in the context of sentiment analysis. Previous research has highlighted that LSTM is less than optimal in capturing the context of words thoroughly and often has difficulty overcoming long distances between words. In addition, recent research highlights that Word2Vec selection has also not reached its optimality.

(2)

Although Word2Vec has great potential in representing words as N-dimensional vectors, certain parameters, such as the skip-gram embedding size and Word2Vec, have not been optimized to their full potential. The results of several studies show that the setting of these parameters affects the performance of the model in classifying sentiment. In this study, it is proposed to use Bi-LSTM and Word2Vec to improve the performance of LSTM as used by [15], in the case of electric car sentiment. Word2Vec optimisation is performed by testing various embedding sizes (100, 200, and 300) on both CBOW and Skip-gram architectures. Afterwards, the best parameters from Word2Vec will be applied to test the Dropout parameters (0.2, 0.5, and 0.7) on the Bi-LSTM model. With the ability to process information from front to back and vice versa, Bi-LSTM can effectively capture more complex contexts and improve the long-range dependency problem often faced by LSTM. This research focuses on optimising sentiment classification models using Bi-LSTM and Word2Vec to improve the accuracy and efficiency of sentiment analysis. By improving the accuracy and efficiency of sentiment analysis, this research can contribute to a better understanding of public sentiment towards electric cars in Indonesia. This can help the government and other policymakers to make more informed decisions to promote a smoother transition towards a more sustainable and environmentally-friendly transport system. The resulting model combines the best parameters from Word2vec and Bi-LSTM, with optimal tuning of the Skip-gram and Word2vec embedding sizes, as well as optimal dropout values in the Bi-LSTM layer. The model also provides better results than research conducted by [12], in the case of sentiment classification using the Bi-LSTM model which gets 91.13% accuracy.

2. RESEARCH METHODOLOGY

2.1 System Design

The system flow of this research can be seen in Figure 1. This research illustrates the system workflow that commences with the collection of Twitter tweet data. The subsequent step involves labeling this data, categorizing it into positive and negative sentiments. The labeled data undergoes a series of preprocessing procedures, including tasks such as dataset cleaning, case folding, tokenizing, and other stages. Upon completion of the preprocessing phase, the dataset is partitioned into two segments: training and testing sets. The data allocated for training proceeds to the word2vec embedding optimization stage, where optimal parameters for word2vec are tested in the subsequent phase, involving classification using the Bi-LSTM method. Model evaluation is conducted using the testing data to gauge the performance and effectiveness of the developed model.

Figure 1. Word2Vec Classification Model on Bi-LSTM 2.2 Collecting Data

This data collection uses data obtained from Twitter data, which contains a mixture of news tweets and public response tweets about electric cars. The process of collecting data is done by crawling data from Twitter using the Pandas and Node.js libraries. The tweet will be downloaded using the tweet-harvest library. This library provides various functions for crawling tweets based on specific keywords and time ranges. In this study, crawling was done with the keyword "electric car" based on tweets from 2020 to October 2023. Next, the prepared data is labeled at a later stage.

2.3 Data Labeling

Data labeling is a process in natural language processing that involves marking or labeling each word in a text based on its syntactic role in a sentence, category, aspect, or opinion. In sentiment labeling, text data will be labeled as positive or negative sentiment. Negative sentiment will be included in opinions that use harsh language or disagreeing sentences. Positive sentiment will encompass viewpoints that contain phrases or words endorsing the existence of electric vehicles [3]. Examples of labeled tweets can be seen in Table 1.

Table 1. Electric Car Tweet Examples

Tweets Sentiment

setuju si kalo ini jadi trend, gw jga kmrin mudik pake mobil listrik emang hemat dan nyaman bgt, di tambah lagi bisa cek lokasi SPKLU pake aplikasi PLN Mobile, jadi praktis banget

Positive

(3)

Tweets Sentiment

@piotrj @nafasidn Apapun itu yg jelas solusinya adalah mobil listrik. Pasti nanti jadi meningkat kualitas udaranya

Positive Ngurangi subsidi transportasi publik tapi ngasih subsidi ke mobil dan motor listrik Negative

@BagusBimantara2 @pejabrut Sumpah ini kebijakan tolol banget, gua heran sama pemikiran generasi boomers ini, mobil listrik di subsidi kendaraan umum malah di mahalin hahaha

Negative

2.4 Preprocessing

Pre-processing is a very important stage in the process of classifying text data [16]. Pre-processing is the initial stage in data processing that aims to clean, change format, and reduce the complexity of text data before it is used in analysis. Pre-processing is the process of converting raw data into data that is ready for analysis. This process is typically done to improve data quality, reduce noise, and prepare data for use by machine learning models [4].

The stages of preprocessing carried out in this study are explained as follows:

a) Dataset cleanup

In the dataset cleaning stage, specific steps were taken to clean the data. These include the removal of HTML tags to ensure the data is not contaminated with markup elements, replacing each URL with '<url>', removing usernames with '<user>', as well as removing mentions that start with '@'. In addition, this process also includes the removal of non-alphabetic characters such as punctuation marks and other symbols. Table 2 shows an example of the dataset cleaning process.

Table 2. Dataset Cleanup Tweets

#PakaiMolisyuk!, https://t.co/7SE0rmhuoX, @toyota, @, #, $, %, 1, 2, 3, Before

<url> <user> After

b) Case folding

The process of changing every letter in a text to lowercase is known as case folding. This process is carried out to homogenize the shape of letters in the text so that words with the same meaning are considered to have the same features.

c) Tokenizing

Tokenizing is the process of breaking down text or data into smaller pieces, called tokens. A token can be a word, phrase, or other piece of text that has meaning. The main purpose of tokenization is to facilitate text processing, analysis,or data processing.

d) Stopword removal

Stopwords are general words that often appear in a text but lack specific meaning in text analysis or natural language processing. The process of removing stopwords is carried out to eliminate these words from the text thereby improving the quality of text analysis by paying attention to more important words. In this study, researchers used the corpus provided by NLTK.

e) Steaming

Steaming is the process of reducing words to their basic form or root word ("stem") by eliminating word inflection or affixation. In this study, researchers used PorterStemmer() from the NLTK module. An illustration of the stemming process can be seen in Table 3.

Table 3. Stemming Before After Mengurangi Kurang

terbarukan Terbaru menyediakan sedia 2.5 Word2Vec Embedding Optimization

In this study, the formation of word vectors was carried out using the Word2Vec model, which is a vector representation model developed by Google. Word2Vec can be used to represent any word in vector form with a specific N dimension size [5]. This technique has been widely applied as an initial step towards predictive models in semantic and information retrieval tasks because it can capture contextual word-to-word relationships in multidimensional spaces [9]. The Word2Vec model has two different but similar architectures viz: CBOW and Skip-gram. In this study, both architectures are evaluated to find the best embedding size. The choice of embedding size is the result of a trade-off between the complexity of the model and the ability of the model to capture sufficient information about the words. A small embedding size tends to provide a concise representation of words but may not go deep enough in capturing the complex nuances of words. Conversely, a large embedding size is able to capture richer information but may make the model more complex.

In this research, the embedding size is explored in the range of 100, 200, and 300. The aim is to evaluate how changing the embedding size affects the performance of the Word2Vec model. The choice of embedding size

(4)

is part of the optimisation as it has a significant impact on the quality of word representation in the Word2Vec model. Illustrations of CBOW and skip-gram architecture can be seen in Figure 2.

Figure 2. CBOW and Skip-gram architecture [17]

These two architectures work in opposite ways because CBOW is a specific word that is predicated based on its surrounding neighbors, while skip-gram predicts a neighbor word based on a given target word [14]. Here is an explanation of each architecture:

1. Continuous Bag of Words (CBOW)

The CBOW architecture works more efficiently on training data in small or medium sizes [14]. This is because CBOW only needs to predict one word at a time. For example, in the sentence "Electric cars are the future", CBOW will try to predict the word "electricity" based on other words in this sentence.

2. Skip-gram

Skip-gram works the opposite of CBOW. The skip-Gram architecture works efficiently on large amounts of training data. This is because the skip-gram can capture more complex word relationships [14]. For example, if the target word is "electric", the skip-gram will try to predict words like "car", "is", and "future" based on the word "electric".

3. Window Size

The window size is a parameter that determines the number of neighboring words that will be used to predict a particular word or neighbor words. For example, if window size = 5, then CBOW will use neighboring words that are in positions 0, 1, 2, 3, and 4 of a particular word. Window sizes that are too small can cause Word2Vec models to be unable to capture complex word relationships. Window sizes that are too large can cause Word2Vec models to become too complex and take a long time to train.

4. Embedding Size

The embedding size is a parameter that determines the number of dimensions on the vector of the word representation. For example, if the embedding size = 100, then each word will be represented in vector form with 100 dimensions. Embedding sizes that are too small can cause Word2Vec models to not capture enough information about Word. Embedding sizes that are too large can cause Word2Vec models to become too complex and take a long time to train.

Previous research, conducted [10] also showed that the selection of Word2Vec was not optimal. The study used Word2Vec with skip-gram architecture to classify sentiment on hotel reviews. The results obtained reached an average accuracy of 85.96%. This shows that there is still potential to improve accuracy by using a different Word2Vec architecture. In this study, researchers implemented Word2Vec using the Gensim library, window size

= 5, using CBOW and skip-gram architectures as a comparison at each embedding size of 100, 200, and 300 for each architecture tested.

2.6 Bi-LSTM Training Model

An advancement on Long Short-Term Memory (LSTM) is Bidirectional Long Short-Term Memory (Bi-LSTM).

A front layer and a back layer, which are accessible both before and after, are combined in Bi-LSTM. LSTM solely utilizes context history, in contrast to Bi-LSTM. As a result, it performs better than LSTM in Bi-LSTM when solving sequential modeling tasks. At the moment, several advancements have been made in the classification of texts using LSTM and Bi-LSTM [18].

Bi-LSTM is one variant of the Long Short-Term Memory (LSTM) architecture used in the field of natural language processing Bi-LSTM can overcome the weaknesses of the LSTM architecture which is only able to process sequences of data in one direction only (forward) [19]. To create a Bi-LSTM network, LSTM neurons are divided into two directions: one to process the sequence of data from beginning to end (forward), and another to process the sequence of data from end to beginning (backward). LSTM is a repeatable neural network architecture designed to solve the vanishing gradient problem in RNNs [20]. Vanishing gradient is a problem that occurs when the value of the learning gradient becomes very small, thus making the training process difficult or even impossible. LSTM solves this problem by using a memory-cell block consisting of an input gate, a forgot gate, and an output gate. These memory-cell blocks serve to store information for a long time, thus allowing LSTM to study long-term dependencies [21].

(5)

LSTM has the disadvantage of only processing one-way words, namely from start to finish. Figure 3 shows the LSTM and Bidirectional LSTM configurations. Thus, Bi-LSTM can combine information from both directions of the data sequence and produce a richer representation of the input data

Figure 3. LSTM and Bi-LSTM configuration [22]

The parameters in Figure 3 play a key role in this neural network architecture. They serve as important components that facilitate the flow and processing of information from the input layer to the output layer. The following is an explanation of these parameters in the context of the Bi-LSTM network:

a) Input layer (X): This is the input layer that receives the data to be processed by the neural network.

b) Forward layer and Backward layer: These two layers are an integral part of the Bi-LSTM architecture. The forward layer is in charge of processing information sequentially from beginning to end, while the backward layer processes information in reverse, from end to beginning. Both layers consist of Long Short-Term Memory (LSTM) units that receive input from the previous layer or in parallel.

c) Output layer: This layer produces the final output of the network after information processing by both LSTM layers.

d) h (Hidden State): Each LSTM unit has a hidden state that serves as a representation of the information that has been processed at a time interval. This supports the understanding and extraction of patterns from the processed data.

e) C (cell state): Records the information in the LSTM cell during a certain time interval.

f) f, i, o (forgot gate, input gate, output gate): These components are the key elements of each LSTM unit that manage the flow of information within the LSTM cells. Forgot gate is responsible for determining the information that needs to be deleted, the input gate determines the information that needs to be updated, and the output gate determines the information to be revealed.

g) tanh Function: This function generates a candidate vector for updating the cell state on each LSTM unit.

Through the use of these parameters, the Bi-LSTM network can incorporate information from both directions of the data sequence, overcome the limitations of LSTMs that only work in one direction, and produce a richer representation of the input data.

2.7 Evaluation

In this study, researchers evaluated the model using cross-validation. Cross-Validation is a model evaluation parameter that indicates the system's ability to make accurate predictions [23]. Researchers evaluated the model using k-fold cross validation to get a more complete picture of the model's performance and reduce the risk of overfitting. In k-fold cross validation, each subset is then used as a test dataset once, while the other k-1 subsets are used as training datasets, as well as removing noise or bias in the words so that a high level of accuracycan be obtained [24]. This process is repeated k times, so each subset is used as a test dataset once. An illustration of K- fold Cross Validation can be seen in Figure 4.

Figure 4. 5-Fold Cross Validation illustration

Researchers utilize confusion matrix for evaluation parameters, such as accuracy which measures the model's ability to classify data correctly, precision which assesses the accuracy of positive predictions, recall which measures the model's ability to identify all positive examples, and F1-Score which provides a balanced measure between precision and recall. This confusion matrix provides the number of true and false predictions based on

(6)

actual values. There are 4 values, namely, True Positives (TP), True Negatives (TN), False Positives (FP), and False Negative (FN). An illustration of the confusion matrix can be seen in Figure 5. The following is an explanation of each evaluation parameter in the confusion matrix.

Explanation:

a) True Positive (TP): Indicates the amount of data correctly classified as positive (positive class) by the model.

b) True Negative (TN): Indicates the amount of data correctly classified as negative (negative class) by the model.

c) False Positive (FP): Indicates the amount of data that is incorrectly classified as positive by the model, when it should be negative.

d) False Negative (FN): Indicates the amount of data that is incorrectly classified as negative by the model, when it should be positive.

1. Accuracy

Accuracy is a measure of how accurately a model can classify data. High accuracy indicates that the model rarely makes mistakes in classifying data.

Accuracy = True Positive+True Negative

True Positive+True Negative+False Positive+False Negative (1)

2. Precision

A metric used to evaluate classification models, precision quantifies how well the model predicts positive classes. The number of accurate positive predictions divided by the total number of positive predictions the model makes yields the precision.

Precison = True Positive

True Positive+False Positive (2)

3. Recall

Recall is a measure of how well a model can identify positive examples. High recall indicates that models rarely miss positive examples. Recall is an important metric for classification models that aim to detect all positive instances.

Recall = True Positive

True Positive+False Negative (3)

4. F1-Score

F1-Score is the harmonic mean of recall and precision. This metric provides a balanced measure between the two metrics.

F1 Score = 2 ×Precision ×Recall

Precision+Recall (4)

Figure 5. Illustration and Confusion Matrix

3. RESULTS AND DISCUSSION

The study used the Bi-LSTM classification method to analyze a dataset containing tweets about electric cars, with two labels: positive and negative. Testing was conducted in two scenarios:

1. Scenario I: Looking for the best embedding size for two Word2Vec architectures: CBOW and skip-gram.

Model evaluation is done using cross-validation.

2. Scenario II: Find the best dropout parameter for the Bi-LSTM model.

This study tested embedding sizes on two types of Word2Vec architecture, CBOW (100, 200, 300), and skip-gram (100, 200, 300), as well as Dropout parameters (0.2, 0.5, 0.7) in the Bi-LSTM model. The main focus of the study was to analyze the impact of parameter changes in Word2Vec and Bi-LSTM on the accuracy of sentiment classification models. The purpose of the second experiment was to assess how differences in Bi-LSTM

(7)

parameters impact sentiment classification model performance. The results suggest that the best scheme for sentiment classification can be identified from the results of both scenarios explored.

3.1 Scenario I

Scenario 1 involves testing on two Word2Vec architectures: Continuous Bag of Words (CBOW) and skip-gram.

Within each architecture, variations in embedding size (100, 200, and 300) were evaluated using 5-fold cross validation to see how they affected sentiment classification. Table 1 shows the test results on two Word2Vec architectures.

From the results of the evaluation carried out, the best architecture that emerged was CBOW with an embedding size of 200. In the context of sentiment classification, CBOW with an embedding size of 200 showed consistency and high performance with 88.38% accuracy, 88.14% precision, 87.57% recall, and 87.85% F1-Score.

A richer word representation of an embedding size of 200 provides an advantage in understanding sentiment context which results in better results in classifying sentiments in text. Therefore, CBOW with an embedding size of 200 is the best choice for sentiment classification purposes in this study. Details of Word2vec architecture testing can be seen in Table 4.

Table 4. CBOW and Skip-gram Architecture Testing

Accuracy Precision Recall F1- Score

CBOW 100 87,25% 86,24% 87,69% 86,96%

CBOW 200 88,38% 88,14% 87,57% 87,85%

CBOW 300 87,44% 90,73 % 84,05% 87.26 %

Skip-gram 100 85,24% 86,71% 85,60% 86,15%

Skip-gram 200 88,25% 86,05% 91,16% 88,53%

Skip-gram 300 87,50% 88,50% 87,10% 87,80%

Figure 6 shows graphs of CBOW and skip-gram architectural test results at three embedding sizes, namely 100, 200, and 300. At an embedding size of 200, the CBOW line rises sharply, while the skip-gram line rises only slightly. This shows that CBOW provides a significant performance improvement compared to skip-gram. Thus, the results of this test show that CBOW with an embedding size of 200 gets the best accuracy for sentiment classification.

Figure 6. Embedding Size CBOW and Skip-gram test results 3.1.1 Scenario II

Scenario II involves testing on a Bi-LSTM model with variations in Dropout values (0.2, 0.5, 0.7) to evaluate model performance on test data. In the results you show in the Table 5 below:

Table 5. Dropout Parameter Test Results

Dropout 0.2 Dropout 0.5 Dropout 0.7

Accuracy 96,31% 95,91% 95,36%

Precision 92,57% 86,28% 84,74%

Recall 98,61% 99,62% 99,12%

F1Score 95,49% 92,47% 91,37%

These results show that varying the Dropout value has an impact on model performance. Dropout is used to prevent overfitting in neural networks, where the Dropout value controls the proportion of units that are randomly "dropped out" or deactivated at each iteration during the training process. Dropout 0.2 gives the best results with the highest accuracy (96.31%). Whereas in [10], using the LSTM and Word2Vec method, an average accuracy of 85.96% was obtained, which was generated by using the Skip-gram parameter as the Word2Vec

(8)

architecture. This proves that Word2Vec optimisation using CBOW architecture with an embedding size of 200 and a dropout parameter of 0.2 in Bi-LSTM provides better results compared to previous research.

Figure 7. Dropout Parameter Test Results

Based on the accuracy illustration obtained in each dropout value tested in Figure 7, this could indicate that at a dropout rate of 0.2, the model tends to be better at generalizing and reducing overfitting, as more information is dropped out with each iteration, forcing the model to learn more general features and rely less on specific training data

4. CONCLUSION

This study focuses on classifying sentiments related to electric cars using the Bi-LSTM classification method on a dataset of tweets categorized into positive and negative labels. Two scenarios were run to explore key parameters.

Results show in the context of sentiment classification on tweets about electric cars, the combination of CBOW with an embedding size of 200 and a Bi-LSTM model with a Dropout value of 0.2 is the best choice and achieves 96.31% accuracy, 92.57% precision, 98.61% Recall, 95.49% F1-Score. Based on the results obtained, it has proven that Word2Vec optimisation using CBOW architecture with an embedding size of 200 and a dropout parameter of 0.2 in Bi-LSTM provides better results compared to previous research. The conclusions of these two scenarios provide important insights into the influence of parameters in Word2Vec and Bi-LSTM in improving the model's ability to interpret sentiment in the context of tweets about electric cars. In future studies, researchers are expected to use more than one representation architecture, such as GloVe and FastText.

REFERENCES

[1] A. Agus Raksodewanto, “MEMBANDINGKAN MOBIL LISTRIK DENGAN MOBIL KONVENSIONAL.” Prosiding Technopex ITI, 2020

[2] A. F. Riyadi, F. R. Rahman, M. A. Nofa Pratama, M. K. Khafidli, and H. Patria, “Pengukuran Sentimen Sosial Terhadap Teknologi Kendaraan Listrik: Bukti Empiris di Indonesia,” EXPERT: Jurnal Manajemen Sistem Informasi dan Teknologi, vol. 11, no. 2, p. 141, Dec. 2021, doi: 10.36448/expert.v11i2.2171.

[3] A. Santoso, A. Nugroho, and A. S. Sunge, “Analisis Sentimen Tentang Mobil Listrik Dengan Metode Support Vector Machine Dan Feature Selection Particle Swarm Optimization,” Journal of Practical Computer Science, vol. 2, no. 1, pp.

24–31, Jul. 2022, doi: 10.37366/jpcs.v2i1.1084.

[4] Y. Pratama, D. Triantoro Murdiansyah, and K. Muslim Lhaksmana, “Analisis Sentimen Kendaraan Listrik Pada Media Sosial Twitter Menggunakan Algoritma Logistic Regression dan Principal Component Analysis,” 2023, doi:

10.30865/mib.v7i1.5575.

[5] W. Widayat, “Analisis Sentimen Movie Review menggunakan Word2Vec dan metode LSTM Deep Learning,” JURNAL MEDIA INFORMATIKA BUDIDARMA, vol. 5, no. 3, p. 1018, Jul. 2021, doi: 10.30865/mib.v5i3.3111.

[6] A. Shrestha, H. Li, J. Le Kernec, and F. Fioranelli, “Continuous Human Activity Classification from FMCW Radar with Bi-LSTM Networks,” IEEE Sens J, vol. 20, no. 22, pp. 13607–13619, Nov. 2020, doi: 10.1109/JSEN.2020.3006386.

[7] M. Arbane, R. Benlamri, Y. Brik, and A. D. Alahmar, “Social media-based COVID-19 sentiment classification model using Bi-LSTM,” Expert Syst Appl, vol. 212, p. 118710, Feb. 2023, doi: 10.1016/j.eswa.2022.118710.

[8] A. Purwarianti and I. A. P. A. Crisdayanti, “Improving Bi-LSTM Performance for Indonesian Sentiment Analysis Using Paragraph Vector,” in 2019 International Conference of Advanced Informatics: Concepts, Theory and Applications (ICAICTA), IEEE, Sep. 2019, pp. 1–5. doi: 10.1109/ICAICTA.2019.8904199.

[9] B. Jang, M. Kim, G. Harerimana, S. Kang, and J. W. Kim, “Bi-LSTM Model to Increase Accuracy in Text Classification:

Combining Word2vec CNN and Attention Mechanism,” Applied Sciences, vol. 10, no. 17, p. 5841, Aug. 2020, doi:

10.3390/app10175841.

(9)

[10] P. F. Muhammad, R. Kusumaningrum, and A. Wibowo, “Sentiment Analysis Using Word2vec And Long Short-Term Memory (LSTM) For Indonesian Hotel Reviews,” Procedia Comput Sci, vol. 179, pp. 728–735, 2021, doi:

10.1016/j.procs.2021.01.061.

[11] L. N. Aqilla, Y. Sibaroni, and S. S. Prasetiyowati, “Word2vec Architecture in Sentiment Classification of Fuel Price Increase Using CNN-BiLSTM Method,” Sinkron, vol. 8, no. 3, pp. 1654–1664, Jul. 2023, doi:

10.33395/sinkron.v8i3.12639.

[12] S. Tam, R. Ben Said, and O. O. Tanriover, “A ConvBiLSTM Deep Learning Model-Based Approach for Twitter Sentiment Classification,” IEEE Access, vol. 9, pp. 41283–41293, 2021, doi: 10.1109/ACCESS.2021.3064830.

[13] L. Xiao, G. Wang, and Y. Zuo, “Research on Patent Text Classification Based on Word2Vec and LSTM,” in 2018 11th International Symposium on Computational Intelligence and Design (ISCID), IEEE, Dec. 2018, pp. 71–74. doi:

10.1109/ISCID.2018.00023.

[14] S. Amin, M. I. Uddin, M. A. Zeb, A. A. Alarood, M. Mahmoud, and M. H. Alkinani, “Detecting Dengue/Flu Infections Based on Tweets Using LSTM and Word Embedding,” IEEE Access, vol. 8, pp. 189054–189068, 2020, doi:

10.1109/ACCESS.2020.3031174.

[15] R. H. Yahya, W. Maharani, and R. Wijaya, “JURNAL MEDIA INFORMATIKA BUDIDARMA Disaster Management Sentiment Analysis Using the BiLSTM Method,” 2023, doi: 10.30865/mib.v7i1.5573.

[16] A. S. W. R. Yuliya Astari, “Analisis Sentimen Multi-Class pada Sosial Media menggunakan metode Long Short-Term Memory (LSTM),” vol. 4, pp. 1–5, Mar. 2021.

[17] G. Futia, A. Vetro, A. Melandri, and J. C. De Martin, “Training Neural Language Models with SPARQL queries for Semi- Automatic Semantic Mapping,” Procedia Comput Sci, vol. 137, pp. 187–198, 2018, doi: 10.1016/j.procs.2018.09.018.

[18] G. Liu and J. Guo, “Bidirectional LSTM with attention mechanism and convolutional layer for text classification,”

Neurocomputing, vol. 337, pp. 325–338, Apr. 2019, doi: 10.1016/j.neucom.2019.01.078.

[19] U. B. Mahadevaswamy and P. Swathi, “Sentiment Analysis using Bidirectional LSTM Network,” Procedia Comput Sci, vol. 218, pp. 45–56, 2023, doi: 10.1016/j.procs.2022.12.400.

[20] H. Gandhi and V. Attar, “Extracting Aspect Terms using CRF and Bi-LSTM Models,” Procedia Comput Sci, vol. 167, pp. 2486–2495, 2020, doi: 10.1016/j.procs.2020.03.301.

[21] D. R. Alghifari, M. Edi, and L. Firmansyah, “Implementasi Bidirectional LSTM untuk Analisis Sentimen Terhadap Layanan Grab Indonesia,” Jurnal Manajemen Informatika (JAMIKA), vol. 12, no. 2, pp. 89–99, Sep. 2022, doi:

10.34010/jamika.v12i2.7764.

[22] R. Wadawadagi and V. Pagi, “Sentiment analysis with deep neural networks: comparative study and performance assessment,” Artif Intell Rev, vol. 53, no. 8, pp. 6155–6195, Dec. 2020, doi: 10.1007/s10462-020-09845-2.

[23] Z. Singla, S. Randhawa, and S. Jain, “Sentiment analysis of customer product reviews using machine learning,” in 2017 International Conference on Intelligent Computing and Control (I2C2), IEEE, Jun. 2017, pp. 1–5. doi:

10.1109/I2C2.2017.8321910.

[24] B. Laurensz and Eko Sediyono, “Analisis Sentimen Masyarakat terhadap Tindakan Vaksinasi dalam Upaya Mengatasi Pandemi Covid-19,” Jurnal Nasional Teknik Elektro dan Teknologi Informasi, vol. 10, no. 2, pp. 118–123, May 2021, doi: 10.22146/jnteti.v10i2.1421.

Referensi

Dokumen terkait

The condition of PGMC (Proton Green Mobility Challenge) competition stated that every participants need to change the ICE engine of the Proton Saga BLM with a DC electric motor,

This sentiment analysis process uses the Classification method with the Naive Bayes Classifier and will be compared with the XGBoost Classifier algorithm.. The results of this

Impacts of Plug-in Hybrid Electric Vehicles Charging on the Distribution Grid and Smart Charging in Proceedings of the IEEE POWERCON 2012: International Conference on Power

Whenever the data has been processed for word embedding segmentation, the weighting in this pre-trained model is word2vec, which converts the shape of the word into a vector, and the

IIT Hyderabad to collaborate on development of lithium ion batteries for electric vehicles The Indian Institute of Technology, Hyderabad, in collaboration with ItsEV Inc, plans to

3.0 SENTI2VEC: FEATURE EXTRACTION METHOD BASED ON POLARITY CLUSTERS FOR WORD2VEC The performance of a classification-based approach for sentiment analysis depends on the features

Sales Prediction Page After calculating with the backpropagation method, Figure 5 uses it to predict electric vehicles with 12 testing data.. To see the success rate of the

The main contribution of this study is to share the impact of various input shape sizes to enhance the performance of gender classification based on ECG signals using LSTM and Bi-LSTM