<?xml version="1.0" encoding="UTF-8"?>
<feed xmlns="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
<title>Architecture and Urban Planning</title>
<link href="https://repository.ju.edu.et//handle/123456789/1207" rel="alternate"/>
<subtitle/>
<id>https://repository.ju.edu.et//handle/123456789/1207</id>
<updated>2026-05-01T17:19:33Z</updated>
<dc:date>2026-05-01T17:19:33Z</dc:date>
<entry>
<title>Product Review Sentiment Analysis in Amharic using XLNet</title>
<link href="https://repository.ju.edu.et//handle/123456789/9243" rel="alternate"/>
<author>
<name>Remla Habib</name>
</author>
<author>
<name>Getachew Mamo</name>
</author>
<author>
<name>Yonas Gido</name>
</author>
<id>https://repository.ju.edu.et//handle/123456789/9243</id>
<updated>2024-04-19T06:09:42Z</updated>
<published>2024-02-01T00:00:00Z</published>
<summary type="text">Product Review Sentiment Analysis in Amharic using XLNet
Remla Habib; Getachew Mamo; Yonas Gido
Artificial Intelligence (AI) has emerged as a transformative force in various domains, &#13;
with Natural Language Processing (NLP) playing a pivotal role in enabling machines &#13;
to comprehend and generate human language. As AI advances, the application of NLP &#13;
becomes crucial for communication with intelligent systems, extending to diverse &#13;
languages, including Amharic. Sentiment Analysis (SA), a subset of NLP, is &#13;
particularly vital for extracting actionable insights from product reviews, aiding &#13;
organizations in understanding consumer sentiments.&#13;
This research addresses the unique challenges of sentiment analysis for Amharic &#13;
product reviews, marked by the absence of labeled data and the intricacies of the &#13;
language. The study focuses on XLNet, an attention mechanism transformer model, to &#13;
overcome limitations associated with masked language modeling.&#13;
Motivated by the cultural significance of Amharic as the second most spoken Semitic &#13;
language globally, the research leverages XLNet to develop robust sentiment analysis &#13;
models. The morphological richness and complex grammar of Amharic pose &#13;
challenges, prompting the investigation of XLNet's ability to capture nuanced &#13;
sentiments and word order sensitivities.&#13;
The research poses three key questions: addressing the lack of labeled data, handling &#13;
linguistic features, tokenizing Amharic pretraining data for XLNet compatibility, and &#13;
tailoring preprocessing techniques for unique linguistic characteristics. The &#13;
overarching goal is to contribute to both Amharic sentiment analysis and the broader &#13;
field of NLP.&#13;
In the pursuit of our research objectives, we conducted comprehensive experiments &#13;
comparing the performance of XLNet with other models. The findings underscore the &#13;
crucial understanding of Amharic nuances, as XLNet consistently outperformed base &#13;
cased models. Employing augmentation techniques such as random insertion, &#13;
swapping, and deletion significantly enhanced dataset variability. The meticulously &#13;
configured XLNet model, with specific hyperparameters and leveraging an augmented &#13;
dataset, showcased exceptional performance, achieving an impressive accuracy of &#13;
xii&#13;
98.10% in Amharic sentiment analysis. Subsequently, in a parallel experiment with &#13;
BERT, which yielded a commendable 90.79% accuracy, it became evident that, when &#13;
comparing both the Custom XLNet model and the BERT model using the same product &#13;
review dataset and pretraining data, the Custom XLNet model demonstrated superior &#13;
performance, further validating its efficacy in Amharic product review sentiment &#13;
analysis. The experimental results demonstrate the significance of XLNet in handling &#13;
Amharic product reviews, showcasing its potential for business strategies, product &#13;
refinement, and global competitiveness. The study contributes a valuable resource by &#13;
addressing the research gap in Amharic SA and showcases the potential of XLNet in &#13;
advancing NLP applications for morphologically rich languages.
</summary>
<dc:date>2024-02-01T00:00:00Z</dc:date>
</entry>
</feed>
