Please use this identifier to cite or link to this item: https://hdl.handle.net/20.500.14279/15699
Title: Answering Open domain questions by respecting Power Law artifacts
Authors: Michael, Stavros 
Keywords: Machine learning;Deep learning;Neural Attention (NA);sequence-to-sequence (seq2seq);Natural Language Processing (NLP)
Advisor: Chatzis, Sotirios P.
Issue Date: May-2019
Department: Department of Electrical Engineering, Computer Engineering and Informatics
Faculty: Faculty of Engineering and Technology
Abstract: It has been an evolution in the last few years of machine learning methods, in which a part of this family is the deep learning models. Neural Attention (NA) is the most recent field in this area, which various methods are implemented or still in progress. One of many models that are based on NA, is sequence-to-sequence (seq2seq); an architecture of Natural Language Processing, used mainly to process data in text format and uncover useful insights. In this paper, we will focus on NA and show what are the challenges with it. We aim to examine a Question Answer (QA) model, whether addressing Power Law artifacts could facilitate modern performance; this is a plausible hypothesis, since we know that language understanding does exhibit such behavior. Also, we examine the performance of the resulting model using the benchmarks datasets.
URI: https://hdl.handle.net/20.500.14279/15699
Rights: Απαγορεύεται η δημοσίευση ή αναπαραγωγή, ηλεκτρονική ή άλλη χωρίς τη γραπτή συγκατάθεση του δημιουργού και κάτοχου των πνευματικών δικαιωμάτων.
Type: MSc Thesis
Affiliation: Cyprus University of Technology 
Appears in Collections:Μεταπτυχιακές Εργασίες/ Master's thesis

Files in This Item:
File Description SizeFormat
Abstract Stavros Michael.pdfAbstract144.65 kBAdobe PDFView/Open
CORE Recommender
Show full item record

Page view(s) 50

253
Last Week
2
Last month
13
checked on Nov 21, 2024

Download(s) 50

87
checked on Nov 21, 2024

Google ScholarTM

Check


Items in KTISIS are protected by copyright, with all rights reserved, unless otherwise indicated.