Enhancement of Bert model with predictive text generation and attention mechanisms for improvements in text autocomplete (Record no. 37424)

000 -LEADER
fixed length control field 02202nam a22001817a 4500
003 - CONTROL NUMBER IDENTIFIER
control field FT8933
005 - DATE AND TIME OF LATEST TRANSACTION
control field 20260112135035.0
050 ## - LIBRARY OF CONGRESS CALL NUMBER
Classification number QA76.9 N38 M34 2025
100 1# - MAIN ENTRY--PERSONAL NAME
Personal name Magno, Martin Angelo M.; Pangindian, Elmer Joaqi F.
245 ## - TITLE STATEMENT
Title Enhancement of Bert model with predictive text generation and attention mechanisms for improvements in text autocomplete
264 #1 - PRODUCTION, PUBLICATION, DISTRIBUTION, MANUFACTURE, AND COPYRIGHT NOTICE
Date of production, publication, distribution, manufacture, or copyright notice c2025
300 ## - PHYSICAL DESCRIPTION
Other physical details Undergraduate Thesis: (Bachelor of Science in Computer Science) - Pamantasan ng Lungsod ng Maynila, 2025
336 ## - CONTENT TYPE
Source text
Content type term text
Content type code text
337 ## - MEDIA TYPE
Source text
Media type term text
Media type code text
338 ## - CARRIER TYPE
Source unmediated
Carrier type term unmediated
Carrier type code unmediated
505 ## - FORMATTED CONTENTS NOTE
Formatted contents note ABSTRACT: Transformer-based fine-tuning strategies have shown effectiveness in-low resource and low-data context. Still, the lack of properly established baselines and benchmark datasets makes it difficult to compare different ways of dealing with low-resource settings. This research consists of three contributions. First, two previously unreleased datasets serve as benchmarks for text classification and low-resource multilabel text classification in the Filipino language. Second, the improved BERT models were pre-trained for application in the Filipino context. Third, presentation of a simple degradation text that measures a model’s susceptibility to performance decline as the number of training samples decreases. The deterioration rates of the pre-trained model and the consideration of using this method to compare models designed for low-resource environments are examined. The findings show that BERT models, even when distilled into a smaller model for low-resource contexts, maintain high performance with a minimum fine-tuning and degrade slowly in low-data conditions, making them well-suited to such restriction. There has also been research showing effective transfer from supervised tasks with large datasets, such as natural language inference (Conneau et al., 2022) and machine translation (McCann et al., 2022). Computer vision research has also demonstrated the importance of transfer learning from large pre-trained models, where an effective recipe is to fine-tune models pre-trained with ImageNet (Yosinski et. al., 2019).
942 ## - ADDED ENTRY ELEMENTS
Source of classification or shelving scheme
Item type Thesis/Dissertation
Holdings
Withdrawn status Lost status Source of classification or shelving scheme Damaged status Not for loan Collection code Permanent Location Current Location Shelving location Date acquired Total Checkouts Full call number Barcode Date last seen Price effective from Item type
          Filipiniana-Thesis PLM PLM Filipiniana Section 2025-10-24   QA76.9 N38 M34 2025 FT8933 2026-01-12 2026-01-12 Thesis/Dissertation

© Copyright 2024 Phoenix Library Management System - Pinnacle Technologies, Inc. All Rights Reserved.