Sign in to use this feature.

Years

Between: -

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (763)

Search Parameters:
Journal = BDCC

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
21 pages, 807 KiB  
Review
Digital Eye-Movement Outcomes (DEMOs) as Biomarkers for Neurological Conditions: A Narrative Review
by Lisa Graham, Rodrigo Vitorio, Richard Walker, Gill Barry, Alan Godfrey, Rosie Morris and Samuel Stuart
Big Data Cogn. Comput. 2024, 8(12), 198; https://doi.org/10.3390/bdcc8120198 - 19 Dec 2024
Viewed by 337
Abstract
Eye-movement assessment is a key component of neurological evaluation, offering valuable insights into neural deficits and underlying mechanisms. This narrative review explores the emerging subject of digital eye-movement outcomes (DEMOs) and their potential as sensitive biomarkers for neurological impairment. Eye tracking has become [...] Read more.
Eye-movement assessment is a key component of neurological evaluation, offering valuable insights into neural deficits and underlying mechanisms. This narrative review explores the emerging subject of digital eye-movement outcomes (DEMOs) and their potential as sensitive biomarkers for neurological impairment. Eye tracking has become a useful method for investigating visual system functioning, attentional processes, and cognitive mechanisms. Abnormalities in eye movements, such as altered saccadic patterns or impaired smooth pursuit, can act as important diagnostic indicators for various neurological conditions. The non-invasive nature, cost-effectiveness, and ease of implementation of modern eye-tracking systems makes it particularly attractive in both clinical and research settings. Advanced digital eye-tracking technologies and analytical methods enable precise quantification of eye-movement parameters, complementing subjective clinical evaluations with objective data. This review examines how DEMOs could contribute to the localisation and diagnosis of neural impairments, potentially serving as useful biomarkers. By comprehensively exploring the role of eye-movement assessment, this review aims to highlight the common eye-movement deficits seen in neurological injury and disease by using the examples of mild traumatic brain injury and Parkinson’s Disease. This review also aims to enhance the understanding of the potential use of DEMOs in diagnosis, monitoring, and management of neurological disorders, ultimately improving patient care and deepening our understanding of complex neurological processes. Furthermore, we consider the broader implications of this technology in unravelling the complexities of visual processing, attention mechanisms, and cognitive functions. This review summarises how DEMOs could reshape our understanding of brain health and allow for more targeted and effective neurological interventions. Full article
Show Figures

Figure 1

27 pages, 9069 KiB  
Article
Forecasting Human Core and Skin Temperatures: A Long-Term Series Approach
by Xinge Han, Jiansong Wu, Zhuqiang Hu, Chuan Li and Boyang Sun
Big Data Cogn. Comput. 2024, 8(12), 197; https://doi.org/10.3390/bdcc8120197 - 19 Dec 2024
Viewed by 254
Abstract
Human core and skin temperature (Tcr and Tsk) are crucial indicators of human health and are commonly utilized in diagnosing various types of diseases. This study presents a deep learning model that combines a long-term series forecasting method with transfer [...] Read more.
Human core and skin temperature (Tcr and Tsk) are crucial indicators of human health and are commonly utilized in diagnosing various types of diseases. This study presents a deep learning model that combines a long-term series forecasting method with transfer learning techniques, capable of making precise, personalized predictions of Tcr and Tsk in high-temperature environments with only a small corpus of actual training data. To practically validate the model, field experiments were conducted in complex environments, and a thorough analysis of the effects of three diverse training strategies on the overall performance of the model was performed. The comparative analysis revealed that the optimized training method significantly improved prediction accuracy for forecasts extending up to 10 min into the future. Specifically, the approach of pretraining the model on in-distribution samples followed by fine-tuning markedly outperformed other methods in terms of prediction accuracy, with a prediction error for Tcr within ±0.14 °C and Tsk, mean within ±0.46 °C. This study provides a viable approach for the precise, real-time prediction of Tcr and Tsk, offering substantial support for advancing early warning research of human thermal health. Full article
Show Figures

Figure 1

21 pages, 1364 KiB  
Article
Arabic Opinion Classification of Customer Service Conversations Using Data Augmentation and Artificial Intelligence
by Rihab Fahd Al-Mutawa and Arwa Yousuf Al-Aama
Big Data Cogn. Comput. 2024, 8(12), 196; https://doi.org/10.3390/bdcc8120196 - 19 Dec 2024
Viewed by 280
Abstract
Customer satisfaction is not just a significant factor but a cornerstone for smart cities and their organizations that offer services to people. It enhances the organization’s reputation and profitability and drastically raises the chances of returning customers. Unfortunately, customer support service through online [...] Read more.
Customer satisfaction is not just a significant factor but a cornerstone for smart cities and their organizations that offer services to people. It enhances the organization’s reputation and profitability and drastically raises the chances of returning customers. Unfortunately, customer support service through online chat is often not rated by customers to help improve the service. This study employs artificial intelligence and data augmentation to predict customer satisfaction ratings from conversations by analyzing the responses of customers and service providers. For the study, the authors obtained actual conversations between customers and real agents from the call center database of Jeddah Municipality that were rated by customers on a scale of 1–5. They trained and tested five prediction models with approaches based on logistic regression, random forest, and ensemble-based deep learning, and fine-tuned two pre-trained recent models: ArabicT5 and SaudiBERT. Then, they repeated training and testing models after applying a data augmentation technique using the generative artificial intelligence, GPT-4, to improve the unbalance in customer conversation data. The study found that the ensemble-based deep learning approach best predicts the five-, three-, and two-class classifications. Moreover, data augmentation improved accuracy using the ensemble-based deep learning model with a 1.69% increase and the logistic regression model with a 3.84% increase. This study contributes to the advancement of Arabic opinion mining, as it is the first to report the performance of determining customer satisfaction levels using Arabic conversation data. The implications of this study are significant, as the findings can be applied to improve customer service in various organizations. Full article
Show Figures

Figure 1

13 pages, 552 KiB  
Article
Mandarin Recognition Based on Self-Attention Mechanism with Deep Convolutional Neural Network (DCNN)-Gated Recurrent Unit (GRU)
by Xun Chen, Chengqi Wang, Chao Hu and Qin Wang
Big Data Cogn. Comput. 2024, 8(12), 195; https://doi.org/10.3390/bdcc8120195 - 18 Dec 2024
Viewed by 448
Abstract
Speech recognition technology is an important branch in the field of artificial intelligence, aiming to transform human speech into computer-readable text information. However, speech recognition technology still faces many challenges, such as noise interference, and accent and speech rate differences. An aim of [...] Read more.
Speech recognition technology is an important branch in the field of artificial intelligence, aiming to transform human speech into computer-readable text information. However, speech recognition technology still faces many challenges, such as noise interference, and accent and speech rate differences. An aim of this paper is to explore a deep learning-based speech recognition method to improve the accuracy and robustness of speech recognition. Firstly, this paper introduces the basic principles of speech recognition and existing mainstream technologies, and then focuses on the deep learning-based speech recognition method. Through comparative experiments, it is found that the self-attention mechanism performs best in speech recognition tasks. In order to further improve speech recognition performance, this paper proposes a deep learning model based on the self-attention mechanism with DCNN-GRU. The model realizes the dynamic attention to an input speech by introducing the self-attention mechanism in a neural network model instead of an RNN and with a deep convolutional neural network, which improves the robustness and recognition accuracy of this model. This experiment uses 170 h of Chinese dataset AISHELL-1. Compared with the deep convolutional neural network, the deep learning model based on the self-attention mechanism with DCNN-GRU accomplishes a reduction of at least 6% in CER. Compared with a bidirectional gated recurrent neural network, the deep learning model based on the self-attention mechanism with DCNN-GRU accomplishes a reduction of 0.7% in CER. And finally, this experiment is performed on a test set analyzed the influencing factors affecting the CER. The experimental results show that this model exhibits good performance in various noise environments and accent conditions. Full article
Show Figures

Figure 1

16 pages, 953 KiB  
Article
Assessing the Guidelines on the Use of Generative Artificial Intelligence Tools in Universities: A Survey of the World’s Top 50 Universities
by Midrar Ullah, Salman Bin Naeem and Maged N. Kamel Boulos
Big Data Cogn. Comput. 2024, 8(12), 194; https://doi.org/10.3390/bdcc8120194 - 18 Dec 2024
Viewed by 561
Abstract
The widespread adoption of Generative Artificial Intelligence (GenAI) tools in higher education has necessitated the development of appropriate and ethical usage guidelines. This study aims to explore and assess publicly available guidelines covering the use of GenAI tools in universities, following a predefined [...] Read more.
The widespread adoption of Generative Artificial Intelligence (GenAI) tools in higher education has necessitated the development of appropriate and ethical usage guidelines. This study aims to explore and assess publicly available guidelines covering the use of GenAI tools in universities, following a predefined checklist. We searched and downloaded publicly accessible guidelines on the use of GenAI tools from the websites of the top 50 universities globally, according to the 2025 QS university rankings. From the literature on GenAI use guidelines, we created a 24-item checklist, which was then reviewed by a panel of experts. This checklist was used to assess the characteristics of the retrieved university guidelines. Out of the 50 university websites explored, guidelines were publicly accessible on the sites of 41 institutions. All these guidelines allowed for the use of GenAI tools in academic settings provided that specific instructions detailed in the guidelines were followed. These instructions encompassed securing instructor consent before utilization, identifying appropriate and inappropriate instances for deployment, employing suitable strategies in classroom settings and assessment, appropriately integrating results, acknowledging and crediting GenAI tools, and adhering to data privacy and security measures. However, our study found that only a small number of the retrieved guidelines offered instructions on the AI algorithm (understanding how it works), the documentation of prompts and outputs, AI detection tools, and mechanisms for reporting misconduct. Higher education institutions should develop comprehensive guidelines and policies for the responsible use of GenAI tools. These guidelines must be frequently updated to stay in line with the fast-paced evolution of AI technologies and their applications within the academic sphere. Full article
Show Figures

Figure 1

16 pages, 1558 KiB  
Article
EIF-SlideWindow: Enhancing Simultaneous Localization and Mapping Efficiency and Accuracy with a Fixed-Size Dynamic Information Matrix
by Javier Lamar Léon, Pedro Salgueiro, Teresa Gonçalves and Luis Rato
Big Data Cogn. Comput. 2024, 8(12), 193; https://doi.org/10.3390/bdcc8120193 - 17 Dec 2024
Viewed by 309
Abstract
This paper introduces EIF-SlideWindow, a novel enhancement of the Extended Information Filter (EIF) algorithm for Simultaneous Localization and Mapping (SLAM). Traditional EIF-SLAM, while effective in many scenarios, struggles with inaccuracies in highly non-linear systems or environments characterized by significant non-Gaussian noise. Moreover, the [...] Read more.
This paper introduces EIF-SlideWindow, a novel enhancement of the Extended Information Filter (EIF) algorithm for Simultaneous Localization and Mapping (SLAM). Traditional EIF-SLAM, while effective in many scenarios, struggles with inaccuracies in highly non-linear systems or environments characterized by significant non-Gaussian noise. Moreover, the computational complexity of EIF/EKF-SLAM scales with the size of the environment, often resulting in performance bottlenecks. Our proposed EIF-SlideWindow approach addresses these limitations by maintaining a fixed-size information matrix and vector, ensuring constant-time processing per robot step, regardless of trajectory length. This is achieved through a sliding window mechanism centered on the robot’s pose, where older landmarks are systematically replaced by newer ones. We assess the effectiveness of EIF-SlideWindow using simulated data and demonstrate that it outperforms standard EIF/EKF-SLAM in both accuracy and efficiency. Additionally, our implementation leverages PyTorch for matrix operations, enabling efficient execution on both CPU and GPU. Additionally, the code for this approach is made available for further exploration and development. Full article
Show Figures

Figure 1

21 pages, 5273 KiB  
Article
Integrating Statistical Methods and Machine Learning Techniques to Analyze and Classify COVID-19 Symptom Severity
by Yaqeen Raddad, Ahmad Hasasneh, Obada Abdallah, Camil Rishmawi and Nouar Qutob
Big Data Cogn. Comput. 2024, 8(12), 192; https://doi.org/10.3390/bdcc8120192 - 16 Dec 2024
Viewed by 821
Abstract
Background/Objectives: The COVID-19 pandemic, caused by Severe Acute Respiratory Syndrome Coronavirus 2 (SARS-CoV-2), led to significant global health challenges, including the urgent need for accurate symptom severity prediction aimed at optimizing treatment. While machine learning (ML) and deep learning (DL) models have [...] Read more.
Background/Objectives: The COVID-19 pandemic, caused by Severe Acute Respiratory Syndrome Coronavirus 2 (SARS-CoV-2), led to significant global health challenges, including the urgent need for accurate symptom severity prediction aimed at optimizing treatment. While machine learning (ML) and deep learning (DL) models have shown promise in predicting COVID-19 severity using imaging and clinical data, there is limited research utilizing comprehensive tabular symptom datasets. This study aims to address this gap by leveraging a detailed symptom dataset to develop robust models for categorizing COVID-19 symptom severity, thereby enhancing clinical decision making. Methods: A unique tabular dataset was created using questionnaire responses from 5654 individuals, including demographic information, comorbidities, travel history, and medical data. Both unsupervised and supervised ML techniques were employed, including k-means clustering to categorize symptom severity into mild, moderate, and severe clusters. In addition, classification models, namely, Support Vector Machine (SVM), Adaptive Boosting (AdaBoost), eXtreme Gradient Boosting (XGBoost), random forest, and a deep neural network (DNN) were used to predict symptom severity levels. Feature importance was analyzed using the random forest model for its robustness with high-dimensional data and ability to capture complex non-linear relationships, and statistical significance was evaluated through ANOVA and Chi-square tests. Results: Our study showed that fatigue, joint pain, and headache were the most important features in predicting severity. SVM, AdaBoost, and random forest achieved an accuracy of 94%, while XGBoost achieved an accuracy of 96%. DNN showed robust performance in handling complex patterns with 98% accuracy. In terms of precision and recall metrics, both the XGBoost and DNN models demonstrated robust performance, particularly for the moderate class. XGBoost recorded 98% precision and 97% recall, while DNN achieved 99% precision and recall. The clustering approach improved classification accuracy by reducing noise and dimensionality. Statistical tests confirmed the significance of additional features like Body Mass Index (BMI), age, and dominant variant type. Conclusions: Integrating symptom data with advanced ML models offers a promising approach for accurate COVID-19 severity classification. This method provides a reliable tool for healthcare professionals to optimize patient care and resource management, particularly in managing COVID-19 and potential future pandemics. Future work should focus on incorporating imaging and clinical data to further enhance model accuracy and clinical applicability. Full article
Show Figures

Figure 1

25 pages, 6393 KiB  
Article
Re-Evaluating Deep Learning Attacks and Defenses in Cybersecurity Systems
by Meaad Ahmed, Qutaiba Alasad, Jiann-Shiun Yuan and Mohammed Alawad
Big Data Cogn. Comput. 2024, 8(12), 191; https://doi.org/10.3390/bdcc8120191 - 16 Dec 2024
Viewed by 471
Abstract
Cybersecurity attacks pose a significant threat to the security of network systems through intrusions and illegal communications. Measuring the vulnerability of cybersecurity is crucial for refining the overall system security to further mitigate potential security risks. Machine learning (ML)-based intrusion detection systems (IDSs) [...] Read more.
Cybersecurity attacks pose a significant threat to the security of network systems through intrusions and illegal communications. Measuring the vulnerability of cybersecurity is crucial for refining the overall system security to further mitigate potential security risks. Machine learning (ML)-based intrusion detection systems (IDSs) are mainly designed to detect malicious network traffic. Unfortunately, ML models have recently been demonstrated to be vulnerable to adversarial perturbation, and therefore enable potential attackers to crash the system during normal operation. Among different attacks, generative adversarial networks (GANs) have been known as one of the most powerful threats to cybersecurity systems. To address these concerns, it is important to explore new defense methods and understand the nature of different types of attacks. In this paper, we investigate four serious attacks, GAN, Zeroth-Order Optimization (ZOO), kernel density estimation (KDE), and DeepFool attacks, on cybersecurity. Deep analysis was conducted on these attacks using three different cybersecurity datasets, ADFA-LD, CSE-CICIDS2018, and CSE-CICIDS2019. Our results have shown that KDE and DeepFool attacks are stronger than GANs in terms of attack success rate and impact on system performance. To demonstrate the effectiveness of our approach, we develop a defensive model using adversarial training where the DeepFool method is used to generate adversarial examples. The model is evaluated against GAN, ZOO, KDE, and DeepFool attacks to assess the level of system protection against adversarial perturbations. The experiment was conducted by leveraging a deep learning model as a classifier with the three aforementioned datasets. The results indicate that the proposed defensive model refines the resilience of the system and mitigates the presented serious attacks. Full article
Show Figures

Figure 1

16 pages, 1694 KiB  
Article
Comparative Study of Filtering Methods for Scientific Research Article Recommendations
by Driss El Alaoui, Jamal Riffi, Abdelouahed Sabri, Badraddine Aghoutane, Ali Yahyaouy and Hamid Tairi
Big Data Cogn. Comput. 2024, 8(12), 190; https://doi.org/10.3390/bdcc8120190 - 16 Dec 2024
Viewed by 282
Abstract
Given the daily influx of scientific publications, researchers often face challenges in identifying relevant content amid the vast volume of available information, typically resorting to conventional methods like keyword searches or manual browsing. Utilizing a dataset comprising 1895 users and 3122 articles from [...] Read more.
Given the daily influx of scientific publications, researchers often face challenges in identifying relevant content amid the vast volume of available information, typically resorting to conventional methods like keyword searches or manual browsing. Utilizing a dataset comprising 1895 users and 3122 articles from the CI&T Deskdrop collection, as well as 7947 users and 25,975 articles from CiteULike-t, we examine the effectiveness of collaborative filtering and content-based and hybrid recommendation approaches in scientific literature recommendations. These methods automatically generate article suggestions by analyzing user preferences and historical behavior. Our findings, evaluated based on accuracy (Precision@K), ranking quality (NDCG@K), and novelty, reveal that the hybrid approach significantly outperforms other methods, tackling some challenges such as cold starts and sparsity problems. This research offers theoretical insights into recommendation model effectiveness and practical implications for developing tools that enhance content discovery and researcher productivity. Full article
Show Figures

Figure 1

23 pages, 4791 KiB  
Article
An Intelligent Self-Validated Sensor System Using Neural Network Technologies and Fuzzy Logic Under Operating Implementation Conditions
by Serhii Vladov, Victoria Vysotska, Valerii Sokurenko, Oleksandr Muzychuk and Lyubomyr Chyrun
Big Data Cogn. Comput. 2024, 8(12), 189; https://doi.org/10.3390/bdcc8120189 - 13 Dec 2024
Viewed by 415
Abstract
This article presents an intelligent self-validated sensor system developed for dynamic objects and based on the intelligent sensor concept, which ensures autonomous data collection and real-time analysis while adapting to changing conditions and compensating for errors. The research’s scientific merit is that an [...] Read more.
This article presents an intelligent self-validated sensor system developed for dynamic objects and based on the intelligent sensor concept, which ensures autonomous data collection and real-time analysis while adapting to changing conditions and compensating for errors. The research’s scientific merit is that an intelligent self-validated sensor for dynamic objects has been developed that integrates adaptive correction algorithms, fuzzy logic, and neural networks to improve the sensors’ accuracy and reliability under changing operating conditions. The proposed intelligent self-validated sensor system provides real-time error compensation, long-term stability, and effective fault diagnostics. Analytical equations are described, considering corrections related to influencing factors, temporal drift, and calibration characteristics, significantly enhancing measurement accuracy and reliability. The fuzzy logic application allows for refining the scaling coefficient that adjusts the relationship between the measured parameter and influencing factors, utilizing fuzzy inference algorithms. Additionally, monitoring and diagnostics implementation for sensor states through LSTM networks enable effective fault detection. Computational experiments on the TV3-117 engine demonstrated high data-restoring accuracy during forced interruptions, reaching 99.5%. A comparative analysis with alternative approaches confirmed the advantages of using LSTM (Long Short-Term Memory) neural networks in improving measurement quality. Full article
Show Figures

Figure 1

42 pages, 1236 KiB  
Systematic Review
Predictive Models for Educational Purposes: A Systematic Review
by Ahlam Almalawi, Ben Soh, Alice Li and Halima Samra
Big Data Cogn. Comput. 2024, 8(12), 187; https://doi.org/10.3390/bdcc8120187 - 13 Dec 2024
Viewed by 641
Abstract
This systematic literature review evaluates predictive models in education, focusing on their role in forecasting student performance, identifying at-risk students, and personalising learning experiences. The review compares the effectiveness of machine learning (ML) algorithms such as Support Vector Machines (SVMs), Artificial Neural Networks [...] Read more.
This systematic literature review evaluates predictive models in education, focusing on their role in forecasting student performance, identifying at-risk students, and personalising learning experiences. The review compares the effectiveness of machine learning (ML) algorithms such as Support Vector Machines (SVMs), Artificial Neural Networks (ANNs), and Decision Trees with traditional statistical models, assessing their ability to manage complex educational data and improve decision-making. The search, conducted across databases including ScienceDirect, IEEE Xplore, ACM Digital Library, and Google Scholar, yielded 400 records. After screening and removing duplicates, 124 studies were included in the final review. The findings show that ML algorithms consistently outperform traditional models due to their capacity to handle large, non-linear datasets and continuously enhance predictive accuracy as new patterns emerge. These models effectively incorporate socio-economic, demographic, and academic data, making them valuable tools for improving student retention and performance. However, the review also identifies key challenges, including the risk of perpetuating biases present in historical data, issues of transparency, and the complexity of interpreting AI-driven decisions. In addition, reliance on varying data processing methods across studies reduces the generalisability of current models. Future research should focus on developing more transparent, interpretable, and equitable models while standardising data collection and incorporating non-traditional variables, such as cognitive and motivational factors. Ensuring transparency and ethical standards in handling student data is essential for fostering trust in AI-driven models. Full article
Show Figures

Figure 1

23 pages, 2292 KiB  
Article
Integrating Generative AI in Hackathons: Opportunities, Challenges, and Educational Implications
by Ramteja Sajja, Carlos Erazo Ramirez, Zhouyayan Li, Bekir Z. Demiray, Yusuf Sermet and Ibrahim Demir
Big Data Cogn. Comput. 2024, 8(12), 188; https://doi.org/10.3390/bdcc8120188 - 13 Dec 2024
Viewed by 556
Abstract
Hackathons have become essential in the software industry, fostering innovation and skill development for both organizations and students. These events facilitate rapid prototyping for companies while providing students with hands-on learning opportunities that bridge theory and practice. Over time, hackathons have evolved from [...] Read more.
Hackathons have become essential in the software industry, fostering innovation and skill development for both organizations and students. These events facilitate rapid prototyping for companies while providing students with hands-on learning opportunities that bridge theory and practice. Over time, hackathons have evolved from competitive arenas into dynamic educational platforms, promoting collaboration between academia and industry. The integration of artificial intelligence (AI) and machine learning is transforming hackathons, enhancing learning experiences, and introducing ethical considerations. This study examines the impact of generative AI tools on technological decision-making during the 2023 University of Iowa Hackathon. It analyzes how AI influences project efficiency, learning outcomes, and collaboration, while addressing the ethical challenges posed by its use. The findings offer actionable insights and strategies for effectively integrating AI into future hackathons, balancing innovation, ethics, and educational value. Full article
Show Figures

Figure 1

28 pages, 5843 KiB  
Article
An Analysis of Vaccine-Related Sentiments on Twitter (X) from Development to Deployment of COVID-19 Vaccines
by Rohitash Chandra, Jayesh Sonawane and Jahnavi Lande
Big Data Cogn. Comput. 2024, 8(12), 186; https://doi.org/10.3390/bdcc8120186 - 13 Dec 2024
Viewed by 474
Abstract
Anti-vaccine sentiments have been well-known and reported throughout the history of viral outbreaks and vaccination programmes. The COVID-19 pandemic caused fear and uncertainty about vaccines, which has been well expressed on social media platforms such as Twitter (X). We analyse sentiments from the [...] Read more.
Anti-vaccine sentiments have been well-known and reported throughout the history of viral outbreaks and vaccination programmes. The COVID-19 pandemic caused fear and uncertainty about vaccines, which has been well expressed on social media platforms such as Twitter (X). We analyse sentiments from the beginning of the COVID-19 pandemic and study the public behaviour on X during the planning, development, and deployment of vaccines expressed in tweets worldwide using a sentiment analysis framework via deep learning models. We provide visualisation and analysis of anti-vaccine sentiments throughout the COVID-19 pandemic. We review the nature of the sentiments expressed with the number of tweets and monthly COVID-19 infections. Our results show a link between the number of tweets, the number of cases, and the change in sentiment polarity scores during major waves of COVID-19. We also find that the first half of the pandemic had drastic changes in the sentiment polarity scores that later stabilised, implying that the vaccine rollout impacted the nature of discussions on social media. Full article
(This article belongs to the Special Issue Application of Semantic Technologies in Intelligent Environment)
Show Figures

Figure 1

22 pages, 487 KiB  
Article
From Fact Drafts to Operational Systems: Semantic Search in Legal Decisions Using Fact Drafts
by Gergely Márk Csányi, Dorina Lakatos, István Üveges, Andrea Megyeri , János Pál Vadász, Dániel Nagy and Renátó Vági
Big Data Cogn. Comput. 2024, 8(12), 185; https://doi.org/10.3390/bdcc8120185 - 10 Dec 2024
Viewed by 506
Abstract
This research paper presents findings from an investigation in the semantic similarity search task within the legal domain, using a corpus of 1172 Hungarian court decisions. The study establishes the groundwork for an operational semantic similarity search system designed to identify cases with [...] Read more.
This research paper presents findings from an investigation in the semantic similarity search task within the legal domain, using a corpus of 1172 Hungarian court decisions. The study establishes the groundwork for an operational semantic similarity search system designed to identify cases with comparable facts using preliminary legal fact drafts. Evaluating such systems often poses significant challenges, given the need for thorough document checks, which can be costly and limit evaluation reusability. To address this, the study employs manually created fact drafts for legal cases, enabling reliable ranking of original cases within retrieved documents and quantitative comparison of various vectorization methods. The study compares twelve different text embedding solutions (the most recent became available just a few weeks before the manuscript was written) identifying Cohere’s embed-multilingual-v3.0, Beijing Academy of Artificial Intelligence’s bge-m3, Jina AI’s jina-embeddings-v3, OpenAI’s text-embedding-3-large, and Microsoft’s multilingual-e5-large models as top performers. To overcome the transformer-based models’ context window limitation, we investigated chunking, striding, and last chunk scaling techniques, with last chunk scaling significantly improving embedding quality. The results suggest that the effectiveness of striding varies based on token count. Notably, employing striding with 16 tokens yielded optimal results, representing 3.125% of the context window size for the best-performing models. Results also suggested that from the models having 8192 token long context window the bge-m3 model is superior compared to jina-embeddings-v3 and text-embedding-3-large models in capturing the relevant parts of a document if the text contains significant amount of noise. The validity of the approach was evaluated and confirmed by legal experts. These insights led to an operational semantic search system for a prominent legal content provider. Full article
Show Figures

Figure 1

15 pages, 5809 KiB  
Article
The Use of Eye-Tracking to Explore the Relationship Between Consumers’ Gaze Behaviour and Their Choice Process
by Maria-Jesus Agost and Vicente Bayarri-Porcar
Big Data Cogn. Comput. 2024, 8(12), 184; https://doi.org/10.3390/bdcc8120184 - 9 Dec 2024
Viewed by 463
Abstract
Eye-tracking technology can assist researchers in understanding motivational decision-making and choice processes by analysing consumers’ gaze behaviour. Previous studies showed that attention is related to decision, as the preferred stimulus is generally the most observed and the last visited before a decision is [...] Read more.
Eye-tracking technology can assist researchers in understanding motivational decision-making and choice processes by analysing consumers’ gaze behaviour. Previous studies showed that attention is related to decision, as the preferred stimulus is generally the most observed and the last visited before a decision is made. In this work, the relationship between gaze behaviour and decision-making was explored using eye-tracking technology. Images of six wardrobes incorporating different sustainable design strategies were presented to 57 subjects, who were tasked with selecting the wardrobe they intended to keep the longest. The amount of time spent looking was higher when it was the chosen version. Detailed analyses of gaze plots and heat maps derived from eye-tracking records were employed to identify different patterns of gaze behaviour during the selection process. These patterns included alternating attention between a few versions or comparing them against a reference, allowing the identification of stimuli that initially piqued interest but were ultimately not chosen, as well as potential doubts in the decision-making process. These findings suggest that doubts that arise before making a selection warrant further investigation. By identifying stimuli that attract attention but are not chosen, this study provides valuable insights into consumer behaviour and decision-making processes. Full article
Show Figures

Figure 1

Back to TopTop