12th International Conference on Signal and Image Processing (Signal 2025)

May 17~ 18, 2025 Zurich, Switzerland

Accepted Papers


Autoencoder for Image Classification with Genetics Algorithms

John Tsiligaridis, Department of Math and Computer Science, Heritage University, Toppenish, WA, USA

ABSTRACT

Autoencoders (AEs) are Deep Learning (DL) models that are well known for their ability to compress and reconstruct data. When an AE compresses input data, a latent space is created which yields a compressed representation of the original data with a smaller set of features. Genetic Algorithms (GAs) based on evolutionary principles can be used to optimize various hyperparameters of a DL model. This work involves two tasks. First, it focuses on the application of an AE on image data along with various configurations of the AE structure and its constituent encoder/decoder structure using Multi-Layer Perceptrons (MLPs). Visualizations of the AE loss functions during training are provided, along with various latent space results obtained using clustering techniques. The second focus of the paper is on the application of the GA on a Convolutional AE where optimization of the Convolutional Neural Networks (CNN) encoder/decoder structures is done by converting the architecture into genes for image classification. We see that the AE is a flexible and robust model that can successfully be applied on a variety of image datasets and the GA model surpasses the AE model.

Keywords

Machine Learning, Deep Learning, Autoencoders, Genetic Algorithms.


Transforming E-Commerce With AI Agents: A Technical Deep Dive

Sai Kiran Padmam, Partha Sarthi Samal,independent,United States of America

ABSTRACT

Online retail has come a long way since its early days of static web pages and manual price comparisons. The new frontier embraces artificial intelligence (AI) to interpret user queries, mediate real-time auctions among multiple vendors, and deliver personalized recommendations at blazing speeds. This paper highlights how such AI agents operate under the hood, drawing upon machine learning, reinforcement learning, and multi-agent coordination principles. We also offer glimpses into emerging research challenges and future directions that may reshape online shopping entirely.


Rethinking Requirement Analysis for AI-based Projects

Andreas Shaji Jacob, Kulwant Singh, Muhammad Shafique, Ruben Movsisyan, Seungbin Lee, and Ugur Randa, Pacific States University, USA

ABSTRACT

The rapid adoption of Artificial Intelligence (AI) across industries has revealed limitations17 in traditional requirement analysis methodologies, which were not designed to address the com-18 plexities and iterative nature of AI-based projects. This paper proposes a refined thought19 process for requirement analysis tailored to the needs of AI-driven initiatives, whether AI is the20 primary focus or an integrated component of a larger system. By emphasizing the dynamic21 interplay between data, models, and deployment environments, the proposed approach departs22 from linear methodologies, advocating for an adaptive and iterative process. Using case studies,23 we demonstrate how this concept ensures better alignment with business goals, enhances data24 utility, and improves model performance while addressing ethical considerations and practical25 constraints. This paper aims to provide practitioners, researchers, and project owners with26 actionable insights to optimize AI project outcomes in an increasingly complex technological27 landscape.


Issues and Challenges in Implementing Agile Scrum Methodology in Different Organizations: Finding Trends and Differences

Lutfur Rahman Fahad, Mukit AL Elahi, Nayem Miah, Himel, Somon, Niaz Dhara, Mia, and Adipta, Department of Computer Science, Pacific States University, USA

ABSTRACT

Agile Scrum methodology is widely regarded as a transformative approach to software de velopment, emphasizing flexibility, collaboration, and iterative progress. However, its adoption is not without challenges, which vary significantly across organizations of different sizes, in-dustries, and geographic distributions. This study explores recurring issues such as resource constraints, communication barriers, and resistance to change, while highlighting trends and successful practices in addressing these challenges. By synthesizing insights from existing lit-erature and limited interviews, this research aims to provide actionable recommendations and a tailored framework for organizations seeking to optimize their Agile Scrum implementation.The findings contribute to a deeper understanding of how industry-specific contexts influence the effectiveness of Agile practices


Edge Computing in Cloud Computing

Jannatul Mawa, Md Nafis Azad Nobel, Sajeet Raj Aryal, Kazi, Taehyun Kim, Pritom Das, Mahbu Khan, Department of Computer Science, Pacific States, University, USA

ABSTRACT

The cloud computing environment offers significant flexibility and access to computing re- sources at a reduced cost. This technology is rapidly transforming the landscape of e-services across various fields. In this paper, we examine cloud computing services and applications, high- lighting examples of services offered by leading Cloud Service Providers (CSPs) such as Google, Microsoft, Amazon, HP, and Salesforce. We also showcase innovative cloud applications in areas like e-learning, Enterprise Resource Planning (ERP), and e-governance. This study aims to help individuals and organizations recognize how cloud computing can deliver customized, reliable, and cost-effective solutions across a diverse range of applications.


From Genes to Insights: A Causal Framework for Diabetes Genomics and Predictive Analytics

Sheresh Zahoor1, Pietro Li`o1, Ga ̈el Dias2, and Mohammed Hasanuzzaman3, 1University of Cambridge, 2Normandie Univ, GREYC, 3Queens University Belfast

ABSTRACT

Diabetes is a global health crisis, demanding advanced genomic approaches to uncover molecular mechanisms and identify therapeutic targets. This study introduces the Genomic Causal Framework (GCF), a novel approach combining genomic data analysis, causal modeling, and predictive analytics to provide actionable insights into diabetes pathology. These include identifying potential therapeutic targets, such as CXCL8, S100A8, and COL1A1, implicated in chronic inflammation and complications like diabetic nephropathy. The framework also highlights regulatory genes, such as ROBO1 and FCGR2A, as upstream drivers of disease progression. Using the Diabetes genome dataset (GSE132831), we identify differentially expressed genes (DEGs) with pyDESeq2, stratifying upregulated and downregulated genes. These DEGs form the basis for constructing a protein-protein interaction (PPI) network, revealing critical functional pathways. The GCF framework integrates Causal Bayesian Networks (CBNs) and Probability Trees (PTrees) to move beyond prediction and enable causal reasoning. CBNs model causal relationships between genes and diabetic outcomes, while PTrees quantify their impact. Achieving 82.22% accuracy and 95% recall, GCF ensures reliable patient identification, with SHAP analysis enhancing interpretability and biological relevance. Its integration of causal reasoning with predictive analytics prioritises biologically relevant features for clinical and research applications. By bridging causal inference with functional genomics, this study advances biomarker discovery and therapeutic target identification, providing a powerful tool for precision medicine in Type 2 Diabetes. Unlike traditional machine learning, our approach enhances interpretability while uncovering critical insights into disease development and progression.

Keywords

Diabetes, Causal Bayesian Network, Probability Trees, Genomics.


How New Technology and ICTS can Be used in Case of Teaching Algorithms

Kostas Dimitrios1 and Kostas Ioannis2, 1National and Kapodistrian University of Athens, 2University of Piraeus

ABSTRACT

In this article we will analyse how Information and communication technologies can be used in teaching Algorithms Information and communication technologies use algorithms can help create numerous applications that can solve a wide variety of problems.by University students. There are many applications in Universities specifically in laboratories teachers and students use algorithms to solve numerical and algebraic system problems. In Partial differential equations, we use Numerical Analysis to solve a system of PDEs. In research, also most people use Algorithms to validate their theoretical findings. Students need to use suitable algorithms to solve problems. Nowadays many people use Algorithms to solve problems in their companies or in their organisation and in every context, algorithms are widely used. This work is about teaching algorithms used to solve problems using the numerical analysis methods.

Keywords

Algorithms, Innovative, technology, Analysis.


Examining Github’s Role in Filter List Maintenance: Challenges and Trends in Open-source Ad-Blocking and Privacy Protection

Mshabab Alrizah, Jazan University, Jazan, Saudi Arabia

ABSTRACT

EasyList is a widely used filter list that enhances online privacy and security by blocking tracking mechanisms, advertisements, and other unwanted web elements. As an open-source project, its sustainability is based on collaborative contributions, issue tracking, and continuous updates to address emerging challenges. GitHub plays a crucial role in facilitating the management of EasyList, offering tools for version control, issue resolution, and community-driven improvements. This study explores the complexities of maintaining EasyList on GitHub by analyzing multiple-year issue reports. Through data collection, trend analysis, and resolution efficiency evaluation, this research provides insight into contributor engagement, frequently reported domains, and the overall effectiveness of the maintenance process. The findings highlight the importance of community participation in maintaining EasyList, the ongoing need for adaptive strategies against evolving tracking and ad-serving techniques, and the broader implications for open-source project management.

Keywords

Ad-blocking, open-source maintenance, GitHub, tracking prevention, filter lists, crowdsourcing collaboration.


Predictive Policing as a Threat to Justice – Why Algorithms Should Serve Communities, Not Control Them

Theodora-Stavroula Korma, Department of Communication and Information studies, Rijkuniversiteit Groningen, Groningen, The Netherlands

ABSTRACT

Predictive policing, an algorithm-driven crime prevention initiative, claims to render the criminal justice system more effective and neutral. Yet, this essay argues that these algorithmic models reinforce system-level prejudices and unfairly focus on over marginalized populations while amplifying injustice. As these models draw from historical data covering four decades shaped by biased police operations, they can magnify racial profiling and harden social hierarchies. Furthermore, these systems lack of transparency and accountability has ethical consequences on surveillance, due process, and civil rights violations. In line with Design Justice principles, this paper calls for a redesign of predictive policing that is not about control by systems but the empowerment of communities. Instead of being used as enforcement tools, these algorithms must be redesigned to address root causes of social harm, promote equitable resource allocation, and engage communities in decision-making. Through participatory governance and moral algorithmic design, predictive technologies can serve justice rather than subvert it, so that communities are protected, not monitored.

Keywords

Predictive policing, algorithmic bias, systemic injustice, racial profiling, Design Justice.


Large Language Models in Clinical Advice: Evaluating Direct Generation, Retrieval Augmented Generation and Human Reference Answers

Iblal Rakha1 and Noorhan Abbas2, 1Oxford University Hospitals NHS Foundation Trust, Oxford, OX3 9DU, UK, 2University of Leeds, Woodhouse, Leeds, LS2 9JT, UK

ABSTRACT

The NHS faces mounting pressures, resulting in workforce attrition and growing care backlogs. Pharmacy services, critical for ensuring medication safety and effectiveness, are often overlooked in digital innovation efforts. This pilot study investigates the potential of Large Language Models (LLMs) to alleviate pharmacy pressures by answering clinical pharmaceutical queries. Two retrieval techniques were evaluated: Vanilla Retrieval Augmented Generation (RAG) and Graph RAG, supported by an external knowledge source designed specifically for this study. ChatGPT 4o without retrieval served as a control. Quantitative and qualitative evaluations were conducted, including expert human assessments for response accuracy, relevance, and safety. Results demonstrated that LLMs can generate high-quality responses. In expert evaluations, Vanilla RAG outperformed other models and even human reference answers for accuracy and risk. Graph RAG revealed challenges related to retrieval accuracy. Despite the promise of LLMs, hallucinations and the ambiguity around LLM evaluations in healthcare remain key barriers to clinical deployment. This pilot study underscores the im-portance of robust evaluation frameworks to ensure the safe integration of LLMs into clinical workflows. However, regulatory bodies have yet to catch up with the rapid pace of LLM development. Guidelines are urgently needed to address the issues of transparency, explainability, data protection, and validation, to facilitate the safe and effective deployment of LLMs in clinical practice.

Keywords

Large Language Model Evaluation, Retrieval Augmented Generation, Clinical Question Answering, Knowledge Graphs, Healthcare Artificial Intelligence.

Optimizing Retrieval-Augmented Generation for Electrical Engineering: A Case Study on ABB Circuit Breakers

Salahuddin Alawadhi1 and Noorhan Abbas2, 1Salahuddin Alawadhi University of Leeds Dubai, UAE, 2Noorhan Abbas School of Computer Science University of Leeds, United Kingdom

ABSTRACT

The integration of Retrieval-Augmented Generation (RAG) with Large Language Models (LLMs) has shown potential in providing precise, contextually relevant responses in knowledge-intensive domains. This study investigates the ap-plication of RAG for ABB circuit breakers, focusing on accuracy, reliability, and contextual relevance in high-stakes engineering environments. By leveraging tailored datasets, advanced embedding models, and optimized chunking strategies, the research addresses challenges in data retrieval and contextual alignment unique to engineering documentation. Key contributions include the development of a domain-specific dataset for ABB circuit breakers and the evaluation of three RAG pipelines: OpenAI GPT-4o, Cohere, and Anthropic Claude. Advanced chunking methods, such as paragraph-based and title-aware segmentation, are assessed for their impact on retrieval accuracy and response generation. Results demonstrate that while certain configurations achieve high precision and relevancy, limitations persist in ensuring factual faithfulness and completeness, critical in engineering contexts. This work underscores the need for iterative improvements in RAG systems to meet the stringent demands of electrical engineering tasks, including design, troubleshooting, and operational decision-making. The findings in this paper help advance research of AI in highly technical domains such as electrical engineering.

Keywords

Retrieval-Augmented Generation (RAG), Electrical Engineering, ABB Circuit Breakers, Chunking, Embeddings

Security Suggestions Based on Laws

Zhiyuan Liu, Computing and Communications School of Lancaster University

ABSTRACT

The essay begins by setting out a detailed scenario for the deployment of face recognition systems in public places. Based on this scenario, two statutes that companies need to focus on and a relevant legal case are critically discussed. The essay then integrates the two statutes into the scenario and makes critical recommendations for security design decisions, both managerial and technical, based on the legal requirements. The essay concludes with a summary of the findings and insights.

Keywords

Network Protocols, Wireless Network, Mobile Network, Virus, Worms &Trojon.

Optimizing Retrieval-Augmented Generation for Electrical Engineering: A Case Study on ABB Circuit Breakers

Salahuddin Alawadhi1 and Noorhan Abbas2, 1Salahuddin Alawadhi University of Leeds Dubai, UAE, 2Noorhan Abbas School of Computer Science University of Leeds, United Kingdom

ABSTRACT

The integration of Retrieval-Augmented Generation (RAG) with Large Language Models (LLMs) has shown potential in providing precise, contextually relevant responses in knowledge-intensive domains. This study investigates the ap-plication of RAG for ABB circuit breakers, focusing on accuracy, reliability, and contextual relevance in high-stakes engineering environments. By leveraging tailored datasets, advanced embedding models, and optimized chunking strategies, the research addresses challenges in data retrieval and contextual alignment unique to engineering documentation. Key contributions include the development of a domain-specific dataset for ABB circuit breakers and the evaluation of three RAG pipelines: OpenAI GPT-4o, Cohere, and Anthropic Claude. Advanced chunking methods, such as paragraph-based and title-aware segmentation, are assessed for their impact on retrieval accuracy and response generation. Results demonstrate that while certain configurations achieve high precision and relevancy, limitations persist in ensuring factual faithfulness and completeness, critical in engineering contexts. This work underscores the need for iterative improvements in RAG systems to meet the stringent demands of electrical engineering tasks, including design, troubleshooting, and operational decision-making. The findings in this paper help advance research of AI in highly technical domains such as electrical engineering.

Keywords

Retrieval-Augmented Generation (RAG), Electrical Engineering, ABB Circuit Breakers, Chunking, Embeddings

Ensemble Learning and Survival Analysis for Predictive Model of Non-alcoholic Fatty Liver Disease (Nafld)

Qi Huamei1 and Md Jahangir Alam2, 1School of Electronics Information Science, Central South University, Changsha, China, 2Department of Computer Science and Technology, Central South University, Changsha, China

ABSTRACT

The rise of Non-Alcoholic Fatty Liver Disease (NAFLD), associated with obesity and metabolic disorders, underscores the importance of developing precise prediction models for early identification. This research employs machine learning and survival analysis techniques to classify and forecast NAFLD using clinical and demographic data. The examined models include Decision Tree, Extra Trees, Random Forest (utilizing 10 estimators), and K-Nearest Neighbours (with K set to 3). For data preparation, KNN imputation was applied to address missing values, and MinMax scaling was used for standardization. Lasso regression (LassoCV) was implemented to select features and highlight significant variables to enhance model efficacy. Alongside classification models, the Kaplan-Meier estimator (KaplanMeierFitter) and Cox Proportional Hazards Model (CoxPHFitter) were utilized to evaluate patient survival rates and to pinpoint risk factors. The ensemble models, specifically the Extra Trees and Random Forest classifiers, surpassed the baseline Decision Tree (88.28) and KNN (91.56) models, achieving accuracies of 92.54 and 92.63, respectively. LassoCV contributed to improved feature significance, while survival analysis offered valuable insights into the progression of NAFLD. This study showcases the efficacy of ensemble methods and survival analysis in developing reliable and interpretable prediction models for NAFLD. Future research should aim to expand the dataset and incorporate additional clinical parameters.

Keywords

NAFLD Prediction, Ensemble Learning, LassoCV Feature Selection, Survival Analysis, Cox Proportional hazards, Kaplan-Meier Estimator.


From Rules to Learning: Hybridizing Knowledge-based Models for Pedestrian Trajectory Prediction

Messaoud MEZATI, Siham BEGGAA, Houria BENBOUBKEUR, Chahd BRAITHEL and Malak GHOULIA, Department of Computer Science and Information Technology, Kasdi Merbah University Ouargla, Algeria

ABSTRACT

Predicting pedestrian trajectories is a key challenge in intelligent transportation systems, robotics, and urban mobility, requiring models that balance accuracy, adaptability, and interpretability. Traditional Knowledge-Based (KB) models, including social force models, agent-based simulations, and reinforcement learning, offer structured decision-making but struggle with rapidly changing and complex environments. In contrast, Deep Learning (DL) techniques, such as LSTMs, Graph Neural Networks (GNNs), and Transformers, capture intricate movement patterns but often lack transparency. This study examines the hybridization of KB and DL models, integrating physics-based constraints with data-driven learning to enhance pedestrian behavior forecasting. A systematic classification of hybrid models is provided based on model structure, prediction tasks, AI integration, and real-world applications. Additionally, the study explores the potential of Reinforcement Learning (RL), Self-Supervised Learning, and Large Language Models (LLMs) in trajectory prediction. By bridging rule-based reasoning with adaptive learning, this work contributes to the development of safer, more flexible, and explainable pedestrian trajectory prediction models for applications in autonomous navigation, smart cities, and crowd management.

Keywords

Pedestrian Trajectory Prediction, Knowledge-Based Models, Deep Learning, Autonomous Driving, Explainable AI.


Lexical Based Methods for Malicious Urls Phishing Detection

Rachana S Potpelwar, U V Kulkarni, J M Waghmare, Shri Guru Gobind Singh Institute of Engineering and Technolgy, Computer Science and Engineering Department, Nanded, 431605, Maharashtra, India

ABSTRACT

Phishing attacks continue to pose a significant threat to online security, making efficient detection methods essential. This paper presents a lexical-based ap- proach for detecting malicious URLs using deep learning algorithms, including Artificial Neural Networks (ANN), Multi-Layer Perceptrons (MLP), and Long Short-Term Memory (LSTM) networks. Our dataset consists of phishing and le- gitimate URLs labeled accordingly. To enhance detection accuracy, the dataset was preprocessed using the Term Frequency-Inverse Document Frequency (TF- IDF) method, converting the raw URL strings into meaningful numerical rep- resentations. The experimental results demonstrate that preprocessing sub- stantially improves model performance. For LSTM, the accuracy improved from 90.05% (without preprocessing) to 90.77% (with preprocessing). These results highlight the effectiveness of combining lexical feature extraction with deep learning algorithms, offering a promising solution for real-time detection systems to safeguard against phishing attacks and enhance cybersecurity.”


'