Journal Description
Computers
Computers
is an international, scientific, peer-reviewed, open access journal of computer science, including computer and network architecture and computer–human interaction as its main foci, published monthly online by MDPI.
- Open Access— free for readers, with article processing charges (APC) paid by authors or their institutions.
- High Visibility: indexed within Scopus, ESCI (Web of Science), dblp, Inspec, and other databases.
- Journal Rank: JCR - Q2 (Computer Science, Interdisciplinary Applications) / CiteScore - Q2 (Computer Networks and Communications)
- Rapid Publication: manuscripts are peer-reviewed and a first decision is provided to authors approximately 17.2 days after submission; acceptance to publication is undertaken in 3.9 days (median values for papers published in this journal in the first half of 2024).
- Recognition of Reviewers: reviewers who provide timely, thorough peer-review reports receive vouchers entitling them to a discount on the APC of their next publication in any MDPI journal, in appreciation of the work done.
Impact Factor:
2.6 (2023);
5-Year Impact Factor:
2.4 (2023)
Latest Articles
Interactive Application as a Teaching Aid in Mechanical Engineering
Computers 2024, 13(7), 170; https://doi.org/10.3390/computers13070170 - 10 Jul 2024
Abstract
This paper examines the integration of interactive 3D applications into the teaching process in mechanical engineering education. An innovative interactive 3D application has been developed as a teaching aid for engineering students. The main advantage is its easy availability through a web browser
[...] Read more.
This paper examines the integration of interactive 3D applications into the teaching process in mechanical engineering education. An innovative interactive 3D application has been developed as a teaching aid for engineering students. The main advantage is its easy availability through a web browser on mobile devices or desktop computers. It includes four explorable 3D gearbox models with assembly animations, linked technical information, and immersive virtual and augmented reality (AR) experiences. The benefits of using this application in the teaching process were monitored on a group of students at the end of the semester. Assessments conducted before and after the use of the interactive 3D application measured learning outcomes. Qualitative feedback from students was also collected. The results demonstrated significant improvements in engagement, spatial awareness, and understanding of gearbox principles compared to traditional methods. The versatility and accessibility of the application also facilitated self-directed learning, reducing the need for external resources. These findings indicate that interactive 3D tools have the potential to enhance student learning and engagement and to promote sustainable practices in engineering education. Future research could explore the scalability and applicability of these tools across different engineering disciplines and educational contexts.
Full article
(This article belongs to the Special Issue Xtended or Mixed Reality (AR+VR) for Education 2024)
►
Show Figures
Open AccessArticle
Internet of Things-Based Robust Green Smart Grid
by
Rania A. Ahmed, M. Abdelraouf, Shaimaa Ahmed Elsaid, Mohammed ElAffendi, Ahmed A. Abd El-Latif, A. A. Shaalan and Abdelhamied A. Ateya
Computers 2024, 13(7), 169; https://doi.org/10.3390/computers13070169 - 8 Jul 2024
Abstract
Renewable energy sources play a critical role in all governments’ and organizations’ energy management and sustainability plans. The solar cell represents one such renewable energy resource, generating power in a population-free circumference. Integrating these renewable sources with the smart grids leads to the
[...] Read more.
Renewable energy sources play a critical role in all governments’ and organizations’ energy management and sustainability plans. The solar cell represents one such renewable energy resource, generating power in a population-free circumference. Integrating these renewable sources with the smart grids leads to the generation of green smart grids. Smart grids are critical for modernizing electricity distribution by using new communication technologies that improve power system efficiency, reliability, and sustainability. Smart grids assist in balancing supply and demand by allowing for real-time monitoring and administration, as well as accommodating renewable energy sources and reducing outages. However, their execution presents considerable problems. High upfront expenditures and the need for substantial and reliable infrastructure changes present challenges. Despite these challenges, shifting to green smart grids is critical for a resilient and adaptable energy future that can fulfill changing consumer demands and environmental aims. To this end, this work considers developing a reliable Internet of Things (IoT)-based green smart grid. The proposed green grid integrates traditional grids with solar energy and provides a control unit between the generation and consumption parts of the grid. The work deploys intelligent IoT units to control energy demands and manage energy consumption effectively. The proposed framework deploys the paradigm of distributed edge computing in four levels to provide efficient data offloading and power management. The developed green grid outperformed traditional grids in terms of its reliability and energy efficiency. The proposed green grid reduces energy consumption over the distribution area by an average of 24.3% compared to traditional grids.
Full article
(This article belongs to the Special Issue Feature Papers in Computers 2024)
►▼
Show Figures
Figure 1
Open AccessArticle
Transforming Digital Marketing with Generative AI
by
Tasin Islam, Alina Miron, Monomita Nandy, Jyoti Choudrie, Xiaohui Liu and Yongmin Li
Computers 2024, 13(7), 168; https://doi.org/10.3390/computers13070168 - 8 Jul 2024
Abstract
The current marketing landscape faces challenges in content creation and innovation, relying heavily on manually created content and traditional channels like social media and search engines. While effective, these methods often lack the creativity and uniqueness needed to stand out in a competitive
[...] Read more.
The current marketing landscape faces challenges in content creation and innovation, relying heavily on manually created content and traditional channels like social media and search engines. While effective, these methods often lack the creativity and uniqueness needed to stand out in a competitive market. To address this, we introduce MARK-GEN, a conceptual framework that utilises generative artificial intelligence (AI) models to transform marketing content creation. MARK-GEN provides a comprehensive, structured approach for businesses to employ generative AI in producing marketing materials, representing a new method in digital marketing strategies. We present two case studies within the fashion industry, demonstrating how MARK-GEN can generate compelling marketing content using generative AI technologies. This proposition paper builds on our previous technical developments in virtual try-on models, including image-based, multi-pose, and image-to-video techniques, and is intended for a broad audience, particularly those in business management.
Full article
(This article belongs to the Special Issue Deep Learning and Explainable Artificial Intelligence)
►▼
Show Figures
Figure 1
Open AccessArticle
Using Artificial Intelligence to Predict the Aerodynamic Properties of Wind Turbine Profiles
by
Ziemowit Malecha and Adam Sobczyk
Computers 2024, 13(7), 167; https://doi.org/10.3390/computers13070167 - 8 Jul 2024
Abstract
This study describes the use of artificial intelligence to predict the aerodynamic properties of wind turbine profiles. The goal was to determine the lift coefficient for an airfoil using its geometry as input. Calculations based on XFoil were taken as a target for
[...] Read more.
This study describes the use of artificial intelligence to predict the aerodynamic properties of wind turbine profiles. The goal was to determine the lift coefficient for an airfoil using its geometry as input. Calculations based on XFoil were taken as a target for the predictions. The lift coefficient for a single case scenario was set as a value to find by training an algorithm. Airfoil geometry data were collected from the UIUC Airfoil Data Site. Geometries in the coordinate format were converted to PARSEC parameters, which became a direct feature for the random forest regression algorithm. The training dataset included 60% of the base dataset records. The rest of the dataset was used to test the model. Five different datasets were tested. The results calculated for the test part of the base dataset were compared with the actual values of the lift coefficients. The developed prediction model obtained a coefficient of determination ranging from 0.83 to 0.87, which is a good prognosis for further research.
Full article
(This article belongs to the Special Issue Deep Learning and Explainable Artificial Intelligence)
►▼
Show Figures
Figure 1
Open AccessArticle
Hardware-Based Implementation of Algorithms for Data Replacement in Cache Memory of Processor Cores
by
Larysa Titarenko, Vyacheslav Kharchenko, Vadym Puidenko, Artem Perepelitsyn and Alexander Barkalov
Computers 2024, 13(7), 166; https://doi.org/10.3390/computers13070166 - 5 Jul 2024
Abstract
►▼
Show Figures
Replacement policies have an important role in the functioning of the cache memory of processor cores. The implementation of a successful policy allows us to increase the performance of the processor core and the computer system as a whole. Replacement policies are most
[...] Read more.
Replacement policies have an important role in the functioning of the cache memory of processor cores. The implementation of a successful policy allows us to increase the performance of the processor core and the computer system as a whole. Replacement policies are most often evaluated by the percentage of cache hits during the cycles of the processor bus when accessing the cache memory. The policies that focus on replacing the Least Recently Used (LRU) or Least Frequently Used (LFU) elements, whether instructions or data, are relevant for use. It should be noted that in the paging cache buffer, the above replacement policies can also be used to replace address information. The pseudo LRU (PLRU) policy introduces replacing based on approximate information about the age of the elements in the cache memory. The hardware implementation of any replacement policy algorithm is the circuit. This hardware part of the processor core has certain characteristics: the latency of the search process for a candidate element for replacement, the gate complexity, and the reliability. The characteristics of the PLRUt and PLRUm replacement policies are synthesized and investigated. Both are the varieties of the PLRU replacement policy, which is close to the LRU policy in terms of the percentage of cache hits. In the current study, the hardware implementation of these policies is evaluated, and the possibility of adaptation to each of the policies in the processor core according to a selected priority characteristic is analyzed. The dependency of the rise in the delay and gate complexity in the case of an increase in the associativity of the cache memory is shown. The advantage of the hardware implementation of the PLRUt algorithm in comparison with the PLRUm algorithm for higher values of associativity is shown.
Full article
Figure 1
Open AccessArticle
Automation Bias and Complacency in Security Operation Centers
by
Jack Tilbury and Stephen Flowerday
Computers 2024, 13(7), 165; https://doi.org/10.3390/computers13070165 - 3 Jul 2024
Cited by 1
Abstract
►▼
Show Figures
The volume and complexity of alerts that security operation center (SOC) analysts must manage necessitate automation. Increased automation in SOCs amplifies the risk of automation bias and complacency whereby security analysts become over-reliant on automation, failing to seek confirmatory or contradictory information. To
[...] Read more.
The volume and complexity of alerts that security operation center (SOC) analysts must manage necessitate automation. Increased automation in SOCs amplifies the risk of automation bias and complacency whereby security analysts become over-reliant on automation, failing to seek confirmatory or contradictory information. To identify automation characteristics that assist in the mitigation of automation bias and complacency, we investigated the current and proposed application areas of automation in SOCs and discussed its implications for security analysts. A scoping review of 599 articles from four databases was conducted. The final 48 articles were reviewed by two researchers for quality control and were imported into NVivo14. Thematic analysis was performed, and the use of automation throughout the incident response lifecycle was recognized, predominantly in the detection and response phases. Artificial intelligence and machine learning solutions are increasingly prominent in SOCs, yet support for the human-in-the-loop component is evident. The research culminates by contributing the SOC Automation Implementation Guidelines (SAIG), comprising functional and non-functional requirements for SOC automation tools that, if implemented, permit a mutually beneficial relationship between security analysts and intelligent machines. This is of practical value to human automation researchers and SOCs striving to optimize processes. Theoretically, a continued understanding of automation bias and its components is achieved.
Full article
Figure 1
Open AccessArticle
Integrity and Privacy Assurance Framework for Remote Healthcare Monitoring Based on IoT
by
Salah Hamza Alharbi, Ali Musa Alzahrani, Toqeer Ali Syed and Saad Said Alqahtany
Computers 2024, 13(7), 164; https://doi.org/10.3390/computers13070164 - 3 Jul 2024
Abstract
Remote healthcare monitoring (RHM) has become a pivotal component of modern healthcare, offering a crucial lifeline to numerous patients. Ensuring the integrity and privacy of the data generated and transmitted by IoT devices is of paramount importance. The integration of blockchain technology and
[...] Read more.
Remote healthcare monitoring (RHM) has become a pivotal component of modern healthcare, offering a crucial lifeline to numerous patients. Ensuring the integrity and privacy of the data generated and transmitted by IoT devices is of paramount importance. The integration of blockchain technology and smart contracts has emerged as a pioneering solution to fortify the security of internet of things (IoT) data transmissions within the realm of healthcare monitoring. In today’s healthcare landscape, the IoT plays a pivotal role in remotely monitoring and managing patients’ well-being. Furthermore, blockchain’s decentralized and immutable ledger ensures that all IoT data transactions are securely recorded, timestamped, and resistant to unauthorized modifications. This heightened level of data security is critical in healthcare, where the integrity and privacy of patient information are nonnegotiable. This research endeavors to harness the power of blockchain and smart contracts to establish a robust and tamper-proof framework for healthcare IoT data. Employing smart contracts, which are self-executing agreements programmed with predefined rules, enables us to automate and validate data transactions within the IoT ecosystem. These contracts execute automatically when specific conditions are met, eliminating the need for manual intervention and oversight. This automation not only streamlines the process of data processing but also enhances its accuracy and reliability by reducing the risk of human error. Additionally, smart contracts provide a transparent and tamper-proof mechanism for verifying the validity of transactions, thereby mitigating the risk of fraudulent activities. By leveraging smart contracts, organizations can ensure the integrity and efficiency of data transactions within the IoT ecosystem, leading to improved trust, transparency, and security. Our experiments demonstrate the application of a blockchain approach to secure transmissions in IoT for RHM, as will be illustrated in the paper. This showcases the practical applicability of blockchain technology in real-world scenarios.
Full article
(This article belongs to the Section Blockchain Infrastructures and Enabled Applications)
►▼
Show Figures
Figure 1
Open AccessArticle
Enhancing the Security of Classical Communication with Post-Quantum Authenticated-Encryption Schemes for the Quantum Key Distribution
by
Farshad Rahimi Ghashghaei, Yussuf Ahmed, Nebrase Elmrabit and Mehdi Yousefi
Computers 2024, 13(7), 163; https://doi.org/10.3390/computers13070163 - 1 Jul 2024
Abstract
This research aims to establish a secure system for key exchange by using post-quantum cryptography (PQC) schemes in the classic channel of quantum key distribution (QKD). Modern cryptography faces significant threats from quantum computers, which can solve classical problems rapidly. PQC schemes address
[...] Read more.
This research aims to establish a secure system for key exchange by using post-quantum cryptography (PQC) schemes in the classic channel of quantum key distribution (QKD). Modern cryptography faces significant threats from quantum computers, which can solve classical problems rapidly. PQC schemes address critical security challenges in QKD, particularly in authentication and encryption, to ensure the reliable communication across quantum and classical channels. The other objective of this study is to balance security and communication speed among various PQC algorithms in different security levels, specifically CRYSTALS-Kyber, CRYSTALS-Dilithium, and Falcon, which are finalists in the National Institute of Standards and Technology (NIST) Post-Quantum Cryptography Standardization project. The quantum channel of QKD is simulated with Qiskit, which is a comprehensive and well-supported tool in the field of quantum computing. By providing a detailed analysis of the performance of these three algorithms with Rivest–Shamir–Adleman (RSA), the results will guide companies and organizations in selecting an optimal combination for their QKD systems to achieve a reliable balance between efficiency and security. Our findings demonstrate that the implemented PQC schemes effectively address security challenges posed by quantum computers, while keeping the the performance similar to RSA.
Full article
(This article belongs to the Section ICT Infrastructures for Cybersecurity)
►▼
Show Figures
Figure 1
Open AccessArticle
Bridging the Gap between Project-Oriented and Exercise-Oriented Automatic Assessment Tools
by
Bruno Pereira Cipriano, Bernardo Baltazar, Nuno Fachada, Athanasios Vourvopoulos and Pedro Alves
Computers 2024, 13(7), 162; https://doi.org/10.3390/computers13070162 - 30 Jun 2024
Abstract
In this study, we present the DP Plugin for IntelliJ IDEA, designed to extend the Drop Project (DP) Automatic Assessment Tool (AAT) by making it more suitable for handling small exercises in exercise-based learning environments. Our aim was to address the limitations of
[...] Read more.
In this study, we present the DP Plugin for IntelliJ IDEA, designed to extend the Drop Project (DP) Automatic Assessment Tool (AAT) by making it more suitable for handling small exercises in exercise-based learning environments. Our aim was to address the limitations of DP in supporting small assignments while retaining its strengths in project-based learning. The plugin leverages DP’s REST API to streamline the submission process, integrating assignment instructions and feedback directly within the IDE. A student survey conducted during the 2022/23 academic year revealed a positive reception, highlighting benefits such as time efficiency and ease of use. Students also provided valuable feedback, leading to various improvements that have since been integrated into the plugin. Despite these promising results, the study is limited by the relatively small percentage of survey respondents. Our findings suggest that an IDE plugin can significantly improve the usability of project-oriented AATs for small exercises, informing the development of future educational tools suitable for mixed project-based and exercise-based learning environments.
Full article
(This article belongs to the Special Issue Future Trends in Computer Programming Education)
►▼
Show Figures
Figure 1
Open AccessArticle
Modeling Autonomous Vehicle Responses to Novel Observations Using Hierarchical Cognitive Representations Inspired Active Inference
by
Sheida Nozari, Ali Krayani, Pablo Marin, Lucio Marcenaro, David Martin Gomez and Carlo Regazzoni
Computers 2024, 13(7), 161; https://doi.org/10.3390/computers13070161 - 28 Jun 2024
Abstract
Equipping autonomous agents for dynamic interaction and navigation is a significant challenge in intelligent transportation systems. This study aims to address this by implementing a brain-inspired model for decision making in autonomous vehicles. We employ active inference, a Bayesian approach that models decision-making
[...] Read more.
Equipping autonomous agents for dynamic interaction and navigation is a significant challenge in intelligent transportation systems. This study aims to address this by implementing a brain-inspired model for decision making in autonomous vehicles. We employ active inference, a Bayesian approach that models decision-making processes similar to the human brain, focusing on the agent’s preferences and the principle of free energy. This approach is combined with imitation learning to enhance the vehicle’s ability to adapt to new observations and make human-like decisions. The research involved developing a multi-modal self-awareness architecture for autonomous driving systems and testing this model in driving scenarios, including abnormal observations. The results demonstrated the model’s effectiveness in enabling the vehicle to make safe decisions, particularly in unobserved or dynamic environments. The study concludes that the integration of active inference with imitation learning significantly improves the performance of autonomous vehicles, offering a promising direction for future developments in intelligent transportation systems.
Full article
(This article belongs to the Special Issue System-Integrated Intelligence and Intelligent Systems 2023)
Open AccessArticle
An NLP-Based Exploration of Variance in Student Writing and Syntax: Implications for Automated Writing Evaluation
by
Maria Goldshtein, Amin G. Alhashim and Rod D. Roscoe
Computers 2024, 13(7), 160; https://doi.org/10.3390/computers13070160 - 25 Jun 2024
Abstract
In writing assessment, expert human evaluators ideally judge individual essays with attention to variance among writers’ syntactic patterns. There are many ways to compose text successfully or less successfully. For automated writing evaluation (AWE) systems to provide accurate assessment and relevant feedback, they
[...] Read more.
In writing assessment, expert human evaluators ideally judge individual essays with attention to variance among writers’ syntactic patterns. There are many ways to compose text successfully or less successfully. For automated writing evaluation (AWE) systems to provide accurate assessment and relevant feedback, they must be able to consider similar kinds of variance. The current study employed natural language processing (NLP) to explore variance in syntactic complexity and sophistication across clusters characterized in a large corpus (n = 36,207) of middle school and high school argumentative essays. Using NLP tools, k-means clustering, and discriminant function analysis (DFA), we observed that student writers employed four distinct syntactic patterns: (1) familiar and descriptive language, (2) consistently simple noun phrases, (3) variably complex noun phrases, and (4) moderate complexity with less familiar language. Importantly, each pattern spanned the full range of writing quality; there were no syntactic patterns consistently evaluated as “good” or “bad”. These findings support the need for nuanced approaches in automated writing assessment while informing ways that AWE can participate in that process. Future AWE research can and should explore similar variability across other detectable elements of writing (e.g., vocabulary, cohesion, discursive cues, and sentiment) via diverse modeling methods.
Full article
(This article belongs to the Special Issue Natural Language Processing (NLP) and Large Language Modelling)
Open AccessArticle
Enhanced Security Access Control Using Statistical-Based Legitimate or Counterfeit Identification System
by
Aisha Edrah and Abdelkader Ouda
Computers 2024, 13(7), 159; https://doi.org/10.3390/computers13070159 - 22 Jun 2024
Abstract
With our increasing reliance on technology, there is a growing demand for efficient and seamless access control systems. Smartphone-centric biometric methods offer a diverse range of potential solutions capable of verifying users and providing an additional layer of security to prevent unauthorized access.
[...] Read more.
With our increasing reliance on technology, there is a growing demand for efficient and seamless access control systems. Smartphone-centric biometric methods offer a diverse range of potential solutions capable of verifying users and providing an additional layer of security to prevent unauthorized access. To ensure the security and accuracy of smartphone-centric biometric identification, it is crucial that the phone reliably identifies its legitimate owner. Once the legitimate holder has been successfully determined, the phone can effortlessly provide real-time identity verification for various applications. To achieve this, we introduce a novel smartphone-integrated detection and control system called Identification: Legitimate or Counterfeit (ILC), which utilizes gait cycle analysis. The ILC system employs the smartphone’s accelerometer sensor, along with advanced statistical methods, to detect the user’s gait pattern, enabling real-time identification of the smartphone owner. This approach relies on statistical analysis of measurements obtained from the accelerometer sensor, specifically, peaks extracted from the X-axis data. Subsequently, the derived feature’s probability distribution function (PDF) is computed and compared to the known user’s PDF. The calculated probability verifies the similarity between the distributions, and a decision is made with 92.18% accuracy based on a predetermined verification threshold.
Full article
(This article belongs to the Special Issue Wireless Sensor Network, IoT and Cloud Computing Technologies for Smart Cities)
Open AccessArticle
Personalized Classifier Selection for EEG-Based BCIs
by
Javad Rahimipour Anaraki, Antonina Kolokolova and Tom Chau
Computers 2024, 13(7), 158; https://doi.org/10.3390/computers13070158 - 21 Jun 2024
Abstract
The most important component of an Electroencephalogram (EEG) Brain–Computer Interface (BCI) is its classifier, which translates EEG signals in real time into meaningful commands. The accuracy and speed of the classifier determine the utility of the BCI. However, there is significant intra- and
[...] Read more.
The most important component of an Electroencephalogram (EEG) Brain–Computer Interface (BCI) is its classifier, which translates EEG signals in real time into meaningful commands. The accuracy and speed of the classifier determine the utility of the BCI. However, there is significant intra- and inter-subject variability in EEG data, complicating the choice of the best classifier for different individuals over time. There is a keen need for an automatic approach to selecting a personalized classifier suited to an individual’s current needs. To this end, we have developed a systematic methodology for individual classifier selection, wherein the structural characteristics of an EEG dataset are used to predict a classifier that will perform with high accuracy. The method was evaluated using motor imagery EEG data from Physionet. We confirmed that our approach could consistently predict a classifier whose performance was no worse than the single-best-performing classifier across the participants. Furthermore, Kullback–Leibler divergences between reference distributions and signal amplitude and class label distributions emerged as the most important characteristics for classifier prediction, suggesting that classifier choice depends heavily on the morphology of signal amplitude densities and the degree of class imbalance in an EEG dataset.
Full article
(This article belongs to the Special Issue Machine and Deep Learning in the Health Domain 2024)
Open AccessArticle
Advancing Skin Cancer Prediction Using Ensemble Models
by
Priya Natha and Pothuraju RajaRajeswari
Computers 2024, 13(7), 157; https://doi.org/10.3390/computers13070157 - 21 Jun 2024
Abstract
There are many different kinds of skin cancer, and an early and precise diagnosis is crucial because skin cancer is both frequent and deadly. The key to effective treatment is accurately classifying the various skin cancers, which have unique traits. Dermoscopy and other
[...] Read more.
There are many different kinds of skin cancer, and an early and precise diagnosis is crucial because skin cancer is both frequent and deadly. The key to effective treatment is accurately classifying the various skin cancers, which have unique traits. Dermoscopy and other advanced imaging techniques have enhanced early detection by providing detailed images of lesions. However, accurately interpreting these images to distinguish between benign and malignant tumors remains a difficult task. Improved predictive modeling techniques are necessary due to the frequent occurrence of erroneous and inconsistent outcomes in the present diagnostic processes. Machine learning (ML) models have become essential in the field of dermatology for the automated identification and categorization of skin cancer lesions using image data. The aim of this work is to develop improved skin cancer predictions by using ensemble models, which combine numerous machine learning approaches to maximize their combined strengths and reduce their individual shortcomings. This paper proposes a fresh and special approach for ensemble model optimization for skin cancer classification: the Max Voting method. We trained and assessed five different ensemble models using the ISIC 2018 and HAM10000 datasets: AdaBoost, CatBoost, Random Forest, Gradient Boosting, and Extra Trees. Their combined predictions enhance the overall performance with the Max Voting method. Moreover, the ensemble models were fed with feature vectors that were optimally generated from the image data by a genetic algorithm (GA). We show that, with an accuracy of 95.80%, the Max Voting approach significantly improves the predictive performance when compared to the five ensemble models individually. Obtaining the best results for F1-measure, recall, and precision, the Max Voting method turned out to be the most dependable and robust. The novel aspect of this work is that skin cancer lesions are more robustly and reliably classified using the Max Voting technique. Several pre-trained machine learning models’ benefits are combined in this approach.
Full article
(This article belongs to the Special Issue Machine and Deep Learning in the Health Domain 2024)
Open AccessArticle
Chef Dalle: Transforming Cooking with Multi-Model Multimodal AI
by
Brendan Hannon, Yulia Kumar, J. Jenny Li and Patricia Morreale
Computers 2024, 13(7), 156; https://doi.org/10.3390/computers13070156 - 21 Jun 2024
Abstract
In an era where dietary habits significantly impact health, technological interventions can offer personalized and accessible food choices. This paper introduces Chef Dalle, a recipe recommendation system that leverages multi-model and multimodal human-computer interaction (HCI) techniques to provide personalized cooking guidance. The application
[...] Read more.
In an era where dietary habits significantly impact health, technological interventions can offer personalized and accessible food choices. This paper introduces Chef Dalle, a recipe recommendation system that leverages multi-model and multimodal human-computer interaction (HCI) techniques to provide personalized cooking guidance. The application integrates voice-to-text conversion via Whisper and ingredient image recognition through GPT-Vision. It employs an advanced recipe filtering system that utilizes user-provided ingredients to fetch recipes, which are then evaluated through multi-model AI through integrations of OpenAI, Google Gemini, Claude, and/or Anthropic APIs to deliver highly personalized recommendations. These methods enable users to interact with the system using voice, text, or images, accommodating various dietary restrictions and preferences. Furthermore, the utilization of DALL-E 3 for generating recipe images enhances user engagement. User feedback mechanisms allow for the refinement of future recommendations, demonstrating the system’s adaptability. Chef Dalle showcases potential applications ranging from home kitchens to grocery stores and restaurant menu customization, addressing accessibility and promoting healthier eating habits. This paper underscores the significance of multimodal HCI in enhancing culinary experiences, setting a precedent for future developments in the field.
Full article
(This article belongs to the Special Issue Harnessing Artificial Intelligence for Social and Semantic Understanding)
►▼
Show Figures
Figure 1
Open AccessArticle
A Contextual Model for Visual Information Processing
by
Illia Khurtin and Mukesh Prasad
Computers 2024, 13(6), 155; https://doi.org/10.3390/computers13060155 - 20 Jun 2024
Abstract
Despite significant achievements in the artificial narrow intelligence sphere, the mechanisms of human-like (general) intelligence are still undeveloped. There is a theory stating that the human brain extracts the meaning of information rather than recognizes the features of a phenomenon. Extracting the meaning
[...] Read more.
Despite significant achievements in the artificial narrow intelligence sphere, the mechanisms of human-like (general) intelligence are still undeveloped. There is a theory stating that the human brain extracts the meaning of information rather than recognizes the features of a phenomenon. Extracting the meaning is finding a set of transformation rules (context) and applying them to the incoming information, producing an interpretation. Then, the interpretation is compared to something already seen and is stored in memory. Information can have different meanings in different contexts. A mathematical model of a context processor and a differential contextual space which can perform the interpretation is discussed and developed in this paper. This study examines whether the basic principles of differential contextual spaces work in practice. The model is developed with Rust programming language and trained on black and white images which are rotated and shifted both horizontally and vertically according to the saccades and torsion movements of a human eye. Then, a picture that has never been seen in the particular transformation, but has been seen in another one, is exposed to the model. The model considers the image in all known contexts and extracts the meaning. The results show that the program can successfully process black and white images which are transformed by shifts and rotations. This research prepares the grounding for further investigations of the contextual model principles with which general intelligence might operate.
Full article
(This article belongs to the Special Issue Feature Papers in Computers 2024)
►▼
Show Figures
Figure 1
Open AccessArticle
Deep Convolutional Generative Adversarial Networks in Image-Based Android Malware Detection
by
Francesco Mercaldo, Fabio Martinelli and Antonella Santone
Computers 2024, 13(6), 154; https://doi.org/10.3390/computers13060154 - 19 Jun 2024
Cited by 1
Abstract
The recent advancements in generative adversarial networks have showcased their remarkable ability to create images that are indistinguishable from real ones. This has prompted both the academic and industrial communities to tackle the challenge of distinguishing fake images from genuine ones. We introduce
[...] Read more.
The recent advancements in generative adversarial networks have showcased their remarkable ability to create images that are indistinguishable from real ones. This has prompted both the academic and industrial communities to tackle the challenge of distinguishing fake images from genuine ones. We introduce a method to assess whether images generated by generative adversarial networks, using a dataset of real-world Android malware applications, can be distinguished from actual images. Our experiments involved two types of deep convolutional generative adversarial networks, and utilize images derived from both static analysis (which does not require running the application) and dynamic analysis (which does require running the application). After generating the images, we trained several supervised machine learning models to determine if these classifiers can differentiate between real and generated malicious applications. Our results indicate that, despite being visually indistinguishable to the human eye, the generated images were correctly identified by a classifier with an F-measure of approximately 0.8. While most generated images were accurately recognized as fake, some were not, leading them to be considered as images produced by real applications.
Full article
(This article belongs to the Special Issue Current Issue and Future Directions in Multimedia Hiding and Signal Processing)
►▼
Show Figures
Figure 1
Open AccessArticle
Empowering Communication: A Deep Learning Framework for Arabic Sign Language Recognition with an Attention Mechanism
by
R. S. Abdul Ameer, M. A. Ahmed, Z. T. Al-Qaysi, M. M. Salih and Moceheb Lazam Shuwandy
Computers 2024, 13(6), 153; https://doi.org/10.3390/computers13060153 - 19 Jun 2024
Abstract
This article emphasises the urgent need for appropriate communication tools for communities of people who are deaf or hard-of-hearing, with a specific emphasis on Arabic Sign Language (ArSL). In this study, we use long short-term memory (LSTM) models in conjunction with MediaPipe to
[...] Read more.
This article emphasises the urgent need for appropriate communication tools for communities of people who are deaf or hard-of-hearing, with a specific emphasis on Arabic Sign Language (ArSL). In this study, we use long short-term memory (LSTM) models in conjunction with MediaPipe to reduce the barriers to effective communication and social integration for deaf communities. The model design incorporates LSTM units and an attention mechanism to handle the input sequences of extracted keypoints from recorded gestures. The attention layer selectively directs its focus toward relevant segments of the input sequence, whereas the LSTM layer handles temporal relationships and encodes the sequential data. A comprehensive dataset comprised of fifty frequently used words and numbers in ArSL was collected for developing the recognition model. This dataset comprises many instances of gestures recorded by five volunteers. The results of the experiment support the effectiveness of the proposed approach, as the model achieved accuracies of more than 85% (individual volunteers) and 83% (combined data). The high level of precision emphasises the potential of artificial intelligence-powered translation software to improve effective communication for people with hearing impairments and to enable them to interact with the larger community more easily.
Full article
(This article belongs to the Special Issue Advanced Image Processing and Computer Vision)
►▼
Show Figures
Figure 1
Open AccessReview
Hybrid Architectures Used in the Protection of Large Healthcare Records Based on Cloud and Blockchain Integration: A Review
by
Leonardo Juan Ramirez Lopez, David Millan Mayorga, Luis Hernando Martinez Poveda, Andres Felipe Carbonell Amaya and Wilson Rojas Reales
Computers 2024, 13(6), 152; https://doi.org/10.3390/computers13060152 - 12 Jun 2024
Abstract
The management of large medical files poses a critical challenge in the health sector, with conventional systems facing deficiencies in security, scalability, and efficiency. Blockchain ensures the immutability and traceability of medical records, while the cloud allows scalable and efficient storage. Together, they
[...] Read more.
The management of large medical files poses a critical challenge in the health sector, with conventional systems facing deficiencies in security, scalability, and efficiency. Blockchain ensures the immutability and traceability of medical records, while the cloud allows scalable and efficient storage. Together, they can transform the data management of electronic health record applications. The method used was the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) methodology to choose and select the relevant studies that contribute to this research, with special emphasis set on maintaining the integrity and security of the blockchain while tackling the potential and efficiency of cloud infrastructures. The study’s focus is to provide a comprehensive and insightful examination of the modern landscape concerning the integration of blockchain and cloud advances, highlighting the current challenges and building a solid foundation for future development. Furthermore, it is very important to increase the integration of blockchain security with the dynamic potential of cloud computing while guaranteeing information integrity and security remain uncompromised. In conclusion, this paper serves as an important resource for analysts, specialists, and partners looking to delve into and develop the integration of blockchain and cloud innovations.
Full article
(This article belongs to the Section Cloud Continuum and Enabled Applications)
►▼
Show Figures
Figure 1
Open AccessArticle
InfoSTGCAN: An Information-Maximizing Spatial-Temporal Graph Convolutional Attention Network for Heterogeneous Human Trajectory Prediction
by
Kangrui Ruan and Xuan Di
Computers 2024, 13(6), 151; https://doi.org/10.3390/computers13060151 - 11 Jun 2024
Abstract
►▼
Show Figures
Predicting the future trajectories of multiple interacting pedestrians within a scene has increasingly gained importance in various fields, e.g., autonomous driving, human–robot interaction, and so on. The complexity of this problem is heightened due to the social dynamics among different pedestrians and their
[...] Read more.
Predicting the future trajectories of multiple interacting pedestrians within a scene has increasingly gained importance in various fields, e.g., autonomous driving, human–robot interaction, and so on. The complexity of this problem is heightened due to the social dynamics among different pedestrians and their heterogeneous implicit preferences. In this paper, we present Information Maximizing Spatial-Temporal Graph Convolutional Attention Network (InfoSTGCAN), which takes into account both pedestrian interactions and heterogeneous behavior choice modeling. To effectively capture the complex interactions among pedestrians, we integrate spatial-temporal graph convolution and spatial-temporal graph attention. For grasping the heterogeneity in pedestrians’ behavior choices, our model goes a step further by learning to predict an individual-level latent code for each pedestrian. Each latent code represents a distinct pattern of movement choice. Finally, based on the observed historical trajectory and the learned latent code, the proposed method is trained to cover the ground-truth future trajectory of this pedestrian with a bi-variate Gaussian distribution. We evaluate the proposed method through a comprehensive list of experiments and demonstrate that our method outperforms all baseline methods on the commonly used metrics, Average Displacement Error and Final Displacement Error. Notably, visualizations of the generated trajectories reveal our method’s capacity to handle different scenarios.
Full article
Figure 1
Highly Accessed Articles
Latest Books
E-Mail Alert
News
Topics
Topic in
Applied Sciences, Computers, Digital, Electronics, Smart Cities
Artificial Intelligence Models, Tools and Applications
Topic Editors: Phivos Mylonas, Katia Lida Kermanidis, Manolis MaragoudakisDeadline: 31 August 2024
Topic in
Biomedicines, Computers, Information, IJERPH, JPM
eHealth and mHealth: Challenges and Prospects, 2nd Volume
Topic Editors: Antonis Billis, Manuel Dominguez-Morales, Anton CivitDeadline: 30 September 2024
Topic in
Applied Sciences, Computers, Electronics, JSAN, Technologies
Emerging AI+X Technologies including Selected Papers from ICGHIT 2024
Topic Editors: Byung-Seo Kim, Hyunsik Ahn, Kyu-Tae LeeDeadline: 31 October 2024
Topic in
Computers, Informatics, Information, Logistics, Mathematics, Algorithms
Decision Science Applications and Models (DSAM)
Topic Editors: Daniel Riera Terrén, Angel A. Juan, Majsa Ammuriova, Laura CalvetDeadline: 31 December 2024
Conferences
Special Issues
Special Issue in
Computers
Xtended or Mixed Reality (AR+VR) for Education 2024
Guest Editors: Veronica Rossano, Michele FiorentinoDeadline: 1 August 2024
Special Issue in
Computers
Best Practices, Challenges and Opportunities in Software Engineering
Guest Editor: Yan LiuDeadline: 31 August 2024
Special Issue in
Computers
Uncertainty-Aware Artificial Intelligence
Guest Editors: Hussain Mohammed Dipu Kabir, Syed Bahauddin Alam, Subrota Kumar Mondal, Jeremy StraubDeadline: 30 September 2024
Special Issue in
Computers
Smart Learning Environments
Guest Editor: Ananda MaitiDeadline: 31 October 2024