Now showing 1 - 10 of 13
  • Publication
    Imporved MPR selection algorithm-based WS-OLSR routing protocol
    Vehicle Ad Hoc Networks (VANETs) have become a viable technology to improve traffic flow and safety on the roads. Due to its effectiveness and scalability, the Wingsuit Search-based Optimised Link State Routing Protocol (WS-OLSR) is frequently used for data distribution in VANETs. However, the selection of MultiPoint Relays (MPRs) plays a pivotal role in WS-OLSR’s performance. This paper presents an improved MPR selection algorithm tailored to WS-OLSR, designed to enhance the overall routing efficiency and reduce overhead. The analysis found that the current OLSR protocol has problems such as redundancy of HELLO and TC message packets or failure to update routing information in time, so a WS-OLSR routing protocol based on improved-MPR selection algorithm was proposed. Firstly, factors such as node mobility and link changes are comprehensively considered to reflect network topology changes, and the broadcast cycle of node HELLO messages is controlled through topology changes. Secondly, a new MPR selection algorithm is proposed, considering link stability issues and nodes. Finally, evaluate its effectiveness in terms of packet delivery ratio, end-to-end delay, and control message overhead. Simulation results demonstrate the superior performance of our improved MR selection algorithm when compared to traditional approaches.
      1
  • Publication
    Deep-Learning Assisting Cerebral Palsy Patient Handgrip Task Translation
    ( 2021-07-26) ; ;
    Phaklen Ehkan
    ;
    Muslim Mustapa
    ;
    An electro-encephalography (EEG) brain-computer interface (BCI) can provide the brain and external environment with separate information sharing and control networks. EEG impulses, though, come from many electrodes, which produce different characteristics, and how the electrodes and features to enhance classification efficiency have been chosen has become an urgent concern. This paper explores the deep convolutional neural network architecture (CNN) hyper-parameters with separating temporal and spatial filters without any pre-processing or artificial extraction processes. It selects the raw EEG signal of electrode pairs over the cortical area as hybrid samples. Our proposed deep-learning model outperforms other neural network models previously applied to this dataset in training time (∼40%) and accuracy (∼6%). Besides, considerations such as optimum order for EEG channels do not limit our model, and it is patient-invariant. The impact of network architecture on decoder output and training time is further discussed.
      1
  • Publication
    Toward Adaptive and Scalable Topology in Distributed SDN Controller
    The increasing need for automated networking platforms like the Internet of Things, as well as network services like cloud computing, big data applications, wireless networks, mobile Internet, and virtualization, has driven existing networks to their limitations. Software-defined network (SDN) is a new modern programmable network architectural technology that allows network administrators to control the entire network consistently and logically centralized in software-based controllers and network devices become just simple packet forwarding devices. The controller that is the network's brain, is mostly based on the OpenFlow protocol and has distinct characteristics that vary depending on the programming language. Its function is to control network traffic and increase network resource efficiency. Therefore, selecting the right controllers and monitoring their performance to increase resource usage and enhance network performance metrics is required. For network performance metrics analysis, the study proposes an implementation of SDN architecture utilizing an open-source OpenDaylight (ODL) distributed SDN controller. The proposed work evaluates the deployment of distributed SDN controller performance on three distinct customized network topologies based on SDN architecture for node-to-node performance metrics such as delay, throughput, packet loss, and bandwidth use. The experiments are conducted using the Mininet emulation tool. Wireshark is used to collect and analyse packets in real-time. The results obtained from the comparison of networks are presented to provide useful guidelines for SDN research and deployment initiatives.
      1
  • Publication
    Breast cancer classification using deep learning and FPGA inferencing
    Implementing deep learning technology with FPGA as an accelerator has become a popular application due to its efficiency and performance. However, given the tremendous data generated on medical diagnosis, normal inference speed is not sufficient. Hence, the FPGA technology is implemented for fast inference. In this context, the FPGA accelerates the deep learning inference process for fast breast cancer classification with minimal latency on real-time deployment. This paper summarizes the findings of model deployment across various computing devices in deep learning technology with FPGA. The study includes model performance evaluation, throughput, and latency comparison with different batch sizes to the extent of expected delay for real-world deployment. The result concludes that FPGA is the most suitable to act as a deep learning inference accelerator with a high throughput-to-latency ratio and fast parallel inference.
      1
  • Publication
    Tuberculosis Classification Using Deep Learning and FPGA Inferencing
    Among the top 10 leading causes of mortality, tuberculosis (TB) is a chronic lung illness caused by a bacterial infection. Due to its efficiency and performance, using deep learning technology with FPGA as an accelerator has become a standard application in this work. However, considering the vast amount of data collected for medical diagnosis, the average inference speed is inadequate. In this scenario, the FPGA speeds the deep learning inference process enabling the real-time deployment of TB classification with low latency. This paper summarizes the findings of model deployment across various computing devices in inferencing deep learning technology with FPGA. The study includes model performance evaluation, throughput, and latency comparison with different batch sizes to the extent of expected delay for real-world deployment. The result concludes that FPGA is the most suitable to act as a deep learning inference accelerator with a high throughput-to-latency ratio and fast parallel inference. The FPGA inferencing demonstrated an increment of 21.8% in throughput while maintaining a 31% lower latency than GPU inferencing and 6x more energy efficiency. The proposed inferencing also delivered over 90% accuracy and selectivity to detect and localize the TB.
      2
  • Publication
    Deep Learning with FPGA: Age and Gender Recognition for Smart Advertisement Board
    ( 2023-10-06)
    Yeoh W.S.
    ;
    ;
    Mustapa M.
    ;
    ; ;
    Mozi A.M.
    Age and gender recognition are helpful in various applications, especially in the field of advertising. To replace the traditional advertising method that can only display the same contents to all audiences, a smart advertisement board capable of detecting age and gender of audiences to display relevant contents is required to increase the effectiveness of advertising. This paper will use two image datasets to train and test the Convolutional Neural Network (CNN) based architecture models for age and gender recognition using deep learning. The dataset that produced the best performing model will be implemented on three different devices to observe the performance of the models on each device. A gender recognition model with accuracy of 91.53% and age recognition model with accuracy of 59.62% is produced. The results have also shown the use of Field Programmable Gated Array (FPGA) has greatly boosted the performance of the models in terms of throughput and latency.
      1
  • Publication
    The impact of inner-parameters B-MAC protocol by Taguchi method for WSN
    The MAC protocols play an important role in the performance of wireless sensor network (WSN). MAC protocols are controlled with set of parameters from being dragged to undesired situation such as reduce the power consumption, listening idle, and overhead. This inner- parameters have direct impact on the efficiency of a MAC protocols and overall network performances. The impacts of theses parameters on reduce the power consumption are less considered. In the literature, a lot of studies concentrates on introducing a new protocols to reduce the power consumption for WSN. This paper aims to analysis the inner- parameters of MAC protocols for WSN power consumption by using Taguchi Delta Analysis (TDA). Moreover, the measure of inner - parameters is very important to find the optimal values to reduce the power consumption. This paper utilized Taguchi method to analysis the impact of B-MAC protocol parameters in WSN scenarios by exploits Taguchi delta analysis. Further, four inner - parameters are investigated in a simulation platform. Moreover, simulation experiments are carried out by OMNET++5 to prove the work in this paper. The obtained results show that inner- parameters B-MAC inner- protocol reduce the power consumption of WSN for two different scenarios.
  • Publication
    Comparing and Assessing the Enhancements of DYMO and OLSR in VANETs
    ( 2020-09-21)
    Khalid Ahmed W.
    ;
    ;
    Khalid Abduljabbar W.
    ;
    The main aspect of this work is to study the differences and define the behaviour of two different routing protocols. The first side is Dynamic MANETs On- Demand (DYMO) while the other side is proactive, optimal link-state routing (OLSR) and both the first and second are interactive routing protocols in the Ad-hoc network (VANET). The efficiency of these protocols was analysed and studied based on the use of three performance indicators: PDR, normal load (NRO) and end-to-end de1ay (E2ED) on the ability to change the size of different nodes. Omnet ++ was used by the INET Framework. We also used the SUMO simulation tool to build random movement patterns for VANET. From full simulation, we noticed that OLSR is doing better than DYMO for VANET at a price. Late and, as a result, the development of OLSR work in VANETs compared to DYMO, packet receipt ratios (PDR), side-to-side delay, normal path load, and VANETs.
  • Publication
    Multipoint Relay Path for Efficient Topology Maintenance Algorithm in Optimized Link State Routing-Based for VANET
    The Optimal Link State Routing (OLSR) protocol employs multipoint relay (MPR) nodes to disseminate topology control (TC) messages, enabling network topology discovery and maintenance. However, this approach increases control overhead and leads to wasted network bandwidth in stable topology scenarios due to fixed flooding periods. To address these challenges, this paper presents an Efficient Topology Maintenance Algorithm (ETM-OLSR) for Enhanced Link-State Routing Protocols. By reducing the number of MPR nodes, TC message generation and forwarding frequency are minimized. Furthermore, the algorithm selects a smaller subset of TC messages based on the changes in the MPR selection set from the previous cycle, adapting to stable and fluctuating network conditions. Additionally, the sending cycle of TC messages is dynamically adjusted in response to network topology changes. Simulation results demonstrate that the ETM-OLSR algorithm effectively reduces network control overhead, minimizes end-to-end delay, and improves network throughput compared to traditional OLSR and HTR-OLSR algorithms.
  • Publication
    Hyper-threading technology: Not a good choice for speeding up CPU-bound code
    Hyper-threading (HT) technology allows one thread to execute its task while another thread is stalled waiting for shared resource or other operations to complete. Thus, this reduces the idle time of a processor. If HT is enabled, an operating system would see two logical cores per each physical core. This gives one physical core the ability to run two threads simultaneously. However, it does not necessarily speed up the performance of a parallel code twice the number of physical cores. This happens when two threads are trying to access the shared CPU resource. The instructions could only be executed one after another at any given time. In this case, parallel CPU-bound code could attain a little improvement in terms of speedup from HT on a quad-core platform, which is Intel i5-2410M@2.30GHz.