1. Introduction
Machine learning has become a ubiquitous feature in everyday life. From self-driving vehicles, facial recognition systems, and real-time interpretation of different languages, to security surveillance, smart home applications, and health monitoring, artificial intelligence has changed almost every society on earth [1,2,3,4]. Due to the extremely high computational requirements of machine learning models, until recently, the majority of these breakthroughs were implemented on high-power stationary computing systems. However, continuous advancements in embedded system design have made the implementation of machine learning models on embedded computing systems for a wide variety of mobile and low-power applications viable. One example of such an application would be [5], a 2020 paper by Ouyang et al., titled “Deep CNN-Based Real-Time Traffic Light Detector for Self-Driving Vehicles”, which proposes a method for recognizing traffic lights for autonomous vehicles. This ever-expanding research field of machine learning implementation in limited environments of embedded systems has been titled “Embedded Machine Learning” [6]. There are many considerations when choosing an embedded system for a specific machine learning application, such as power limitations, specific sensor outputs, model architecture, and monetary cost. In this review paper, we focus on the system models and assess which systems are better suited for which specific applications and sensing schemes.
As stated, machine learning algorithms are trained and used for many different applications, such as hand gesture recognition [7] and speech source identification [8]. They usually have a very high performance and memory requirement for both training and inference. Effective implementation would require the tuning and modification of the machine learning model architecture as well as the selection of the appropriate system depending on the priorities of the application. All machine learning applications aim to consume as little power and computation and be as fast and accurate as possible, however, improvement in one of these areas almost always comes at a relative cost to the other ones. Since embedded systems can vary drastically in power consumption, processing power, memory, storage, and pricing, it is prudent to select the appropriate system for each specific application. As an example, a system for pedestrian detection for autonomous vehicles [9] would prioritize performance speed and accuracy much more so than a system designed for recognizing marine life [10], even if it comes at a much higher monetary cost.
Training a machine learning model for any task requires a dataset, which can consist of megabytes to terabytes of images, video files, audio files, graphs, etc., and their corresponding annotation files. The specific files of a dataset used for training depend on the intended application of the machine learning model, an image classification model, for example, would use a dataset of image files and label annotations associated with each image. The sensing schemes used for collecting these files, both for the initial training and testing datasets, as well as for the inference of the trained machine learning algorithm on an embedded system, are varied. Another subject of analysis in this research was the correlation between the type of sensor scheme used in each system to the overall implementation of the system.
Most of the papers reviewed in this work utilized some form of computer vision, mainly in areas such as obstacle detection for autonomous vehicles (such as speed bumps) [11] or safety and security measures (such as violent assault identification) [12]. However, several also presented embedded machine learning methods for medical applications (such as patient heart monitoring) [13] or automating more aspects of city management (such as managing the direction and flow of vehicular traffic) [14].
Essentially, in this review, we emphasized specific applications, embedded hardware platforms, and sensors, then compared them based on the nature of those networks and applications, while any other embedded machine learning review papers have a greater focus on the performance of specific lines of hardware [15], or the network architecture implemented on the hardware [16]. The paper is structured in the following format: 1. Abstract; 2. Introduction; 3. Hardware System Considerations; 4. Specific Hardware Systems Covered In The Review; 5. Sensing Systems; 6. Network Applications; 7. Comprehensive Comparisons; 8. Conclusions. This layout is also displayed in Figure 1. If the readers are interested in machine learning algorithms, models, and databases, please refer to other review and benchmark papers such as the ones used as sources in this work [15,16,17]. Works such as [18,19,20,21] and [15,17] provide a comprehensive performance analysis and benchmark of the embedded systems used in their specified applications, while works such as [22,23] conduct a more in-depth study on improvement methods for both system hardware and model architecture for their specific applications.
2. Objective and Method
To reiterate, the goal of this study is to summarize the current state-of-the-art research in the embedded machine learning area for different applications, so that the researchers could have an overview of the cutting-edge methods and results, as well predict the general trajectory of embedded machine learning advances. The method of research for this study was the compilation of the results gathered by the research papers referenced for this work. Excluding the related works in the Benchmark and Review section of the references, all of the papers presented in this review included a proposal or implementation of embedded machine learning for a specified application with the results of each study including one or all of the following findings: accuracy, inference speed, and power consumption.
3. Hardware
Embedded systems are computer hardware systems designed for performing dedicated functions in a combination with a larger system. They include and are used in many everyday items from mobile phones and household appliances. Embedded computer devices are a subset of embedded systems used for computational tasks for more dedicated or remote operations, such as running machine learning algorithms in real time on small unmanned aerial vehicles, connecting systems connected to the internet of things, and even security monitoring. While the variety of the embedded computer devices produced and used is quite wide, most academic research conducted on embedded machine learning is focused on using Raspberry Pi and NVIDIA Jetson devices. Some other devices used include the ASUS Tinker board series, Google’s Coral TPU dev series, ODROID-XU4 Boards, and the Banana Pi board series.
3.1. General Considerations
When choosing an embedded computing device for specific applications, many different parameters need to be kept in mind. The parameters include, but are not limited to, system processing speed, affected by the integrated CPU and GPU of a system, system memory affected by the RAM, system storage space, system bus and drivers, the overall power consumption of a system, and its cost of purchase. Generally, systems with higher performance and memory are capable of performing more complex machine learning tasks at a greater speed but have high power consumption rates and monetary prices. On the other hand, cheaper and less power-intensive systems have lower performances and memory, making them perform their dedicated task far slower.
3.2. Processor Units
Processing units are the integrated electrical circuits responsible for performing the fundamental algorithmic and arithmetic logic processes for running a computer device. There are different categories of processors, with the most common ones in embedded computer systems being CPUs and GPUs. Central Processing Units, or CPUs, are the processors present in most electrical devices and are responsible for the execution of programs and applications, they are usually composed of multiple cores and have their performance measured in gigahertz. Graphical Processing Units, or GPUs, are dedicated processors used for graphical rendering, allowing devices to allocate graphically intensive tasks, such as real-time object recognition, to them. All of the embedded computer devices presented in this review contain both a CPU and GPU unit, with the CPUs being various ARM Cortex multicore processors [24,25,26,27,28,29,30,31,32,33,34]. The GPUs for each system were more varied in both clock speed and power consumption. More detailed descriptions are given within each systems subsection.
3.3. Memory Units
System memory generally refers to a computing system’s Random Access Memory or RAM, which is responsible for storing application data for quick access. The larger a system’s RAM, the quicker the system can run simultaneous applications, making RAM proportional to the overall performance of a system. Embedded computing devices are packaged with their own memory component, with most embedded systems in this review having between 1 GB, 2 GB, and 4 GB of RAM [30,31], while the most recent NVIDIA kits have between 8 GB and 16 GB [24,28]. Memory Bandwidth is another important parameter of system memory, indicating the rate at which data can be accessed and edited, with the bandwidth of the system included in this review ranging from 128-bit to 256-bit.
3.4. Storage Units
Computer storage refers to the component of a computing device responsible for retaining longtime application and computation data. While access and alteration to storage data by the CPU are much slower than its access to RAM data, it consumes far less power and processing capability. Storage systems come in many varieties such as flash drives, hard drives, solid state drives, SD cards, and embedded MultiMediaCard memory or eMMC. Hard drives have been the most common form of storage up until recently, with their advantage over other alternatives being their overall size and their downside being their relatively slow data access speed. Solid state drives or SSDs have provided far faster data access at the cost of storage size, however, in recent years, SSDs have made leaps in storage capacity and are now comparable in overall storage size to hard drives. Flash drives are quick and easy to connect or disconnect from different computing devices while having very small storage space, they are very similar to SSDs in terms of performance. Secure digital cards or SD cards are also similar to flash storage but have both much smaller sizes and storage capacities. eMMCs are architecturally similar to flash storage and are generally used in small laptops and embedded computing systems. Most development kit embedded computing systems contain eMMCs, this being very much the case in NVIDIA Jetson, Coral Edge, and ASUS Tinker board devices, and others, such as ODROID-XU4 boards, do not have their own integrated storage devices but instead have flash storage interface. Raspberry Pi boards have interfaces for both SD cards and Flash drives.
3.5. Operating Systems
Operating systems are responsible for managing and running all of the applications on a computing device, allowing applications to make requests for services through a defined application program interface (API). This makes the creation and usage of various applications much simpler, as all low-level functions, such as allocating disk space for an app, can be delegated to the OS. Operating systems rely on a library of device drivers to their services to specific hardware environments, so while every application makes a common call to a storage device, it is the OS that receives that call and uses the corresponding driver to translate the call into commands needed for the underlying hardware. Hardware capabilities are divided into three sections: providing UI through a CLI or GUI, launching and managing application execution, and identifying and exposing system hardware resources to the applications. Most personal computing devices utilize general-purpose operating systems, such as Windows, Mac OS, and Linux, and while there are specific embedded operating systems, mainly used in ATMs, Airplanes, and ioT devices, most embedded computing systems either utilize operating systems based on or very similar to general-purpose computer operating systems. For example, Nvidia Jetson boards have Linux for Tegra included in their development software kits [35].
3.6. Bus and Drivers
Computer buses are communication systems responsible for transferring data between the various components of a computing system. While most home computer systems have 32-bit to 64-bit buses, embedded devices have far smaller bit rates between 4-bit and 8-bit. Drivers refer to the systems responsible for communicating the software of a computer device to its hardware component. They generally run at a high privilege level in the OS run time environment, and in many cases are directly linked to the OS kernel, which is a portion of an OS such as Windows, Linux, or Mac OS, which remains memory-resident and handles execution for all other code. Drivers are what defines the messages from the OS to a specific device that facilitate the devices’ fulfillment of the OS’s request. The device drivers used in each embedded computing system are related to the operating systems of each device. For example, Raspberry Pi devices mainly use Raspberry Pi’s own operating system which is based on Debian, while Nvidia Jetson boards mainly rely on JetPack, Nvidia’s proprietary Software Development Kit (SDK) for their Jetson board series, which includes the Linux for Tegra (L4T) operating system. This means the driver kernels for both of these embedded system product lines are similar to that of a Linux computer [36].
Firmware refers to software formats that are directly embedded in specific devices, giving users low-level control over them. Essentially, firmware is responsible for giving simple devices their operation and system communication instructions. They are unique to other software in that they do not rely on APIs, OSs, or device drivers for operation. They are the first part of device programming to start sending instructions when the device is powered on, and in some more simple devices such as keyboards, they never pause their operations. They are mostly installed on a ROM for software protection and proximity to the physical component of their specific device. They can only work with a basic or low-level binary language known as machine language [37]. All of this applies to the components within an embedded system, meaning each device within the system has its own unique firmware with varying levels of complexity based on the function of the device.
4. Specific Systems
4.1. Nividia Jetson
Jetson is the name of a series of machine learning embedded systems by NVIDIA used for autonomous devices and various embedded applications. While Jetson Developer kits vary in capability and performance, they are generally very reliable for implementing machine learning tasks—this is especially true for more graphically intensive applications. The downside to this is that NVIDIA Jetson boards also tend to be more costly than market alternatives. Most of the sources shown in this review either only made use of Jetson boards or used their combination with other devices. These specific developer kits were the NVIDIA Jetson Nano, NVIDIA Jetson TX1, NVIDIA Jetson TX2, NVIDIA Jetson AGX Xavier, and NVIDIA Jetson Xavier NX.
NVIDIA Jetson Nano is one of the smaller Jetson kits specialized for machine learning tasks like image classification, object detection, segmentation, and speech processing. It has a 128-core Maxwell GPU, a Quad-core ARM Cortex A57 1.4Remote Sensing of EnvironmentHz CPU, 4 GB 64-bit LPDDR4 25.6 GB/s Memory, 2x MIPI CSI-2 DPHY lanes camera, Ethernet, HDMI, and USB connection ports. Unlike most other NVIDIA kits, Nano does not have an integrated storage unit and has to rely on SD cards for that purpose. It has a power consumption of 5–10 Watts and with a price range of USD 300–USD 500, it is the more affordable option out of all of the NVIDIA development kits [24].
The Jetson TX1 and TX2 series are a discontinued line of embedded system development kits with flexible capabilities that include great performance for machine learning tasks. As the discontinuation of this line of kits is especially recent for the TX2 series, research publications that utilize the TX2 board are not uncommon, with the TX1 being much rarer. The TX1 has a 256-core Maxwell GPU, a Quad-core ARM® Cortex®-A57 CPU, a 4 GB LPDDR4 memory, a 16 GB eMMC 5.1 Flash Storage, a 5 MP Fixed Focus MIPI CSI Camera, Ethernet, HDMI, and USB type A and Micro AB connection ports. The TX2 has NVIDIA Pascal™ Architecture GPU, 2 64-bit CPUs, Quad-Core Cortex®-A57 Complexes, an 8 GB L128 bit DDR4 memory, a 32 GB eMMC 5.1 Flash Storage, a 16 GB eMMC 5.1 Flash Storage, a 5 MP Fixed Focus MIPI CSI Camera, and Ethernet, HDMI, and USB type A and Micro AB connection ports. The power consumption of the TX1 is around 15 Watts and that of the TX2 is about 25 Watts [25,26].
The Jetson AGX Xavier is one of the most powerful developer kits produced by NVIDIA. It is mainly used for creating and deploying end-to-end AI robotics applications for manufacturing, delivery, retail, and agriculture, but it could also be applied for less intensive machine learning applications. It has a 512-core Volta GPU with Tensor Cores, an 8-core ARM v8.2 64-bit CPU, a 32 GB 256-Bit LPDDR4x memory, a 32 GB eMMC 5.1 Flash storage, as well as two USB C ports, and an HDMI and camera connector. It has a price of about USD 4000 and has a power consumption of 30 Watts, making it much more costly in both price and electricity than the other Jetson kits [27].
The Jetson Xavier NX kits is another series of NVIDIA developer kits designed as the successor to the TX series. It is power-efficient and compact, making it suitable for machine learning application development. It has an NVIDIA Volta architecture GPU with 384 NVIDIA CUDA® cores and 48 Tensor cores, a six-core NVIDIA Carmel ARM®v8.2 64-bit CPU, an 8 GB 128-bit LPDDR4x memory, two MIPI CSI-2 DPHY lanes cameras, and Ethernet, HDMI, and USB type A and Micro AB connection ports. It has an integrated storage component of its own, instead of relying on a micro SD storage interface. It has a power consumption of 10 Watts and a price range of around USD 2000. Its well-rounded quality makes it a very good, if somewhat expensive, the choice for machine learning implementation on embedded systems [28].
4.2. Google Coral
Google Coral Dev Board is a single-board computer by Coral that can be used to perform fast machine learning (ML) inferencing in a small form factor; it is mainly used for prototyping custom embedded systems, but it can also be used for embedded machine learning on its own. It has an Edge TPU coprocessor that is capable of performing 4 trillion operations per second, as well as being compatible with TensorFlow Lite. It has a quad Cortex-A53 CPU, integrated GC7000 Lite Graphics, 1 GB/2 GB/4 GB LPDDR4 memory, 8 GB eMMC storage as well as a MicroSD slot, Type C, A, and microB USB, Gigabit Ethernet, and HMDI 2.0 ports. The overall board has a low power cost of 6–10 Watts and at USD 130, the price for the board is relatively low [29].
4.3. Raspberry Pi
Raspberry Pi is a series of extremely popular embedded computers developed by the Raspberry Pi Foundation in the United Kingdom. The uses for these systems are extremely wide, including machine learning. Like the Jetson series, Raspberry Pi products are very commonly used in embedded machine-learning implementation projects. For this review, the three systems of Raspberry Pi that were commonly utilized were the Raspberry Pi 3 Model B, the Raspberry Pi 3 Model B+, and the Raspberry Pi 4 Model B.
The Raspberry Pi 3 Model B is the first iteration of the third-generation Raspberry Pi computers. It has a Quad Core 1.2 GHz Broadcom BCM2837 64bit CPU, a 400 MHz VideoCore IV video processor, a 1 GB LPDDR2 memory, a microSD port for storage, a 100 Base Ethernet, 4 USB 2.0, and full-size HDMI ports. It has an extremely low power consumption of 1.5 Watts and a monetary cost of about USD 40 [30].
The Raspberry Pi 3 Model B+ is the final iteration of the third-generation Raspberry Pi Computers. It has a Quad Core 1.4 GHz Broadcom BCM2837B0, Cortex-A53 (ARMv8) 64-bit SoC CPU, a 400 MHz VideoCore IV video processor, a 1 GB LPDDR2 memory, a microSD port for storage, a 1000 Base Ethernet, 4 USB 2.0, and full-size HDMI ports. Its main advantage to model 3b is its processor’s higher clock speed and its PoE (power over Ethernet) support. At 2 Watts, its power consumption is still low but higher than that of the model 3b series. It also has a very close monetary cost ranging around USD 40.
The Raspberry Pi 4 Model B is the first iteration of the fourth-generation Raspberry Pi Computer. It has a Quad Core 1.5 GHz Broadcom BCM2837B0, Cortex-A72 (ARMv8) 64-bit SoC CPU, a 400 MHz VideoCore IV video processor, a choice between 1 GB, 2 GB, 4 GB, and 8 GB LPDDR2 memory, a microSD port for storage, a Gigabit Ethernet, 4 USB 2.0, and full size HDMI ports. Its main advantage to model 3b is its processor’s higher clock speed and its PoE (power over Ethernet) support. Its newer processor and option for memory make it a superior choice compared to the previous iteration of Raspberry pi. It has a relatively low power consumption of 4 Watts and a monetary cost of about USD 40–USD 80 depending on the memory size [31].
4.4. ODROID XU4
The ODROID XU4 is an energy-efficient single-board embedded computing system by Hardkernel Co. located in Rm704 Anyang K Center 1591-9 Gwanyang-dong Dongan-gu, Anyang-si, Gyeonggi-do, South Korea. It is compatible with open-source software and can use different versions of Linux, such as Ubuntu, as its operating system. It has Exynos5422 Cortex™-A15 2 Ghz and Cortex™-A7 Octa core CPUs, a Mali-T628 MP6 GPU, a 2 GB LPDDR3 memory, 2 GB eMMC5.0 LPDDR3 Flash Storage as well as a microSD slot, 2 USB 3.0 and 1 USB 2.0, Gigabit Ethernet, and HMDI 1.4 ports. It has an operating power of 5 Watts and its cost is generally around USD 100 [32].
4.5. Banana Pi
Banana Pi is an open-source hardware platform by Shenzhen SINOVOIP Co. located in 7/F, Comprehensive Building of Zhongxing Industry City, Chuangye Road, Nanshan District, Shenzhen, China. Like other embedded systems, it has a wide range of applications, amongst them, embedded machine learning implementation. It has an H3 Quad-core Cortex-A7 H.265/HEVC 4K, a Mali400MP2 GPU, 1 GB DDR3 Memory, an 8 GB eMMC Onboard Storage, two USB 2.0 ports, an HDMI port, and an Ethernet interface. Its overall power consumption is about 5 Watts and it has a price range of USD 50–USD 75 [33].
4.6. ASUS Tinker Board
The ASUS Tinker Board S is a powerful SBC board with a wide range of functions such as computer vision, gesture recognition, image stabilization, and processing, as well as computational photography. It has a Rockchip Quad-Core RK3288 CPU, an ARM® Mali™-T764 GPU, a 2 GB Dual-Channel DDR3 Memory and 16 GB eMMC Onboard Storage 4 USB 2.0, and an Ethernet port, and RTL GB LAN connectivity. It has a maximum power consumption of 5 Watts and is a relatively low-price system for all of its capabilities ranging in price from USD 100–USD 150 [34].
The ASUS Tinker Edge R is specifically developed for AI applications, containing an integrated Machine Learning (ML) accelerator that speeds up processing efficiency, lowers power demands, and makes it easier to build connected devices and intelligent applications. It has an Arm® big.LITTLE™ A72+A53 Hexa-core CPU, an ARM® Mali™-T860 MP4 GPU, a 4 GB Dual-CH LPDDR4 memory on its system, and a 2 GB LPDDR3 on the Rockchip NPU, a 16 GB eMMC Flash Storage as well as a microSD slot, 3 USB 3.2 type A and 1 USB 3.2 Type C, Gigabit Ethernet, and HMDI ports. It can maintain a maximum power supply of 65 Watts and is a relatively lo- price system for all of its capabilities ranging in price from USD 200–USD 270 [38].
All of the inforamtion related to hardware specification has been summarised in Table 1.
5. Sensors
Electrical sensors are components responsible for gathering input from a given physical environment. The specific input that a sensor responds to varies from sensor to sensor could be temperature, ultrasound waves, light waves, pressure [39,40], or motion. Sensors do this by acting as switches in a circuit, controlling the flow of electric charges through their overall systems. Sensors can be split into two separate overarching categories, active sensors, and passive sensors. Active sensors emit their own radiation such as ultrasound waves and laser, from an internal power source, which is then reflected from the objects in the environment, the sensor then detects these reflections as inputs. radars are an example of active sensors. Passive sensors simply detect the radiation or signature emitted from their targets, such as body heat [41].
The most important characteristics of sensor performance are transfer function, sensitivity, span, uncertainty, hysteresis, noise, resolution, and bandwidth. The transfer function shows the functional relationship between the physical input signal and the electrical output signal. The sensitivity is defined in terms of the relationship between the input physical signal and the output electrical signal. The span is the range of input physical signals that may be converted to electrical signals by the sensor. Uncertainty is generally defined as the largest expected error between actual and ideal output signals. Hysteresis is the width of the expected error in terms of the measured quantity for sensors that do not return to the same output value when the input stimulus is cycled up or down. Output noise is generated by all sensors in addition to the output signal, and since there is an inverse relationship between the bandwidth and measurement time, it can be said that the noise decreases with the square root of the measurement time. The resolution is defined as the minimum detectable signal fluctuation. The bandwidth is the frequency range between the upper and lower cutoff frequencies, which respectively correspond to the reciprocal of the response and decay times [42].
Once sensors acquire input and convert it into electrical current, they can communicate their data to the rest of an overarching system through a variety of means, the main methods being to transfer data over a wired interface, or transfer data wirelessly [43,44]. Since the embedded systems studied in this research all made use of wired communication for their sensing systems, we focus only on analog communication. Standard wired interfaces between sensors and computing devices use serial ports, which transfer data between the data terminal equipment (DTE) and data circuit-terminating equipment (DCE). For successful data communication, the DTE and DCE must agree on a communication standard, the transmission speed, the number of bits per character, and whether stop and parity framing bits are used. Most modern-day computing devices and embedded systems use USB standards for their communication, connection, and power peripherals, which includes any additional sensor systems. USBs have had many port-type iterations since their inception; USB 1.x (up to 12 Mbps speed), USB 2.0 (up to 480 Mbps speed), USB 3.0 (up to 5 Gbps speed), and USB4 (super speed, up to 40 Gbps), most devices have ports for the USB 2.0 and USB 3.0 port types, with the USB4 being mostly suited for mobile smartphone devices. One of the main advantages of USB devices, including sensor systems, is that they can have multiple functionalities through a single connection port, for example, a USB camera can record both video and audio. These devices are referred to as composite devices and each of their functionalities is assigned to a specific address. USB devices can draw 5V and a maximum of 500mA from a USB host, allowing both data interface for sensor systems as well as powering the sensor component [45].
5.1. Sensor-to-Computation Pipeline
Once sensor systems receive input, they convert the input into digital data and transfer it to a display or a larger system. The format of the gathered data depends on the specific input a sensor collects, cameras would collect videos or images and microphones would collect audio. The environmental data collected by sensors are then stored within internal or external storage components connected to the overall system. These data are then used for whatever purpose the overall system that employed the sensor has been designed for.
As the focus of these research projects is over-viewing the capability of different embedded systems for running machine learning models, all of the sensor data are transferred to a previously trained machine learning algorithm or used to train a new algorithm based on existing architecture. In cases of trained model deployment, depending on the exact application of the model as well as its architecture, the stored data collected by the sensor systems is transferred to the model to perform predictions. For example, image identification and object recognition models will compare images files to the dataset images they have been trained with to either identify the specific objects of interest or the entire image, while forest biomass estimation models would compare the results gathered from lidar sensors to their trained dataset to estimate the concentration of vegetation in certain areas of forests [46].
5.2. Specific Sensors
Much like the different embedded computing systems that were used for machine learning implementation, many different sensors were used in each of our review sources depending on the application of the research. Not all sources made active use of a sensor within their work, and mainly explored the theoretical implementation of their machine-learning models using sensor systems. Amongst those that did implement their systems in some capacity, many implemented some form of object detection, image recognition, image segmentation, and other forms of computer vision, making extensive use of different integrated and separate image and video cameras. These cameras included infrared, RGB, Depth, Thermal, and 360-degree cameras. Other sensors used included microphones, electrocardiograms, radar, motion sensors, LIDAR, and multi-sensors.
5.2.1. RGB Cameras
RGB color cameras or visible imaging sensors are sensor systems that collect and store visible light waves as electrical signals that are then reorganized as rendered colored images. The images and videos they capture replicate human vision, capturing lightwave with (400–700) nm wavelength through light-sensitive electrical diodes, then saving them as pixels. Modern-day cameras can capture high-definition images [47]. The main use of these sensors is for object detection and image classification algorithms. Among the sources in this review, the main application in which an RGB camera was implemented included autonomous vehicles for pedestrian and sign detection, security cameras for intruder detection, facial recognition, and employee safety monitoring, and drones for search and rescue, domestic animal monitoring [48,49], agricultural crops, and wildlife observation [50].
5.2.2. Infrared Cameras
Infrared cameras or thermal imaging sensors are sensor systems that collect and store the heat signature that is emitted from objects as electronic images that show the apparent surface temperature of the captured object. They contain sensor arrays, consisting of thousands of detector pixels arranged in a grid on which infrared energy is focused. The pixels then generate an electrical signal that is used to create a color map image corresponding to the heat signature detected on an object ranging from violet to red, yellow, and finally white, with deep violet corresponding to the lowest detected heat signature and bright white corresponding to the highest detected heat signature [51]. In a similar sense to RGB cameras, the main use of these sensors is for object detection and image classification algorithms, albeit for more specialized tasks. Applications proposed by the sources in this review included autonomous vehicles for pedestrian detection, hand gesture, sign language, and facial expression recognition, thermal monitoring of electrical equipment, and profile recognition in smart cities.
5.2.3. Depth Cameras
Depth or range cameras are specific forms of sensor systems used to measure the exact three-dimensional depth of a given environment. They work by illuminating the scene with infrared light and measuring the time-of-flight. There are two operation principles for these sensors, pulsed light, and continuous wave amplitude modulation. In a sense, depth camera operation is very similar to Lidar, with it relying on infrared radiation reflection instead of laser [52]. The main application depth cameras used in among the sources of this paper were for quad-copter drone formation control, ripe coffee beans identification, and personal fall detection.
5.2.4. 360 Degree Cameras
360-degree cameras are sensor systems used to record images or video from all directions in 3D space using two over-180-degree cameras facing the front and rear of the device, the borders of the two images or videos are then stitched together to create a seamless single 360 image or video file. Users and automated applications can then select a specific section of the captured 360-image or footage for the intended use. Other than the over 180-field of view for each camera lens, 360 cameras work in an identical fashion to RGB cameras capturing visible spectrum light and storing it as digital data in pixel format [53,54]. While 360 cameras have various applications, from recreational ones such as vlogging and nature photography to navigational ones such as Google Maps, the sources used in this paper mainly relied on them for biometric recognition and marine life research.
5.2.5. Radar
RADAR, short for Radio Detecting And Ranging, is a radio transmission-based sensor system designed for detecting objects. They operate using short-pulse electromagnetic waves, these pulses are then reflected from objects in the path of the RADAR sensor and are then reflected back at it. Essentially, “When these pulses intercept precipitation, part of the energy is scattered back to the RADAR” [55]. RADAR systems can rely on 14 different frequency bands depending on the application. RADAR systems have a wide variety of applications, from meteorology to military surveillance and astronomical studies. Among the sources used for this review, RADAR systems were scarcely used, and within these cases, the main usage was for electric hybrid car deep learning-based car following systems as well as multi-target classification for security monitoring.
5.2.6. LiDar
Lidar (light detection and ranging) sensors are sensor systems that emit millions of laser waveforms and then collect their reflection to precisely measure the shape and distance of physical objects in a 3D environment. Essentially, they are laser-based radar systems. This process is repeated millions of times per second to create a precise real-time three-dimensional map of an area called a point cloud, which can then be used for navigation systems [56]. While the technology itself is decades old, with improvements in Lidar performance in terms of range detection, accuracy, power consumption, as well as physical features such as dimension and weight, its popularity has been rising in recent years, especially in the fields of robotics, navigation, remote sensing, and advanced driving assistance [57]. Lidars’ main usage among our sources was for locating people in danger in search and rescue operations, such as one following an earthquake, and optimizing trajectory tracking for small multi-rotor aerial drones.
5.2.7. Microphones
Microphones are sound sensors that act as transducers, converting sound waves into electrical current audio signals carrying the sound data. When sound waves interact with the microphone diaphragm, the vibrations created are converted into a coinciding audio signal via electromagnetic or electrostatic principles that will be outputted [58]. This audio signal can then be stored as digital data and replayed or used in other applications such as training sound recognition machine learning models. The sources presented in this review mainly used microphones for real-time speech source localization.
5.2.8. Body Motion Sensors
Body motion sensors, also known as motion capture sensors, are a series of sensor systems that are used to keep track of a person or a physical movement or physical posture. They generally work by making use of other sensing systems, including photosensors, angle sensors, IR sensors, optical sensors, accelerometers, inertial sensors [59], and magnetic bearing sensors [60]. Mocap sensors have been widely known for their use in the entertainment industry, but with recent advances, they have become more affordable and accurate for common consumer use. The application for which motion capture was used among the sources in this review is complex posture detection.
5.2.9. Electrocardiograms
Electrocardiograms are heart monitoring sensors used for quick analysis of a patient’s heart [61,62,63]. Heart contractions generate natural electrical impulses that are measurable by nonintrusive devices, such as lead wires placed on a patient’s skin. The measured pulses are then converted into an electric signal that can be used to measure irregularities in the patient’s heart rate [64]. Naturally, electrocardiograms are mainly used in medical facilities or by caregivers and nurses to monitor heart health [65,66], however, the sources used for this review have also utilized them for identifying epileptic seizures.
5.2.10. Electroencephalograms
Electroencephalograms are brain monitoring sensors used for analyzing a patient’s brain activity. The brain’s processes are the result of electrical current traveling through its neurons at varying levels depending on the current state of a patient, what they are doing, or how they are feeling. Electroencephalograms record these currents across the various brain regions using painless electrodes placed around a patient’s scalp. These fluctuations recordings are then saved as either a paper or digital graph [67]. Much like electrocardiograms, electroencephalograms are mainly used in medical facilities or by caregivers and nurses to monitor heart health, however, sources used for this review have also utilized them for anesthesia patient monitoring.
6. Applications
Embedded machine learning applications are all either of a remote nature or require more mobile systems to be implemented. The applications which are covered in this review are divided into the following categories: autonomous driving, security, personal health and safety, unmanned aerial vehicle navigation, and agriculture.
6.1. Autonomous Driving
Autonomous driving refers to the ever-expanding field of assisted and self-driving vehicles. It involves the implementation of a machine learning algorithm designed to detect obstacles, street signs, pedestrians, and other vehicles. Almost all self-driving vehicle AI models are computer vision models such as object and depth detection and distance measurement, with some exceptions that rely on Lidar or Radar for obstacle detection. Due to the nature of the application, the highest priority for models developed on embedded systems for self-driving vehicles is performance speed. Driving requires extremely short reaction time and that makes the speed at which a model can identify objects and allow the other car systems to make driving decisions very important.
6.2. Security and Safety
Security applications of machine learning can be related to many different sections such as intruder detection or personnel safety in hazardous worksites [68]. Once again, most of these models are trained for computer vision purposes in order to identify different individuals and ensure authorized access to secure locations and information. They do this through facial recognition and biometric identification using embedded system-operated camera systems, to name a few avenues. Ensuring personnel safety in hazardous work environments also involves constant monitoring by camera systems, to see if any of the employers are showing visible signs of illness or injury. Accuracy and computational speed are both of very high import in these applications.
6.3. Healthcare
Monitoring the health of hospital and nursing home patients is one of the fields in which machine learning has been found to be increasingly useful. The AI models trained for these purposes are varied depending on the exact nature of the task they are created to accomplish [69,70]. Applications involving the monitoring of the status of specific organs of patients can rely on various different medical equipment as well as visual and thermal cameras, such as monitoring a patient’s heart rate or brain activity, which are achieved with electrocardiograms and electroencephalograms. Fast performance of the machine learning models is of even greater importance in these scenarios as they can quite literally be about "life and death". Other health monitoring applications can refer to posture recognition and monitoring systems that rely on motion sensors and cameras to identify the posture of a given patient and inform their caretakers in case of any danger.
6.4. Drones
Aerial drones, or unmanned aerial vehicles, have a long history of military use, but have become increasingly utilized in everyday life over the past decade, be it for package delivery, remote video recording, wildlife research, or simply for recreational purposes. Many of these drones are of the quadcopter variety [71]. While most drones require remote piloting, there has been an increasing element of automation to their navigation [72,73], odometry, landing, and trajectory systems. AI models trained for these purposes use pathways, object images, and balance data models. While performance speed is an important factor for these models, accuracy takes far greater precedence as even the slightest misclassification can result in damage to or the destruction of the drone.
6.5. Agriculture
Different agricultural sectors have also started making use of machine learning. Object detection and facial recognition models are customized for recognizing individual animals during feeding and drinking to measure their overall consumption as well as monitor animal behavior and health. Object detection machine learning models are also used in farming crops for identifying weeds within the field, damaged crops, and crops ready for harvest, as well as any damage to the field and its fences. In both instances, the detection accuracy and energy consumption of the models are far more important than the performance speed.
7. Application Based System Comparison
As previously discussed, most review work on embedded machine learning has been focused on the implementation of modified ML architecture on specific embedded devices, whereas in this work, our focus is on identifying the advantages certain systems provide for specific applications and sensing schemes. For this purpose, we have divided our sources into the following categories with a summary of each presented in the Table 2, Table 3, Table 4, Table 5, Table 6, Table 7, Table 8, Table 9, Table 10, Table 11, Table 12 after the conclusion section. The systems are then compared by their performance and cost, the former being assessed differently depending on the task for which the machine learning model is trained. The method used for analyzing the performance is different from source to source and heavily dependent on the specific application and sensory system. Each sourced paper used a different method for analyzing model accuracy and inference speed. Alongside the power consumption, the mean of all the final results is used to assess the overall performance of each embedded system and presented in Figure 2, Figure 3, Figure 4, Figure 5, Figure 6, Figure 7, Figure 8 and Figure 9.
7.1. Image Recognition, Object Detection, and Computer Vision
As previously stated, different machine learning methods have been seeing an ever-increasing application within various fields, among these methods is the broad field of computer vision, which includes image and object detection. These applications can range from security and agriculture to autonomous vehicles—we have further divided these applications into the specific field in which they are applied.
7.1.1. Crop Identification
As previously discussed, like many other professions, machine learning has been seeing an increasing level of application within the field of crop and animal agriculture. This application can range from smart affordable farming solutions such as in [74] to the monitoring of ripened produce as in [75]. While time is valuable in any discipline, for agricultural machine learning applications, it is not nearly as much of a priority as power consumption and accuracy. Most of the applications covered in this review involve the usage of object recognition algorithms for the detection of various field or crop features but there are other applications that are analyzed as well. The performance of these applications is covered in Table 2 in addition to a comparison graph provided in Figure 2.
Figure 2Average inference time in agricultural computer vision for devices used in this application.
[Figure omitted. See PDF]
Table 2Computer Vision in Agriculture.
| Paper Title | Hardware | Application | Sensor | Accuracy | Power Consumption | Inference Time |
|---|---|---|---|---|---|---|
| [76] | ASUS Tinker Board S | Crop identification via aerial drone | Logitech C925e wWebcam | 89.44% | 8 Watts for both sensor and system | 0.7 s |
| [77] | Google Edge TPU, NVIDIA Jetson TX2 | Vineyard Landmark extraction for robot navigation in steep slope vineyard environment through vine trunk identification | Raspberry Pi infrared camera, Mako G-125C infrablue camera | 52.98% | 15 Watts for both sensor and system | 54.20 ms |
| [78] | Raspberry Pi 3 B+, with and without a neural compute stick, (Intel Movidius) NVIDIA Jetson Nano | Protect crops from ungulate attacks | Camera module (Raspberry Pi) | 62.41% | 10 Watts for both sensor and system (Jetson) 3.4 Watts for both sensor and system (RaPi) | 67.57 ms (Jetson) 1.25 s (RaPi) |
| [79] | NVIDIA Jetson Nano | Detection of ripe coffee beans | Intel realsense depth camera D435 | 97.23% | 14 Watts for both sensor and system | 17.49 ms |
| [80] | NVIDIA Jetson TX2 | Crop recognition for robotic weeding | Canon PowerShot SX150 IS camera | 95.9% | 12.5 Watts for both sensor and system | 8.9 ms |
| [81] | NVIDIA Jetson TX2 | Accurate weed detection for micro aerial vehicles | Multispectral camera | 79.9% | 15 Watts for both sensor and system | 0.56 s |
| [82] | Raspberry Pi 4 | Weed identification for herbicide | Raspberry Pi camera module version 2.0 with an 8-megapixel Sony IMX219 sensor | 96% | 6.88 Watts for both sensor and system | 0.167 s |
| [83] | NVIDIA Jetson TX2 | Loose fruit detection for oil palm | Camera | 94% | 10 Watts for both sensor and system | Not Stated |
| [84] | NVIDIA Jetson TX2 | Intelligent pest detection | High-resolution optical drone camera | 89.72% | 7.5 Watts | 114.89 ms |
7.1.2. Face and Expression Recognition
Facial recognition is one of the most well known applications in the field of computer vision—many personal projects, academic research studies, and computer applications have been developed regarding or using facial recognition. There are also many specialized models based on facial recognition, such as facial recognition models for animals [85], or facial expression recognition models that make use of existing facial recognition technologies as a baseline [86]. The priority in facial recognition models is dependent on the application as models used for security purposes would need to have both high accuracy and inference speed, while commercial application models are not under as much scrutiny. Most of the sources used in this review either implement facial recognition directly [87], or use it as a basis for emotion and personality assessment as well [85]. The performance of these applications is covered in Table 3 in addition to a comparison graph provided in Figure 3.
Figure 3Average Inference time in facial recognition for devices used in this application.
[Figure omitted. See PDF]
7.1.3. Depth Estimation
Depth estimation is a sub-field of machine learning that attempts to estimate depth within 2D images. It involves the use of pixel shape and orientation for the identification of the distance of objects within 2D images and video from the device that recorded it. Its utility is mainly in photography and depth estimation for self-driving vehicles, while within our sources, it was mostly used for personal projects such as in [88]. The performance of these applications is covered in Table 4 as well as a comparison graph being provided in Figure 4.
Table 3Computer Vision in Face Recognition.
| Paper Title | Hardware | Application | Sensor | Accuracy | Power Consumption | Inference Time |
|---|---|---|---|---|---|---|
| [86] | Banana Pi | Emotion and Personality Recognition | Thermal Camera (Vanadium Oxide Microbolometer with Chalcogenide Lens and a Field of View 36O.) | 87.87% | 4 Watts for both sensor and system | 3.851 s |
| [89] | Nvidia Jetson Nano, Nvidia Jetson TX2, Nvidia Jetson Xavier NX, Nvidia Jetson Xavier AGX | Facial recognition inference comparison between edge and cloud devices | None | 99.63% | 5 Watts (Nano) 7.5 Watts (TX2) 10 Watts (Xavier NX & AGX) | 0.37 s (Nano) 0.4 s (TX2) 0.18 s (Xavier NX) 0.28 s (AGX) |
| [2] | NVIDIA Jetson Nano | Analyze face structure from video feed and detect drowsiness from facial features | Webcam camera | 83.31% | 15 Watts for both sensor and system | 2 s |
| [90] | NVIDIA Jetson Nano | Face mask detection system | TGCAM-2000STAR camera | 99.02% | 17 Watts for both sensor and system | 30.18 ms |
| [87] | Raspberry Pi 3 model B | Facial biometric scan | Pi camera | 97.1% | 2.8 Watts for both sensor and system | 2.283 min |
| [91] | Raspberry Pi 4 | High-accuracy facial recognition | Webcam | 75.26% | 14 Watts for both sensor and system | 74.15 ms |
| [92] | Raspberry Pi 4 | Facial recognition and facial expression recognition | Logitech c270 camera | 98% | 14 Watts for both sensor and system | 71.14 ms |
| [93] | NVIDIA Jetson Nano, NVIDIA Jetson TX2 | Facial ID for security | Camera | 94% | 5 Watts (Nano) 7.5 Watts (TX2) | 0.1 s (Nano) 33.33 ms (TX2) |
| [94] | NVIDIA Jetson TX2 | Lightweight facial recognition for embedded systems | Camera | 58.7% | 1.4 Watts | 29 ms |
Computer Vision in Depth Estimation.
| Paper Title | Hardware | Application | Sensor | Accuracy | Power Consumption | Inference Time |
|---|---|---|---|---|---|---|
| [88] | NVIDIA Jetson TX1 | Monocular depth estimation (MDE) (estimating depth from a single image or video frame) | Camera | 78.3% | 5 Watts | 32.26 ms |
| [95] | ODROID XU4 NVIDIA Jetson TX2 | Collision checking for small aerial vehicles navigation | FLIR thermal imaging camera | 35.3% | 1.5 Watts (ODROID) 7.5 Watts (TX2) | 30 ms (ODROID) |
| [75] | ODROID XU4 | Computationally inexpensive misclassification minimization for aerial vehicles | D435i Depth Camera | 45.8% | 1.5 Watts 4.9 Watts for System and Sensor | 36.46 ms |
| [96] | NVIDIA Jetson Xavier NX | Depth estimation | Monocular camera | 87.8% | 10 Watts | 0.03 s |
| [97] | NVIDIA Jetson TX2 | Personal fall detection system | Image depth camera, RGB camera | 98% | 7.5 Watts | 66.67 ms |
Avg. inference time in depth estimation for devices used in this application.
[Figure omitted. See PDF]
7.1.4. Autonomous Vehicle Obstacle Recognition
One of the most widespread and focused implementations of machine learning, specifically, embedded machine learning, is in autonomous or assisted vehicles. Self-driving cars have been a staple of both science fiction and practical research for decades, but in the past decade, they have come increasingly close to reality. Advances in machine learning have been one of, if not the largest, driving factors behind this. While there are many different aspects of driving that a machine-earning algorithm could automate, from speed adjustment to the piloting of the vehicle in different directions, the focus in this review is mainly on the implementations of detection schemes for the various obstacles a vehicle can encounter, from other cars to pedestrians [98], road signs [99], traffic lights [5], and speed bumpers [11]. Due to the extremely dangerous nature of this application, systems used for these implementations need to be both as accurate and as fast as possible. The performance of these applications is covered in Table 5 in addition to a comparison graph provided in Figure 5.
Figure 5Average inference time in autonomous vehicle obstacle recognition in devices used in this application.
[Figure omitted. See PDF]
Table 5Computer Vision in Autonomous vehicles.
| Paper Title | Hardware | Application | Sensor | Accuracy | Power Consumption | Inference Time |
|---|---|---|---|---|---|---|
| [98] | ODROID XU4 NVIDIA Jetson Xavier | Nighttime pedestrian detection systems for cars | FLIR A325sc thermal camera | 75.7% | 1.5 Watts (ODROID) 10 Watts (Xavier) | 103 ms (ODROID) 43.3 ms (Xavier) |
| [5] | NVIDIA Jetson TX1, NVIDIA Jetson TX2 | Lightweight real-time traffic light detection for autonomous vehicles | AVT camera (only used for data collection) | 99.3% | 5 Watts (TX1) 7.5 Watts (TX2) | 83.3 ms (TX1) 71.4 ms (TX2) |
| [1] | NVIDIA Jetson TX2 | Road marking detection for autonomous vehicles | Camera | 96.9% | 7.5 Watts | 47 ms |
| [100] | NVIDIA Jetson TX2 | Lightweight road object detection for autonomous vehicles | Camera | 80.39% | 7.5 Watts | 31 ms |
| [101] | NVIDIA Jetson Xavier | Lightweight Multitask object detection and semantic segmentation for autonomous vehicles | N/A | 98.31% | 10 Watts | 17.36 ms |
| [102] | NVIDIA Jetson Xavier NX | Path Planning for self-driving vehicles and robotic systems | Camera | 93% | 10 Watts | 48.57 ms |
| [103] | NVIDIA Jetson Nano | Thermal object detection for assisted driving | LWIR prototype thermal camera | 86.6% | 5 Watts | 333.33 ms |
| [104] | NVIDIA Jetson Xavier NX | Road obstacle detection for vehicles | 20 Hz stereo camera | 98.1% | 10 Watts | 28.23 ms |
| [99] | NVIDIA Jetson TX1 | Traffic sign identification for smart vehicles | USB webcam | 96% | 5 Watts | 670 ms |
| [105] | NVIDIA Jetson AGX Xavier | Object detection and recognition and energy management for autonomous vehicles | N/A (can theoretically use onboard camera or radar) | 99.63% | 10 Watts | 260 ms |
| [106] | Raspberry Pi 3 Model B+ | Scalable and computationally cheap networks for autonomous driving | Raspberry Pi camera | 97.75% | 2.1 Watts | 3 ms |
| [11] | Raspberry Pi 3 Model B+ | Speed bump detection for autonomous vehicles | Raspberry Pi camera | 97.89% | 2.1 Watts | 104 ms |
| [107] | NVIDIA Jetson Nano | Algorithm review for self-driving car navigation | Mini camera IMX-219 | 80.5% | 5 Watts | Not Stated |
| [9] | NVIDIA Jetson TX1 | Real-time pedestrian detection for autonomous vehicles | Zed Stereo camera | 88.44% | 5 Watts | 33.3 ms |
| [108] | NVIDIA Jetson TX2 | Real-time vehicle detection on embedded systems | N/A | 85.6% | 7.5 Watts | 59.52 ms |
| [109] | NVIDIA Jetson AGX Xavier | Uncertainty-based real-time object detection for autonomous vehicles | Camera | 68.7% | 10 Watts | 14.35 ms |
7.1.5. Computer Vision in Medical Diagnosis and Disability Assistance
An interesting and beneficial application of computer vision is its use in the diagnosis of medical conditions and in assisting individuals with disabilities. Many of the sources presented in this review made use of RGB and thermal imaging of patients to perform object detection and image classification to find any signs of medical conditions such as melanoma [110] or diabetes [111], while others presented systems for assisting the visually impaired [112]. In both presented fields of application, while a very high accuracy is of extreme importance, a high inference speed is also paramount to any aides to special needs individuals. The result of these benchmarks is covered in Table 6 in addition to a comparison graph provided in Figure 6.
Figure 6Average inference time in medicine and disability assistance in devices used in these applications.
[Figure omitted. See PDF]
7.1.6. Computer Vision in Safety and Security
A more novel application of Computer vision models is its use in security systems as well as safety oversight networks. The sources presented in this section cover applications in detecting violent assaults [12] and mining personnel safety [3] to detecting survivors of severe natural disasters [113]. Most of these applications make use of RGB video and image cameras to perform detection and recognition. The result of these benchmarks is covered in Table 7 in addition to a comparison graph provided in Figure 7.
Figure 7Average inference time in safety and security in devices used in these applications.
[Figure omitted. See PDF]
Table 6Computer Vision in Medical and Special Aide Applications.
| Paper Title | Hardware | Application | Sensor | Accuracy | Power Consumption | Inference Time |
|---|---|---|---|---|---|---|
| [112] | NVIDIA Jetson TX2 | Visual aid system for the blind via real-time object detection | Webcam | 99.82% | 7.5 Watts | Not Stated |
| [114] | NVIDIA Jetson TX2 | Localize veins from color skin images. | 2-CCD multi-spectral prism camera (JAI AD-080-CL) | 78.27% | 7.5 Watts | 530 ms |
| [115] | Raspberry Pi 4, NVIDIA Jetson Xavier | COVID Identification through chest CT scans | CT Scanner | 98.8% | 4 Watts (Pi 4) 10 Watts (Xavier) | 23.3 s (Pi 4) 2.9 s (Xavier) |
| [116] | NVIDIA Jetson Nano | Posture recognition system for medical surveillance | RGB camera | 83% | 5 Watts | 476 ms |
| [117] | NVIDIA Jetson TX2 | Diabetes diagnosis | Jetson TX2 onboard camera | 91.8% | 7.5 Watts | 48 ms |
| [118] | Raspberry Pi 3 Model B+ | Reading assistance for blind people | Raspberry Pi camera module V2 | 100% | 2.1 Watts | 1 s |
| [110] | Raspberry Pi 3 Model B+ | Early skin cancer detection | IR camera | 98% | 2.1 Watts | 62 ms |
| [119] | Raspberry Pi | Cervical cancer prevention | PiCamera | 90% | Not Stated | 5.2 s |
| [120] | Raspberry Pi 4 Model B | Dog health monitoring through posture analysis | Smart camera network | 100% | 4 Watts | 69.24 s |
| [111] | NVIDIA Jetson Nano | Diabetic ulcer detection | Thermal Camera | 97.9% | 5 Watts | Unspecified |
| [121] | NVIDIA Jetson Xavier NX | Colonoscopy | Colonoscopy camera | 100% | 10 Watts | Unspecified |
| [122] | NVIDIA Jetson Nano | Travel assistance for the visually impaired | Optical RGB camera | 94.87% | 5 Watts | 22.22 ms |
| [123] | Raspberry Pi 3 Model B+ | Activity recognition for medical monitoring and rehab | Wearable Sensor | 96.63% | 2.1 Watts | 167.773 ms |
Computer Vision in Safety and Security Applications.
| Paper Title | Hardware | Application | Sensor | Accuracy | Power Consumption | Inference Time |
|---|---|---|---|---|---|---|
| [124] | Raspberry Pi | Sign language recognition | Thermal camera | 99.52% | Not Stated | 30 ms |
| [125] | NVIDIA Jetson Xavier NX | Proposal of a fast and accurate method of power line edge intelligent inspection | UAV camera | 55.6% | 10 Watts | 3.5 ms |
| [3] | NVIDIA Jetson TX1 | Production safety oversight in coal mines | Video Surveillance camera | 76.7% | 5 Watts | 27.25 ms |
| [126] | NVIDIA Jetson Nano | Passenger safety monitoring | 360◦ view camera | 85% | 5 Watts | Not Stated |
| [127] | NVIDIA Jetson TX2, NVIDIA Jetson Nano | Hard hat detection on construction site | Surveillance camera | 97.14% | 7.5 Watts (TX2) 5 Watts (Nano) | 68.03 ms (TX2) 111 ms (Nano) |
| [128] | NVIDIA Jetson TX2 | Detecting and tracking sinkholes via video streaming | Video camera | 90.61% | 7.5 Watts | 17 ms |
| [129] | NVIDIA Jetson TX2 | Concrete damage detection on the surface of buildings | Logitech Camera | 94.24% | 7.5 Watts | 33 ms |
| [130] | NVIDIA Jetson AGX Xavier | Railway defect detection | Camera | 93.5% | 10 Watts | 29.94 ms |
| [131] | Raspberry Pi 4 Model B | Biometric scan for entry control | Raspberry Pi NoIR camera | 97.2% | 4 Watts | Not Stated |
| [132] | Raspberry Pi 4 | Real-time fire detection | Camera | 97.5% | 4 Watts | 100 ms |
| [12] | Raspberry Pi 4 | Violent assault recognition | Surveillance camera (no actual live testing) | 92.05% | 4 Watts | 250 ms |
| [133] | Raspberry Pi 3 Model B+, Intel Neural Compute Stick 2 | Security surveillance | Surveillance camera | 94% | 2.1 Watts | 5.5 ms |
| [134] | NVIDIA Jetson Nano | Security surveillance for abnormal activity detection | Logitech C270 Camera | 89% | 5 Watts | 250 ms |
| [135] | NVIDIA Jetson Nano | Security surveillance for unusual behavior | HD camera | 97.5% | 5 Watts | Not Stated |
| [136] | NVIDIA Jetson Xavier NX | Fire and smoke detection | Camera | 100% | 10 Watts | 100 ms |
| [137] | NVIDIA Jetson TX2 | Monitoring vehicle driver tiredness in real time | Infrared Camera | 94% | 7.5 Watts | 45.45 ms |
| [138] | NVIDIA Jetson TX2 | Real-time security surveillance for acts of violence | RaspiCam camera, panoramic spherical camera | Not Stated | 7.5 Watts | 185 ms |
| [139] | NVIDIA Jetson Nano, Raspberry Pi 3 Model B+ | Rescue operation robot computer vision | No IR filter camera, LiDAR, Raspi Cam NOIR V2.1 | 78.6% | 7.5 Watts (Nano) 2.1 Watts (Pi 3) | 50 ms (Nano) 500 ms (Pi 3) |
| [140] | Raspberry Pi | CPU heat tracking | Infrared thermal sensor | 90.72% | Not Stated | 12.3 ms |
| [141] | NVIDIA Jetson Xavier NX | Real-time image processing for fusion diagnostics | Thermal image camera | Not Stated | 10 Watts | 48.97 ms |
| [142] | NVIDIA Jetson Nano | Automobile fog lamp intelligent control | IMX219 camera | 97.5% | 5 Watts | Not Stated |
| [113] | NVIDIA Jetson TX2 | Rescue of natural disaster survivors through drone object detection | Zenmuse XT2 gimbal camera | 61.97% | 7.5 Watts | 37.6 ms |
| [143] | NVIDIA Jetson Nano | Power system cyber security | N/A | 99.96% | 5 Watts | Not Stated |
7.1.7. Smart City Management
Smart cities are an increasingly used term within tech circles that refers to, among other things, the usage of machine learning and AI for the automation of many aspects of city management. Many of these applications are related to traffic management [14] or to the profiling of individuals [144]. It is very important for these models to be able to handle a large number of objects at any given time; for this reason, inference time is of a higher priority for these applications. Most of these applications make use of RGB video cameras to perform detection and recognition. The result of these benchmarks is covered in Table 8 as well as a comparison graph being provided in Figure 8.
Figure 8Average inference time in devices used in city management applications.
[Figure omitted. See PDF]
7.1.8. General Embedded Computer Vision
Many of the sources presented in this review could not fit into a large enough application category of their own. These sources ranged from works that were focused on the visual location of robotic limb grasping points [145] to ones studying the identification of individuals via their clothing [146]. For that purpose, these sources were all included within a generalized category presented in Table 9 as well as the comparison graphs shown in Figure 9.
Figure 9Average inference time in embedded computer vision devices.
[Figure omitted. See PDF]
Table 8Computer Vision in City Management.
| Paper Title | Hardware | Application | Sensor | Accuracy | Power Consumption | Inference Time |
|---|---|---|---|---|---|---|
| [14] | NVIDIA Jetson TX2 | Traffic flow detection and management | Canon EOS550D camera | 92% | 7.5 Watts | 26.39 ms |
| [147] | NVIDIA Jetson Nano | Real-time metro passenger volume enumeration | HD video recording camera | 97.1% | 5 Watts | 128.2 ms |
| [148] | Raspberry Pi 4 Model B | Smart Urban waste management | Pi Camera | 91.76% | 4 Watts | 358.9598 ms |
| [149] | Raspberry Pi 4 Model B | Garbage identification for recycling | Camera | 92.62% | 4 Watts | 630 ms |
| [144] | Raspberry Pi 3 Model B | Pedestrian profile recognition | FLIR Lepton thermal camera | 74.63% | 1.4 Watts | 111 ms |
| [150] | NVIDIA Jetson Nano | Car counter Traffic management | Logitech c922 webcam | Not Stated | 5 Watts | Not Stated |
| [151] | NVIDIA Jetson Nano | Smart city traffic management | Camera | 90% | 5 Watts | 25 ms |
| [152] | NVIDIA Jetson Nano | Visual garbage detection | N/A (most likely a video camera) | 94.56% | 5 Watts | 40 ms |
| [153] | NVIDIA Jetson Nano | AI traffic light control | Raspberry Pi camera | 90% | 5 Watts | Not Stated |
General Embedded Computer Vision.
| Paper Title | Hardware | Application | Sensor | Accuracy | Power Consumption | Inference Time |
|---|---|---|---|---|---|---|
| [146] | NVIDIA Jetson AGX Xavier | Person detection using top clothing | N/A | 92.57% | 10 Watts | 41.67 ms |
| [154] | NVIDIA Jetson TX1 | Detecting, tracking, and geolocating based on a monocular camera of an aerial drone | Monocular Camera | 97.6% | 5 Watts | 75.76 ms |
| [155] | NVIDIA Jetson TX2 | Drone detection | Spherical Camera (Ricoh Theta S) | 88.9% | 5 Watts | 33.33 ms |
| [156] | NVIDIA Jetson TX2 | Resource-constrained object tracking | N/A | 55% | 7.5 Watts | 72.89 ms |
| [157] | NVIDIA Jetson TX2 | Object detection and object tracking on drones with limited power and computational resources | Logitech BRIO camera | 90% | 7.5 Watts | 243.9 ms |
| [145] | NVIDIA Jetson Nano | Identifying and detecting suitable grasping point on objects for robotic limbs | A Basler acA2500-14uc industrial RGB camera with Computer M3514-MP lens | Not Stated | 5 Watts | 48 ms |
| [158] | NVIDIA Jetson TX2 | Navigation for indoor autonomous drones | Fisheye lens on the PointGrey Firefly camera | 75.5% | 7.5 Watts | 34.54 ms |
| [159] | NVIDIA Jetson TX2, NVIDIA Jetson Nano | Object detection via template tracking | N/A | Not Stated | 7.5 Watts (TX2) 5 Watts (Nano) | Not Stated |
| [160] | NVIDIA Jetson TX2 | Target tracking amongst static and dynamic obstacles | Drone camera | Not Stated | 7.5 Watts | Not Stated |
| [161] | NVIDIA Jetson TX2 | Underwater object gripping point detection | ZED binocular camera | Not Stated | 7.5 Watts | 90.09 ms |
| [162] | NVIDIA Jetson TX2 | Intelligent weapon targeting system | N/A | 68.9% | 7.5 Watts | 60 ms |
| [163] | NVIDIA Jetson AGX Xavier | Object recognition for unmanned surface vehicles | High-definition photoelectric vision sensor | 81.74% | 10 Watts | 37.36 ms |
| [164] | Raspberry Pi 3 Model B+ | Drone landing automation | Raspberry Pi v1.3 camera with a fisheye lens | Not Stated | 2.1 Watts | 37.36 ms |
| [10] | Raspberry Pi 3 model B | Image recognition for sea life | Pi Camera v2.1 | 89.81% | 1.4 Watts | 33.33 ms |
| [165] | Raspberry Pi 3 Model B+ | Image classification | N/A | 83.7% | 2.1 Watts | 180 ms |
| [166] | Raspberry Pi | Counting individuals within a given video feed | Camera | 90% | 1.4 Watts | Not Stated |
| [167] | Raspberry Pi | Fish recognition for underwater drones | 360 degrees panoramic camera | 87% | 1.4 Watts | 6 s |
| [168] | NVIDIA Jetson Nano | Identifying different plant species | Photo camera | 97.5% | 5 Watts | Not Stated |
| [169] | Nvidia Jetson Nano, Nvidia Jetson TX1, Raspberry Pi 4 | Artistic photography aesthetic score prediction | N/A | 91.02% | 5 Watts (Nano and TX1) 4 Watts (Pi 4) | 37 ms (Nano) 17.9 ms (TX1) 1.14 s (Pi 4) |
| [170] | NVIDIA Jetson Nano | Underwater object detection | N/A (visual camera in case of field testing) | 74.77% | 5 Watts | 125 ms |
7.2. Non-Vision-Related Machine Learning
Among the sources used for this review, a number were unrelated to any sub-field of computer vision and relied on different sensing schemes from LiDar [171] to ultrasound [13] for gathering training data and implementation, in applications from waste management [148] to heart monitoring [13]. While the sensing scheme and overall application of these models vastly differed from one another, their numbers for each application and sensor were not sufficient for a proper basis-by-basis comparison. For this reason, they are displayed within Table 10.
7.3. Embedded Machine Learning Optimization
Some of the sources in this review did not look into new applications of machine learning, but rather sought to optimize the performance of existing machine learning architecture on embedded system devices. The optimizations ranged from improving the effectiveness of image captioning models on the NVIDIA Jetosn TX2 [172] to pruning deep neural nets [173]. It should be noted that unlike the other sources in this review, most of these papers did not have sensing schemes. The result of these benchmarks is covered in Table 11 in addition to a comparison graph provided in Figure 10.
Figure 10Average inference time in devices used for testing model optimization methods.
[Figure omitted. See PDF]
7.4. Benchmarks, Reviews, and Machine Learning Enhancements
Among the sources used for this review, there were works of research that were not focused on the introduction of a specific application or a new method for the implementation of machine learning tasks for any field. These papers either attempted to perform benchmarks of different embedded system hardware via the implementation of specific machine learning architectures on them [20] or tried to augment the learning rate of machine learning models and implement their work on embedded computing systems [23]. While most of the work that fell into this category did not include any sensing schemes, the data gathered in them were highly relevant to this work and were for that reason included in this review. The result of these benchmarks are covered in Table 12 and a comparison graph is provided in Figure 11.
Table 10LiDar, Radar, Audio, and Motion Recognition Models.
| Paper Title | Hardware | Application | Sensor | Accuracy | Power Consumption | Inference Time |
|---|---|---|---|---|---|---|
| [13] | NVIDIA Jetson Nano, Raspberry Pi 3 | Early cardiovascular disease prevention through ultrasound | Ultrasound | 90.7 % | 5 Watts (Nano) |
2.78 ms (Nano) |
| [174] | Raspberry Pi 3 | Patient anesthesia monitoring | Electroencephalogram | 95% | 1.4 Watts | 20 ms |
| [175] | Raspberry Pi 3 | Human posture detection | Wireless body sensors (motion sensors, inertial sensors) | 98.28% | 1.4 Watts | 20 ms |
| [176] | NVIDIA Jetson Nano | Epileptic seizure detection | Electrocardiogram | 91.58% | 5 Watts | Not Stated |
| [177] | NVIDIA Jetson TX2 | Low-power multimodal data classification | Stand-alone dual-mode Tongue Drive System | 98% | 7.5 Watts | 1.6 ms |
| [178] | Raspberry Pi Model 3 | Driver behavior monitoring | IMU sensor, Shimmer Version 3 wearable body sensors | 73.02% | 1.4 Watts | 4.357 s |
| [179] | Raspberry Pi 3 Model B+ | Smart Urban waste management | Ultrasonic sensor | 88.43% | 2.1 Watts | 960 ms |
| [180] | Raspberry Pi 3 Model B | Fault detection in AC electrical systems | Photoelectric sensor | 99.37% | 1.4 Watts | 31 ms |
| [181] | Raspberry Pi 3 Model B+ | Target classification at road gates with radar SVM | Radar | Not Stated | 2.1 Watts | Not Stated |
| [182] | Raspberry Pi 3 Model B+ | Human activity recognition | Wearable multimodal sensors | 99.21% | 2.1 Watts | 153 ms |
| [183] | Raspberry Pi 3B+ | Speech recognition | Audio sensor | 96.82% | 2.1 Watts | 270 ms |
| [4] | Raspberry Pi 3B, NVIDIA Jetson TX1, NVIDIA Jetson TX2 | Psychological stress monitoring | Heart rate and accelerometer sensors | 96.7% | 1.4 Watts (Pi 3) 5 Watts (TX1) 7.5 Watts (TX2) | 189 ms (Pi 3) 2.8 ms (TX1) 4.7 ms (TX2) |
| [184] | Raspberry Pi 3 Model B | Motor fault diagnosis | Hall effect sensor | 97.05% | 1.4 Watts | 3.4 s |
| [185] | Raspberry Pi 4 Model B | Machine state monitoring | Vibration Sensor, Accelerometers | 98% | 4 Watts | 1.002 s |
| [186] | Raspberry Pi | Asthma risk prediction | SDS011 air quality sensor | 99% | 1.4 Watts | Not Stated |
| [8] | Raspberry Pi 3 Model B | Speech source identification | SSL sensors, microphones | 89.68% | 4 Watts | 21 ms |
| [187] | NVIDIA Jetson Nano | Battery charge management | GY169 current converter sensor module | RMSE of 1.976 | 5 Watts | Not Stated |
| [188] | NVIDIA Jetson TX2 | Food quality analysis | Nuclear magnetic resonance spectrometer, infrared spectrometer | 95% | 7.5 Watts | 4 ms |
| [189] | NVIDIA Jetson Nano | Pot plant species identification and watering needs monitoring | Capacitive Soil Moisture sensor, Water Level Sensor | Not Stated | 5 Watts | Not Stated |
| [190] | NVIDIA Jetson Nano | Radio frequency ID recognition | Universal software radio peripheral | 89.27% | 5 Watts | 18 min |
| [171] | NVIDIA Jetson Xavier NX | Trajectory tracking for small drones | Velodyne Lite 16 Lidar sensor | 83% | 10 Watts | 100 ms |
Embedded Machine Learning Optimization Papers.
| Paper Title | Hardware | Application | Sensor | Accuracy | Power Consumption | Inference Time |
|---|---|---|---|---|---|---|
| [172] | NVIDIA Jetson TX2 | Improve the effectiveness of Image Captioning | N/A | 65.7% | 7.5 Watts | 230 ms |
| [191] | NVIDIA Jetson TX2, NVIDIA Jetson Nano | Latency estimation on embedded systems | N/A | 96.39 % (Nano) 95.82 % (TX2)) | 5 Watts (Nano) 7.5 Watts (TX2) | 13.74 ms (Nano) 6.7 ms (TX2) |
| [192] | NVIDIA Jetson Nano | Real-time video analysis for edge computing | Video camera | 85% | 5 Watts | 11.21 ms |
| [193] | NVIDIA Jetson TX2 | Low-power and real-time deep learning-based multiple object visual tracking | 5MP CSI camera | N/A | 7.5 Watts | 100 ms |
| [173] | NVIDIA Jetson TX2 | Filter Pruning DNNs | N/A | 93.51% | 7.5 Watts | 8.01 ms |
| [194] | NVIDIA Jetson AGX Xavier | Energy-efficient acceleration of deep neural networks | N/A | N/A | 10 Watts | Not Stated |
| [195] | NVIDIA Jetson TX1 | Semantic Segmentation for autonomous vehicles | N/A | 87.3% | 5 Watts | 24 ms |
| [196] | NVIDIA Jetson TX2 | Improve semantic segmentation performance in contexts of various sizes and types in diverse environments | N/A | 92.74% | 7.5 Watts | 92.46 ms |
| [197] | NVIDIA Jetson TX2, Edge tensor processing unit, neural compute stick, and neural compute stick2 | Fusion Pruning DNNs | N/A | 90.66% | 7.5 Watts | 4.7 ms |
| [198] | NVIDIA Jetson TX2 | Reduce computational complexity and memory consumption of CNNs architecture on low-power devices | N/A | 93% | 7.5 Watts | 66.14 ms |
| [199] | NVIDIA Jetson TX2 | Reduce computational complexity and memory consumption of CNNs architecture on low-power devices | N/A | 99.3% | 7.5 Watts | 894.85 ms |
| [200] | NVIDIA Jetson AGX Xavier | Improve embedded system performance in autonomous vehicles | N/A | 98.3% | 10 Watts | 690 ms |
| [201] | NVIDIA Jetson TX1 | Provide a less resource costly object detection model for embedded systems | N/A | 65.7% | 5 Watts | 135.2 ms |
| [202] | NVIDIA Jetson Nano | Efficient video understanding | Video camera | 74.1% | 5 Watts | 13.51 ms |
| [106] | Raspberry Pi 3 Model B+ | Scalable and computationally cheap networks for autonomous driving | Raspberry Pi camera | 75.78% | 5 Watts | 284 ms |
Benchmark and Review Papers.
| Paper Title | Hardware | Application | Sensor | Accuracy | Power Consumption | Inference Time |
|---|---|---|---|---|---|---|
| [23] | NVIDIA Jetson Nano, Coral Edge TPU, custom convolutional neural network accelerator | Enhance learning rate for ML model with smaller training datasets | N/A (Benchmark paper) | 49.6% (Nano) 49.8% (TPU) | 5 Watts (Nano) 2 Watts (TPU) | 0.3294 s (Nano) 19.8 ms (TPU) |
| [20] | NVIDIA Jetson Nano, NVIDIA Jetson AGX Xavier | Benchmark analysis of 3d object detection | USB attached video camera (Benchmark paper) | 70% | 5 Watts (Nano) 10 Watts (AGX) | 0.56 s (Nano) 47.61 ms (AGX) |
| [18] | NVIDIA Jetson Nano, NVIDIA Jetson TX2, Raspberry PI 4 | Performance analysis of different hardware for object detection CNNs | N/A (Benchmark paper) | 93.8 % (Nano) 93.9% (TX2) 91.6% (Pi) | 5 Watts (Nano) 7.5 Watts (TX2) 4 Watts (Pi) | 58 s (Nano) 32 s (TX2) 372 s (Pi) |
| [19] | NVIDIA Jetson TX1 | Analysis of DNN architecture in image recognition | N/A (Benchmark paper) | 69.52% | 5 Watts | 10.55 ms |
| [15] | Asus Tinker Edge R, Raspberry Pi 4, Google Coral Dev Board, NVIDIA Jetson Nano | Presentation and comparison of the performance of the presented systems in terms of inference time and power consumption | N/A (Benchmark paper) | 92.5% | 4.75 Watts (Tinker) 2.75 Watts (Coral) 2.1 Watts (Pi) 0.9 Watts (Nano) | 0.33 s (Tinker) 0.28 s (Coral) 0.21 s (Pi) 0.137 s (Nano) |
| [22] | Raspberry Pi 4 | Space exploration landing site selection | N/A (dataset acquired from images taken by the Mars HiRISE camera) | 95% | 4 Watts | 89 ms |
| [21] | NVIDIA Jetson Nano, NVIDIA Jetson TX1, NVIDIA Jetson AGX Xavier | Benchmarking paper | N/A | Accuracy Rates Not Stated | 5 Watts (Nano & TX1) 10 Watts (AGX) | 94 ms (Nano) 84 ms (TX1) 46 ms (AGX) |
| [17] | NVIDIA Jetson TX2, NVIDIA Jetson Xavier NX, and NVIDIA Jetson AGX Xavier | Benchmarking NVIDIA Jetson systems for visual odometry of flying drones | N/A | Accuracy Rates Not Stated | 7.5 Watts (TX2) 10 Watts (NX & AGX) | Speed Rates Not Stated |
Average inference time in devices covered in referenced benchmark papers.
[Figure omitted. See PDF]
8. Conclusions
Rapid advances have been made in the field of machine learning, causing an explosion of model variety, application, and performance. While many of these models are implemented on powerful stationary computer devices, there are many applications that are faced with cost, power, and size limitations for the specific usage of their models. For this reason, the field of embedded machine learning, which is the implementation of machine learning on embedded computing systems, has also faced a great deal of attention recently. The main challenges faced in embedded machine learning are caused by the severe limitations of embedded system devices in terms of computational performance and power, with different devices having different performances, power requirements, and purchasing costs. In this review, a large collection of research work and implementation of embedded machine learning on Raspberry Pi, NVIDIA Jetson, and a few other series of devices is presented alongside the overall power consumption, inference time, and accuracy of these implementations. In addition, unlike many other reviews of this topic, this paper also includes a presentation of the overall sensing scheme present in many of the works. It was believed that this was a major dimension of embedded machine learning study overlooked by most other reviews on the subject matter. The hope of this review is to familiarize interested researchers in the field of embedded machine learning by giving them a general introduction to it.
Overall, this study contained studies of several generations of embedded systems, specifically, the Nvidia Jetson and Raspberry Pi systems, showing that much like dedicated computing systems, embedded devices have been experiencing steady improvements in the fields of performance and power consumption. More recent Jetson boards such as the TX2 have a far higher performance rate compared to the TX1 while having the same power consumption levels. As these advances continue, it stands to reason that embedded machine learning will see even greater attention and become even more widespread. All of the systems discussed in this work have their own distinct advantages and disadvantages that users would need to consider when choosing a system for their embedded machine learning application. More robust systems with high performance and relatively efficient power usage such as the Jetson Board and Coral Dev Board line tend to be more monetarily expensive, while more affordable options such as the Raspberry and Banana Pi boards tend to have far lower performances. More remote applications such as agricultural object detection systems might need a greater number of low-power systems while not having much emphasis on performance, while autonomous vehicle applications would have a far greater emphasis on performance and accuracy than on cost and power usage. A general table of all sources’ hardware, application, ML architecture, sensor is provided in Table 13 for interested readers.
Conceptualization was performed by W.T. and A.B. Validation of the research was performed by W.T. Investigation of sources for the review was completed by A.B. Resources were identified by W.T. and A.B. Writing of the original draft of the paper was done by A.B. Final review and editing were completed by W.T. Supervision over the research was provided by W.T. All authors have read and agreed to the published version of the manuscript.
Not applicable.
The authors declare no conflict of interest.
The following abbreviations are used in this manuscript:
| ADAS | Advanced Driver-Assistance System |
| AI | Artificial Intelligence |
| ANN | artificial neural network |
| API | Application Programming Interface |
| BDR | Break Down Rate |
| CNN | Convolutional Neural Network |
| CPU | Central Processing Unit |
| CSI | Camera Serial Interface |
| CT | Computerized Tomography |
| DCE | Data Circuit-terminating Equipment |
| DCNN | Deep Convolutional Neural Network |
| DNN | Deep Nerual Network |
| DRL | Deep Reinforcement Learning |
| DTE | Data Terminal Equipment |
| FCN | Fully Convolutional Network |
| FLIR | Forward Looking InfraRed |
| GPU | Graphical Processing Unit |
| GRU | Gated Recurrent Unit |
| IR | Infra-Red |
| KNN | K-Nearest Neighbors |
| L4T | Linux for Tegra |
| LFFD | Light and Fast Face Detector |
| LGHP | Local Gradient Hexa Pattern |
| LSTM | Long Short-Term Memory |
| LiDAR | Light Detection And Ranging |
| MDE | Monocular Depth Estimation |
| ML | Machine Learning |
| MLP | Multilayer Perceptron |
| MMSN | Multi-Mapping Spherical Normalization |
| MPC | Model Predictive Control |
| MTCNN | Multi-Task Cascaded Convolutional Neural Network |
| MoCap | Motion Capture |
| OCR | Optical Character Recognition |
| OS | Operating System |
| RAM | Random Access Memory |
| RAM | Random Access Memory |
| RCNN | Region-Based Convolutional Neural Network |
| RGB | Red Green Blue |
| RNN | Recurrent Neural Network |
| RPN | Region Proposal Network |
| RaDAR | Radio Detecting And Ranging |
| SDK | Software Development Kit |
| SSD | Single Shot Detector |
| SVM | Support Vector Machine |
| TPU | Tensor Processing Unit |
| TSM | Temporal Shift Module |
| UAV | Unmanned Aerial Vehicle |
| USB | Universal Serial Bus |
| VP-CNN | Vein and Periocular Pattern-based Convolutional Neural Network |
| YOLO | You Only Look Once |
Footnotes
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Figure 1. Paper Layout Showing the Distribution of Subjects Covered in the Review.
Hardware specifications.
| Hardware | Processor | RAM | Storage | Power | Maker |
|---|---|---|---|---|---|
| ASUS Tinker Board S | Rockchip Quad-Core RK3288 Processor | 2 GB Dual-Channel DDR3 Memory | 16 GB eMMC Onboard Storage | 5 W | Asus |
| Banana Pi BPI-M2+ | H3 Quad-core Cortex-A7 H.265/HEVC 4K | 1 GB DDR3 Memory | 8 GB eMMC Onboard Storage | 5 W | Shenzhen SINOVOIP Co. |
| Coral TPU Dev Board | NXP i.MX 8M Quad-core Cortex-A53 | 1 GB LPDDR4 Memory | 8 GB eMMC Onboard Storage | (6–10) W | Coral |
| ODROID-XU4 Board | Exynos5422 Cortex-A15 2 Ghz, Cortex™-A7 Octa core | 2 GB LPDDR3 Memory | Flash Storage Interface | 15 W | Hardkernel Co. |
| ASUS Tinker Edge R | Cortex-A72, Cortex-A53, Mali-T860 | 4 GB LPDDR4 Memory | 16 GB eMMC Onboard Storage | 65 W | ASUS |
| NVIDIA Jetson Nano | ARM Cortex-A57 MPCore | 4 GB 64-bit LPDDR4 | 16 GB eMMC 5.1 Onboard Storage | (5–10) W | NVIDIA |
| NVIDIA Jetson TX1 | 4 Core ARM Cortex-A57 MPCore | 4 GB 64-bit LPDDR4 | 16 GB eMMC 5.1 Onboard Storage | 15 W | NVIDIA |
| NVIDIA Jetson TX2 | 6 Core ARM Cortex-A57 MPCore | 8 GB 64-bit LPDDR4 | 16 GB eMMC 5.1 Onboard Storage | 25 W | NVIDIA |
| NVIDIA Jetson AGX Xavier | 8 Core ARM v8.2 64-bit MPCore | 16 GB 256-Bit LPDDR4x | 32 GB eMMC 5.1 Onboard Storage | (10–30) W | NVIDIA |
| NVIDIA Jetson Xavier NX | 6 Core NVIDIA Carmel ARM v8.2 64-bit MPCore | 8 GB 128-bit LPDDR4x | microSD storage interface | 10 W | NVIDIA |
| Raspberry Pi 3 Model B | 1.2 GHz Broadcom BCM2837 (64 Bit) | 1 GB LPDDR2 | microSD storage interface | (1.3–1.4) W | Raspberry Pi Foundation |
| Raspberry Pi 3 Model B+ | 1.2 GHz Quad-Core ARM Cortex-A53 (64 Bit) | 1 GB LPDDR2 | microSD storage interface | (1.9–2.1) W | Raspberry Pi Foundation |
| Raspberry Pi 4 Model B | 1.2 GHz Quad-Core ARM Cortex-A72 (64 Bit) | (1/2/4) GB LPDDR2 | microSD storage interface | (3.8–4) W | Raspberry Pi Foundation |
Hardware specifications.
| Paper Title | Hardware | Application | ML Architecture | Sensor |
|---|---|---|---|---|
| [ |
ASUS Tinker Board S | Crop identification via aerial drone | SegNet, FCN-AlexNet | Logitech C925e webcam |
| [ |
Banana Pi | Emotion and Personality Recognition | Hidden Markov Model | Thermal Camera (Vanadium Oxide Microbolometer with Chalcogenide Lens and a Field of View 36O.) |
| [ |
NVIDIA Jetson Nano, Coral Edge TPU, custom convolutional neural network accelerator | Enhance learning rate for ML model with smaller training datasets | Siamese Neural Network | N/A |
| [ |
NVIDIA Jetson TX1 | Monocular depth estimation (MDE) (estimating depth from a single image or video frame) | Separable Pyramidal pooling Encoder-Decoder (Custom Architecture) | Camera |
| [ |
Google Edge TPU, NVIDIA Jetson TX2 | Vineyard Landmark extraction for robot navigation in steep slope vineyard environment through vine trunk identification | MobileNet V1, MobileNet V2 | Raspberry Pi infrared camera, Mako G-125C infrablue camera |
| [ |
ODROID XU4, NVIDIA Jetson Xavier | Nighttime pedestrian detection systems for cars | YOLOv2 | FLIR A325sc thermal camera |
| [ |
ODROID XU4, NVIDIA Jetson TX2 | Collision checking for small aerial vehicles navigation | Custom pyramid-based spatial partitioning | FLIR thermal imaging camera |
| [ |
ODROID XU4 | Computationally inexpensive misclassification minimization for aerial vehicles | Siamese Neural Network | D435i Depth Camera |
| [ |
NVIDIA Jetson Nano, NVIDIA Jetson AGX Xavier | Benchmark analysis of 3D object detection | Complex YOLOv3, Complex YOLOv4 | USB attached video camera (Benchmark paper) |
| [ |
NVIDIA Jetson Nano, NVIDIA Jetson TX2, Raspberry PI4 | Performance analysis of different hardware for object detection CNNs | Custom Deep-CNN | N/A (Benchmark paper) |
| [ |
NVIDIA Jetson TX1 | Analysis of DNN architecture in image recognition | AlexNet, GoogLeNet, SENet, MobileNet | N/A (Benchmark paper) |
| [ |
Asus Tinker Edge R, Raspberry Pi 4, Google Coral Dev Board, NVIDIA Jetson Nano | Presentation and comparison of the performance of the presented systems in terms of inference time and power consumption | MobileNetV2, MobileNetV2 Lite, MobileNetV2 Quant. Lite | N/A (Benchmark paper) |
| [ |
NVIDIA Jetson TX2 | Visual aid system for the blind via real-time object detection | CNN YOLOv2 | Webcam |
| [ |
NVIDIA Jetson Xavier NX | Proposal of a fast and accurate method of power line edge intelligent inspection | RepYOLO, YOLOv5 | UAV camera |
| [ |
NVIDIA Jetson Nano, Raspberry Pi 3 | Early cardiovascular disease prevention through ultrasound | DNN (custom models for different tasks) | Ultrasound |
| [ |
NVIDIA Jetson Nano | Passenger safety monitoring | DNN (YOLO, SSD) | 360◦ view camera |
| [ |
NVIDIA Jetson TX1 | Production safety oversight in coal mines | FL-YOLO | Video surveillance camera |
| [ |
NVIDIA Jetson TX2 | Traffic flow detection and management | YOLOv3, DeepSORT | Canon EOS550D camera |
| [ |
NVIDIA Jetson TX2 | Improve the effectiveness of image captioning | Captioning. BDR-GRU | N/A |
| [ |
Raspberry Pi 4, NVIDIA Jetson Xavier | COVID Identification through chest CT scans | Anam-Net | CT Scanner |
| [ |
NVIDIA Jetson TX2, NVIDIA Jetson Nano | Latency estimation on embedded systems | AlexNet, VGG16 ResNet-50, MobileNetV2 | N/A |
| [ |
Nvidia Jetson AGX, Raspberry Pi 4 | Hand gesture recognition | Custom Deep CNN model | Thermal camera |
| [ |
Nvidia Jetson Nano, Nvidia Jetson TX2, Nvidia Jetson Xavier NX, Nvidia Jetson Xavier AGX | Facial recognition inference comparison between edge and cloud devices | MTCNN detector, FaceNet | None |
| [ |
NVIDIA Jetson AGX Xavier | Person detection using top clothing | Mask-R-CNN, YOLACT++ | N/A |
| [ |
NVIDIA Jetson TX1, NVIDIA JetsonTX2 | Lightweight real-time traffic light detection for autonomous vehicles | Lightweight Convolution Neural Network | AVT camera (only used for data collection) |
| [ |
NVIDIA Jetson Nano | Real-time video analysis for edge computing | Custom architecture consisting of Front-CNN and Back-CNN | Video camera |
| [ |
NVIDIA Jetson TX2 | Low-power and real-time deep learning-based multiple object visual tracking | CNN-based custom architecture | 5MP CSI camera |
| [ |
NVIDIA Jetson TX2 | Localize veins from color skin images. | CNN | 2-CCD multi-spectral prism camera (JAI AD-080-CL) |
| [ |
Raspberry Pi 3 B+, with or without a neural compute stick (Intel Movidius), NVIDIA Jetson Nano | Protect crops from ungulate attacks | YOLO, Tiny-YOLO | Camera module (Raspberry Pi) |
| [ |
NVIDIA Jetson TX2 | Detecting and tracking sinkholes via video streaming | Cascaded CNN | Video camera |
| [ |
NVIDIA Jetson Nano | Analyze face structure from video feed and detect drowsiness from facial features | OpenCV facial recognition | Webcam camera |
| [ |
NVIDIA Jetson TX1 | Detecting, tracking, and geolocating based on a monocular camera of an aerial drone | YOLOv3 | Monocular Camera |
| [ |
NVIDIA Jetson TX2 | Drone detection | YOLOv3 | Spherical Camera (Ricoh Theta S) |
| [ |
NVIDIA Jetson TX2 | Filter Pruning DNNs | VGG-16, ResNet-56, LeNet, FCNet-120 | N/A |
| [ |
NVIDIA Jetson TX2 | Resource constrained object tracking | CNN | N/A |
| [ |
NVIDIA Jetson AGX Xavier | Energy-efficient acceleration of deep neural networks | DNN | N/A |
| [ |
NVIDIA Jetson TX2 | Road marking detection for autonomous vehicles | CNN | Camera |
| [ |
NVIDIA Jetson TX1 | Semantic Segmentation for autonomous vehicles | DNN | N/A |
| [ |
NVIDIA Jetson TX2 | Improve semantic segmentation performance in contexts of various sizes and types in diverse environments | Segmentation CNN | N/A |
| [ |
NVIDIA Jetson Nano | Face mask detection system | CNN | TGCAM-2000STAR camera |
| [ |
NVIDIA Jetson Xavier NX | Depth estimation | FastMDE custom model | monocular camera |
| [ |
NVIDIA Jetson TX2, Edge tensor processing unit, neural compute stick, and neural compute stick2 | Fusion Pruning DNNs | DNN | N/A |
| [ |
NVIDIA Jetson TX2 | Object detection and object tracking on drones with limited power and computational resources | CNN | Logitech BRIO camera |
| [ |
NVIDIA Jetson Nano | Detection of ripe coffee beans | CNN | Intel realsense depth camera D435 |
| [ |
NVIDIA Jetson TX2 | Intelligent pest detection | Tiny-YOLOv3 | High-resolution optical drone camera |
| [ |
NVIDIA Jetson TX2 | Personal fall detection system | Gaussian mixture model (GMM) | Image depth camera, RGB camera |
| [ |
NVIDIA Jetson TX2 | Reduce computational complexity and memory consumption of CNNs architecture on low-power devices | Light-YOLOv4 | N/A |
| [ |
NVIDIA Jetson TX2 | Reduce computational complexity and memory consumption of CNNs architecture on low-power devices | CNN | N/A |
| [ |
NVIDIA Jetson Nano | Identify and detect suitable grasping point on objects for robotic limbs | ASP U-Net (DCNN) | A Basler acA2500-14uc industrial RGB camera with Computer M3514-MP lens |
| [ |
NVIDIA Jetson TX2 | Lightweight road object detection for autonomous vehicles | CNN | Camera |
| [ |
NVIDIA Jetson Xavier | Lightweight Multitask object detection and semantic segmentation for autonomous vehicles | DCNN | N/A |
| [ |
NVIDIA Jetson Xavier NX | Path Planning for self-driving vehicles and robotic systems | LSTM | Camera |
| [ |
NVIDIA Jetson Nano | Thermal object detection for assisted driving | Thermal-YOLO | LWIR prototype thermal camera |
| [ |
NVIDIA Jetson AGX Xavier | Improve embedded system performance in autonomous vehicles | DNN | N/A |
| [ |
NVIDIA Jetson Xavier NX | Trajectory tracking for small drones | MPC | Velodyne Lite 16 Lidar sensor |
| [ |
NVIDIA Jetson TX2 | Navigation for indoor autonomous drones | SSD | Fisheye lens on the PointGrey Firefly camera |
| [ |
NVIDIA Jetson TX2, NVIDIA Jetson Nano | Object detection via template tracking | OpenCV | N/A |
| [ |
NVIDIA Jetson Nano | Epileptic seizure detection | DNN | Electrocardiogram |
| [ |
NVIDIA Jetson Nano | Posture recognition system for medical surveillance | MobilenetV2, LSTM | RGB camera |
| [ |
NVIDIA Jetson TX2 | Concrete damage detection on the surface of buildings | YOLO-v3 | Logitech Camera |
| [ |
NVIDIA Jetson TX2 | Crop recognition for robotic weeding | ResNet-10 | Canon PowerShot SX150 IS camera |
| [ |
NVIDIA Jetson AGX Xavier | Railway defect detection | TensorRT | Camera |
| [ |
NVIDIA Jetson Nano | Real-time metro passenger volume enumeration | CircleDet | HD video recording camera |
| [ |
NVIDIA Jetson TX2 | Target tracking amongst static and dynamic obstacles | Model Predictive Control (MPC) | Drone camera |
| [ |
NVIDIA Jetson TX2 | Underwater object gripping point detection | real-time lightweight object detector (RLOD) | ZED binocular camera |
| [ |
NVIDIA Jetson Xavier NX | Road obstacle detection for vehicles | Siamese Neural network | 20 Hz stereo camera |
| [ |
NVIDIA Jetson TX2 | Intelligent weapons targeting system | YOLOv5 | N/A |
| [ |
NVIDIA Jetson TX1, NVIDIA Jetson TX2, NVIDIA Jetson TK1 | Review of assisted driving in resource constrained hardware | ADAS | N/A |
| [ |
NVIDIA Jetson TX2 | Diabetes diagnosis | R-CNN with InceptionV2 | Jetson TX2 onboard camera |
| [ |
NVIDIA Jetson TX2, NVIDIA Jetson Xavier NX, and NVIDIA Jetson AGX Xavier | Benchmarking NVIDIA Jetson systems for visual odometry of flying drones | VINS-Mono, VINS-Fusion, Kimera, ALVIO, Stereo-MSCKF, ORB-SLAM2 stereo, and ROVIO | N/A |
| [ |
NVIDIA Jetson TX2 | Low-power multimodal data classification | DCNN | Stand-alone Dual-mode Tongue Drive System |
| [ |
NVIDIA Jetson TX1 | Provide a less resource costly object detection model for embedded systems | Tiny-YOLO-V3, Tinier-YOLO | N/A |
| [ |
NVIDIA Jetson Nano | Power system cyber security | recurrent neural networks (RNN) | N/A |
| [ |
NVIDIA Jetson TX1 | Traffic sign identification for smart vehicles | deep convolutional neural network (DCNN) | USB webcam |
| [ |
NVIDIA Jetson Nano | Efficient video understanding | Temporal Shift Module (TSM) | Video camera |
| [ |
NVIDIA Jetson TX2 | Rescue of natural disaster survivors through drone object detection | YOLOV3, YOLOV3-MobileNetV1, YOLOV3-MobileNetV3 | Zenmuse XT2 gimbal camera |
| [ |
NVIDIA Jetson AGX Xavier | Object detection and recognition and energy management for autonomous vehicles | Deep reinforcement learning (DRL), YOLO | N/A (can theoretically use onboard camera or radar) |
| [ |
NVIDIA Jetson AGX Xavier | Object recognition for unmanned surface vehicles | YOLOv4, Siamese-RPN | High-definition photoelectric vision sensor |
| [ |
NVIDIA Jetson TX2 | Accurate weed detection for micro aerial vehicles | SegNet | Multispectral camera |
| [ |
Raspberry Pi 4 | Smart Urban waste management | SSD MobileNetV2 | Pi Camera |
| [ |
Raspberry Pi 4B | Garbage identification for recycling | MobileNetV3 | Camera |
| [ |
Raspberry Pi 4 Model B | Biometric scan for entry control | Vein and Periocular Pattern-based Convolutional Neural Network (VP-CNN). | Raspberry Pi NoIR camera |
| [ |
Raspberry Pi 4 | Real time fire detection | CNN | Camera |
| [ |
Raspberry Pi 3 | Patient anesthesia monitoring | DNN | Electroencephalogram |
| [ |
Raspberry Pi 3 | Human posture detection | Multi-Mapping Spherical Normalization (MMSN) | Wireless body sensors (motion sensors, inertial sensors) |
| [ |
Raspberry Pi 3 Model B+ | Reading assistance for blind people | OCR CNN | Raspberry Pi camera module V2 |
| [ |
Raspberry Pi 3 | Driver behavior monitoring | DCNN | IMU sensor, Shimmer Version 3 wearable body sensors |
| [ |
Raspberry Pi 3 Model B+ | Scalable and computationally cheap networks for autonomous driving | DNN | Raspberry Pi camera |
| [ |
Raspberry Pi 3 Model B+ | Early skin cancer detection | CNN | IR camera |
| [ |
Raspberry Pi 3 Model B+ | Smart Urban waste management | Keras | Ultrasonic sensor |
| [ |
Raspberry Pi 3B | Fault detection in AC electrical systems | ArcNet (CNN) | Photoelectric sensor |
| [ |
Raspberry Pi 4B | Space exploration landing site selection | SegNet, FCN | N/A (dataset acquired from images taken by the Mars HiRISE camera) |
| [ |
Raspberry Pi 3 Model B+ | Target classification at road gates with radar | SVM | Radar |
| [ |
Raspberry Pi 3 Model B+ | Activity recognition for medical monitoring and rehab | CNN | Wearable Sensor |
| [ |
Raspberry Pi | Sign language recognition | CNN | Thermal camera |
| [ |
Raspberry Pi 3+ | Speed bump detection for autonomous vehicles | CNN | Raspberry Pi camera |
| [ |
Raspberry Pi 3B+ | Human activity recognition | CNN | Wearable multimodal sensors |
| [ |
Raspberry Pi 3B+ | Drone landing automation | DNN | Raspberry Pi v1.3 camera with a fisheye lens |
| [ |
Raspberry Pi | Cervical cancer prevention | PiHRME | PiCamera |
| [ |
Raspberry Pi 3B+ | Speech recognition | EdgeRNN | Audio sensor |
| [ |
Raspberry Pi | CPU heat tracking | Adaptive learning | Infrared thermal sensor |
| [ |
Raspberry Pi 4 | High accuracy facial recognition | EfficientNet-Lite (CNN-KNN) | Webcam |
| [ |
Raspberry Pi 3B, NVIDIA Jetson TX1, NVIDIA Jetson TX2 | Psychological stress monitoring | KNN, SVM | Heart rate and accelerometer sensors |
| [ |
Raspberry Pi 3 model B | Image recognition for sea life | CNN-based animal recognition | Pi Camera v2.1 |
| [ |
Raspberry Pi 3 model B | Facial biometric scan | LGHP | Pi camera |
| [ |
Raspberry Pi 3 Model B+, Intel Neural Compute Stick 2 | Security surveillance | Mask R-CNN | Surveillance camera |
| [ |
Raspberry Pi 3B+, NVIDIA Jetson TX2 | Scalable and computationally cheap networks for embedded systems | DNN, MobileNetv2 | N/A |
| [ |
Raspberry Pi 4 | Weed identification for herbicide | Varied, includes CNN and KNN | The Raspberry Pi camera module version 2.0 with an 8-megapixel Sony IMX219 sensor |
| [ |
Raspberry Pi 3 Model B | Motor fault diagnosis | CNN | Hall effect sensor |
| [ |
Raspberry Pi 4 Model B | Machine state monitoring | CNN | Vibration Sensor, Accelerometers |
| [ |
Raspberry Pi 4 | Violent assault recognition | mobile CNN | Surveillance camera (no actual live testing) |
| [ |
Raspberry Pi | Asthma risk prediction | CNN, DNN | SDS011 air quality sensor |
| [ |
Raspberry Pi 3 Model B+ | Image classification | MobiHisNet (based on MobileNet) | N/A |
| [ |
Raspberry Pi 4 | Facial recognition and facial expression recognition | CNN | Logitech c270 camera |
| [ |
Raspberry Pi | Counting individuals within a given video feed | Hidden Makarov Model | Camera |
| [ |
Raspberry Pi 4 Model B | Dog health monitoring through posture analysis | Mask R-CNN | Smart camera network |
| [ |
Raspberry Pi 3 Model B | Pedestrian profile recognition | 2-layer CNN | FLIR Lepton thermal camera |
| [ |
NVIDIA Jetson TX2 | Lightweight facial recognition for embedded systems | Facial action unit | Camera |
| [ |
Raspberry Pi 3 Model B | Speech source identification | CNN | SSL sensors, microphones |
| [ |
Raspberry Pi | Fish recognition for underwater drones | LeNet, AlexNet, GoogLeNet | 360 degrees panoramic camera |
| [ |
NVIDIA Jetson Nano | AI traffic light control | SSD algorithm | Raspberry Pi camera |
| [ |
NVIDIA Jetson Nano | Battery charge management | Long Short-Term Memory (LSTM) | GY169 current converter sensor module |
| [ |
NVIDIA Jetson Nano | Automobile fog lamp intelligent control | CN-FWR5 | IMX219 camera |
| [ |
NVIDIA Jetson Nano, NVIDIA Jetson TX1, NVIDIA Jetson AGX Xavier | Benchmarking paper | PointNet | N/A |
| [ |
NVIDIA Jetson Nano | Identifying different plant species | AlexNet, ResNet50, and MobileNetv2, within Python’s Tensorflow framework | Photo camera |
| [ |
NVIDIA Jetson Nano | Car counter Traffic management | TeleBot API | Logitech c922 webcam |
| [ |
NVIDIA Jetson Nano | Diabetic ulcer detection | VGGNet, MatConvNet, and DenseNet | Thermal Camera |
| [ |
NVIDIA Jetson Nano | Smart city traffic management | MobileNetSSD and YOLOv4 | Camera |
| [ |
NVIDIA Jetson TX2 | Food quality analysis | Support Vector Machines (SVM), Naive Bayes, k-Nearest Neighbours algorithm (K-NN), Decision Tree, Random Forest, Logistic Regression, Neural Networks | Nuclear magnetic resonance spectrometer, infrared spectrometer |
| [ |
NVIDIA Jetson Xavier NX | Colonoscopy | Mobilenet | Colonoscopy camera |
| [ |
NVIDIA Jetson TX2 | Loose fruit detection for oil palm | Faster R-CNN | Camera |
| [ |
NVIDIA Jetson TX2, NVIDIA Jetson Nano | Hard hat detection on construction site | Histogram of Oriented Gradients | Surveillance camera |
| [ |
NVIDIA Jetson TX2 | Monitoring vehicle driver tiredness in real time | MobileNetV3 | Infrared Camera |
| [ |
NVIDIA Jetson Nano | Visual garbage detection | MobileNetV3Lite | N/A (most likely a video Camera) |
| [ |
NVIDIA Jetson Nano | Pot plant species identification and watering needs monitoring | MOBILENET SSD V2 | Capacitive Soil Moisture sensor, Water Level Sensor |
| [ |
NVIDIA Jetson Nano | Algorithm review for self-driving car navigation | SVM, ANN-MLP, CNN-LSTM | Mini camera IMX-219 |
| [ |
NVIDIA Jetson TX2 | Real-time security surveillance for acts of violence | Local Maximal Occurrence (LOMO), Crossview Quadratic Discriminant Analysis (XQDA) | RaspiCam camera, panoramic spherical camera |
| [ |
NVIDIA Jetson Nano, Raspberry Pi 3 Model B+ | Rescue operation robot computer vision | Haar Cascade, YOLO Tiny | No IR filter camera, LiDAR, Raspi Cam NOIR V2.1 |
| [ |
NVIDIA Jetson Nano | Security surveillance for abnormal activity detection | YOLOv5 | Logitech C270 Camera |
| [ |
NVIDIA Jetson Nano, NVIDIA Jetson TX2 | Facial ID for security | LFFD, ResNet50, SeNet50, LFFD+ ResNet50, LFFD+ SeNet50 | Camera |
| [ |
NVIDIA Jetson Nano | Radio frequency ID recognition | Baseline LSTM, baseline CNN, baseline CNMN, CNN with ResNet, CNMN with ResNet | Universal software radio peripheral |
| [ |
NVIDIA Jetson Xavier NX | Real-time image processing for fusion diagnostics | Max-Tree Representation | Thermal image camera |
| [ |
NVIDIA Jetson Nano | Security surveillance for unusual behavior | 2D CNN | HD camera |
| [ |
NVIDIA Jetson Xavier NX | Fire and smoke detection | YOLOv3 | Camera |
| [ |
NVIDIA Jetson Nano | Travel assistance for the visually impaired | MobileNet, SSD | Optical RGB camera |
| [ |
NVIDIA Jetson TX1 | Real-time pedestrian detection for autonomous vehicles | Modified YOLO v2 (Model H) | Zed Stereo camera |
| [ |
Nvidia Jetson Nano, Nvidia Jetson TX1, Raspberry Pi 4 | Artistic photography aesthetic score prediction | YOLO-CNN, Mobilenet, multi-threaded aesthetic predictor | N/A |
| [ |
NVIDIA Jetson TX2 | Real-time vehicle detection on embedded systems | EfficientDet-Lite, Yolov3-tiny | N/A |
| [ |
NVIDIA Jetson AGX Xavier | Uncertainty-based real-time object detection for autonomous vehicles | tiny YOLOv3, Gaussian YOLOv3 | Camera |
| [ |
NVIDIA Jetson Nano | Underwater object detection | YOLO v3, YOLO Nano Underwater | N/A (visual camera in case of field testing) |
References
1. Hoang, T.M.; Nam, S.H.; Park, K.R. Enhanced Detection and Recognition of Road Markings Based on Adaptive Region of Interest and Deep Learning. IEEE Access; 2019; 7, pp. 109817-109832. [DOI: https://dx.doi.org/10.1109/ACCESS.2019.2933598]
2. Inthanon, P.; Mungsing, S. Detection of Drowsiness from Facial Images in Real-Time Video Media using Nvidia Jetson Nano. Proceedings of the 2020 17th International Conference on Electrical Engineering/Electronics, Computer, Telecommunications and Information Technology (ECTI-CON); Phuket, Thailand, 24–27 June 2020; pp. 246-249. [DOI: https://dx.doi.org/10.1109/ECTI-CON49241.2020.9158235]
3. Xu, Z.; Li, J.; Zhang, M. A Surveillance Video Real-Time Analysis System Based on Edge-Cloud and FL-YOLO Cooperation in Coal Mine. IEEE Access; 2021; 9, pp. 68482-68497. [DOI: https://dx.doi.org/10.1109/ACCESS.2021.3077499]
4. Attaran, N.; Puranik, A.; Brooks, J.; Mohsenin, T. Embedded Low-Power Processor for Personalized Stress Detection. IEEE Trans. Circuits Syst. II Express Briefs; 2018; 65, pp. 2032-2036. [DOI: https://dx.doi.org/10.1109/TCSII.2018.2799821]
5. Ouyang, Z.; Niu, J.; Liu, Y.; Guizani, M. Deep CNN-Based Real-Time Traffic Light Detector for Self-Driving Vehicles. IEEE Trans. Mob. Comput.; 2020; 19, pp. 300-313. [DOI: https://dx.doi.org/10.1109/TMC.2019.2892451]
6. Dean, J. The Deep Learning Revolution and Its Implications for Computer Architecture and Chip Design. arXiv; 2019; [DOI: https://dx.doi.org/10.48550/ARXIV.1911.05289] arXiv: 1911.05289
7. Breland, D.S.; Dayal, A.; Jha, A.; Yalavarthy, P.K.; Pandey, O.J.; Cenkeramaddi, L.R. Robust Hand Gestures Recognition Using a Deep CNN and Thermal Images. IEEE Sens. J.; 2021; 21, pp. 26602-26614. [DOI: https://dx.doi.org/10.1109/JSEN.2021.3119977]
8. Hao, Y.; Küçük, A.; Ganguly, A.; Panahi, I.M.S. Spectral Flux-Based Convolutional Neural Network Architecture for Speech Source Localization and its Real-Time Implementation. IEEE Access; 2020; 8, pp. 197047-197058. [DOI: https://dx.doi.org/10.1109/ACCESS.2020.3033533] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/33981519]
9. Harisankar, V.; Karthika, R. Real Time Pedestrian Detection Using Modified YOLO V2. Proceedings of the 2020 5th International Conference on Communication and Electronics Systems (ICCES); Coimbatore, India, 10–12 June 2020; pp. 855-859. [DOI: https://dx.doi.org/10.1109/ICCES48766.2020.9138103]
10. Demir, H.S.; Christen, J.B.; Ozev, S. Energy-Efficient Image Recognition System for Marine Life. IEEE Trans. Comput. Aided Des. Integr. Circuits Syst.; 2020; 39, pp. 3458-3466. [DOI: https://dx.doi.org/10.1109/TCAD.2020.3012745]
11. Dewangan, D.K.; Sahu, S.P. Deep Learning-Based Speed Bump Detection Model for Intelligent Vehicle System Using Raspberry Pi. IEEE Sens. J.; 2021; 21, pp. 3570-3578. [DOI: https://dx.doi.org/10.1109/JSEN.2020.3027097]
12. Vieira, J.C.; Sartori, A.; Stefenon, S.F.; Perez, F.L.; de Jesus, G.S.; Leithardt, V.R.Q. Low-Cost CNN for Automatic Violence Recognition on Embedded System. IEEE Access; 2022; 10, pp. 25190-25202. [DOI: https://dx.doi.org/10.1109/ACCESS.2022.3155123]
13. Sahani, A.K.; Srivastava, D.; Sivaprakasam, M.; Joseph, J. A Machine Learning Pipeline for Measurement of Arterial Stiffness in A-Mode Ultrasound. IEEE Trans. Ultrason. Ferroelectr. Freq. Control; 2022; 69, pp. 106-113. [DOI: https://dx.doi.org/10.1109/TUFFC.2021.3109117] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/34460373]
14. Chen, C.; Liu, B.; Wan, S.; Qiao, P.; Pei, Q. An Edge Traffic Flow Detection Scheme Based on Deep Learning in an Intelligent Transportation System. IEEE Trans. Intell. Transp. Syst.; 2021; 22, pp. 1840-1852. [DOI: https://dx.doi.org/10.1109/TITS.2020.3025687]
15. Baller, S.P.; Jindal, A.; Chadha, M.; Gerndt, M. DeepEdgeBench: Benchmarking Deep Neural Networks on Edge Devices. Proceedings of the 2021 IEEE International Conference on Cloud Engineering (IC2E); Timisoara, Romania, 27–30 October 2021; pp. 20-30. [DOI: https://dx.doi.org/10.1109/IC2E52221.2021.00016]
16. Ajani, T.S.; Imoize, A.L.; Atayero, A.A. An Overview of Machine Learning within Embedded and Mobile Devices–Optimizations and Applications. Sensors; 2021; 21, 4412. [DOI: https://dx.doi.org/10.3390/s21134412] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/34203119]
17. Jeon, J.; Jung, S.; Lee, E.; Choi, D.; Myung, H. Run Your Visual-Inertial Odometry on NVIDIA Jetson: Benchmark Tests on a Micro Aerial Vehicle. IEEE Robot. Autom. Lett.; 2021; 6, pp. 5332-5339. [DOI: https://dx.doi.org/10.1109/LRA.2021.3075141]
18. Süzen, A.A.; Duman, B.; Şen, B. Benchmark Analysis of Jetson TX2, Jetson Nano and Raspberry PI using Deep-CNN. Proceedings of the 2020 International Congress on Human-Computer Interaction, Optimization and Robotic Applications (HORA); Ankara, Turkey, 26–28 June 2020; pp. 1-5. [DOI: https://dx.doi.org/10.1109/HORA49412.2020.9152915]
19. Bianco, S.; Cadene, R.; Celona, L.; Napoletano, P. Benchmark Analysis of Representative Deep Neural Network Architectures. IEEE Access; 2018; 6, pp. 64270-64277. [DOI: https://dx.doi.org/10.1109/ACCESS.2018.2877890]
20. Choe, M.; Lee, S.; Sung, N.M.; Jung, S.; Choe, C. Benchmark Analysis of Deep Learning-based 3D Object Detectors on NVIDIA Jetson Platforms. Proceedings of the 2021 International Conference on Information and Communication Technology Convergence (ICTC); Jeju Island, Republic of Korea, 20–22 October 2021; pp. 10-12. [DOI: https://dx.doi.org/10.1109/ICTC52510.2021.9621072]
21. Ullah, S.; Kim, D.H. Benchmarking Jetson Platform for 3D Point-Cloud and Hyper-Spectral Image Classification. Proceedings of the 2020 IEEE International Conference on Big Data and Smart Computing (BigComp); Busan, Republic of Korea, 19–22 February 2020; pp. 477-482. [DOI: https://dx.doi.org/10.1109/BigComp48618.2020.00-21]
22. Claudet, T.; Tomita, K.; Ho, K. Benchmark Analysis of Semantic Segmentation Algorithms for Safe Planetary Landing Site Selection. IEEE Access; 2022; 10, pp. 41766-41775. [DOI: https://dx.doi.org/10.1109/ACCESS.2022.3167763]
23. Lungu, I.A.; Aimar, A.; Hu, Y.; Delbruck, T.; Liu, S.C. Siamese Networks for Few-Shot Learning on Edge Embedded Devices. IEEE J. Emerg. Sel. Top. Circuits Syst.; 2020; 10, pp. 488-497. [DOI: https://dx.doi.org/10.1109/JETCAS.2020.3033155]
24. Nvidia Corporation. Jetson Nano Developer Kit; Nvidia Corporation: Santa Clara, CA, USA, 2019.
25. Nvidia Corporation. Jetson TX1 Developer Kit; Nvidia Corporation: Santa Clara, CA, USA, 2016.
26. Nvidia Corporation. Jetson TX2 Developer Kit; Nvidia Corporation: Santa Clara, CA, USA, 2019.
27. Nvidia Corporation. Jetson AGX Xavier Developer Kit; Nvidia Corporation: Santa Clara, CA, USA, 2019.
28. Nvidia Corporation. Jetson Xavier NX Developer Kit; Nvidia Corporation: Santa Clara, CA, USA, 2020.
29. Coral.ai. Get Started with the Dev Board. Available online: https://coral.ai/docs/dev-board/get-started (accessed on 29 May 2022).
30. Raspberry Pi Foundation. Raspberry Pi 3 Model B; Raspberry Pi Foundation: Cambridge, UK, 2016.
31. Raspberry Pi Foundation. Raspberry Pi 4 Model B; Raspberry Pi Foundation: Cambridge, UK, 2019.
32. Hardkernel Co. ODROID XU4; Hardkernel Co.: Anyang, Gyeonggi-do, Republic of Korea, 2015.
33. SinoVoip Co., Ltd. Banana PI M2; SinoVoip Co., Ltd.: Shenzhen, China.
34. ASUSTek Computer Inc. Tinker Board S; ASUSTek Computer Inc.: Taipei, Taiwan, 2017.
35. Bigelow, S.J. TechTarget, Operating System (OS). Available online: https://www.techtarget.com/whatis/definition/operating-system-OS (accessed on 11 July 2022).
36. Gillis, A.S. TechTarget, Device Driver. Available online: https://www.techtarget.com/searchenterprisedesktop/definition/device-driver (accessed on 4 July 2022).
37. Chakraborty, K. Firmware. Techopedia. Available online: https://www.techopedia.com/definition/2137/firmware (accessed on 27 June 2022).
38. ASUSTek Computer Inc. Tinker Edge R; ASUSTek Computer Inc.: Taipei, Taiwan, 2020.
39. Hu, Q.; Tang, X.; Tang, W. A Real-Time Patient-Specific Sleeping Posture Recognition System Using Pressure Sensitive Conductive Sheet and Transfer Learning. IEEE Sens. J.; 2021; 21, pp. 6869-6879. [DOI: https://dx.doi.org/10.1109/JSEN.2020.3043416]
40. Hu, Q.; Tang, X.; Tang, W. A Smart Chair Sitting Posture Recognition System Using Flex Sensors and FPGA Implemented Artificial Neural Network. IEEE Sens. J.; 2020; 20, pp. 8007-8016. [DOI: https://dx.doi.org/10.1109/JSEN.2020.2980207]
41. Science Learning Hub, Electricity and Sensors. Available online: https://www.sciencelearn.org.nz/resources/1602-electricity-and-sensors (accessed on 12 July 2022).
42. Wilson, J.S. Sensor Technology Handbook; Newnes: Oxford, UK, 2004.
43. Hu, Q.; Yi, C.; Kliewer, J.; Tang, W. Asynchronous communication for wireless sensors using ultra wideband impulse radio. Proceedings of the 2015 IEEE 58th International Midwest Symposium on Circuits and Systems (MWSCAS); Fort Collins, CO, USA, 2–5 August 2015; pp. 1-4. [DOI: https://dx.doi.org/10.1109/MWSCAS.2015.7282170]
44. Hu, Q.; Tang, X.; Tang, W. Integrated Asynchronous Ultra-Wideband Impulse Radio with Intrinsic Clock and Data Recovery. IEEE Microw. Wirel. Components Lett.; 2017; 27, pp. 416-418. [DOI: https://dx.doi.org/10.1109/LMWC.2017.2678438]
45. McGrath, M.J.; Ní Scanaill, C. Key Sensor Technology Components: Hardware and Software Overview; Apress: Berkeley, CA, USA, 2014; pp. 51-77.
46. Gleason, C.J.; Im, J. Forest biomass estimation from airborne LiDAR data using machine learning approaches. Remote. Sens. Environ.; 2012; 125, pp. 80-91. [DOI: https://dx.doi.org/10.1016/j.rse.2012.07.006]
47. Infiniti Electro-Optics, Visible Imaging Sensor (RGB Color Camera). Available online: https://www.infinitioptics.com/glossary/visible-imaging-sensor-400700nm-colour-cameras (accessed on 11 July 2022).
48. Tang, W.; Biglari, A.; Ebarb, R.; Pickett, T.; Smallidge, S.; Ward, M. A Smart Sensing System of Water Quality and Intake Monitoring for Livestock and Wild Animals. Sensors; 2021; 21, 2885. [DOI: https://dx.doi.org/10.3390/s21082885] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/33924135]
49. Biglari, A.; Tang, W. A Vision-Based Cattle Recognition System Using TensorFlow for Livestock Water Intake Monitoring. IEEE Sens. Lett.; 2022; 6, pp. 1-4. [DOI: https://dx.doi.org/10.1109/LSENS.2022.3215699]
50. Ibarra, V.; Araya-Salas, M.; Tang, Y.; Park, C.; Hyde, A.; Wright, T.F.; Tang, W. An RFID Based Smart Feeder for Hummingbirds. Sensors; 2015; 15, 29886. [DOI: https://dx.doi.org/10.3390/s151229886]
51. Fluke. How Infrared Cameras Work. Available online: https://www.fluke.com/en-us/learn/blog/thermal-imaging/how-infrared-cameras-work (accessed on 14 July 2022).
52. Langmann, B.; Hartmann, K.; Loffeld, O. Depth Camera Technology Comparison and Performance Evaluation. Proceedings of the International Conference on Pattern Recognition Applications and Methods; Algarve, Portugal, 6–8 February 2012.
53. Adams, J. Digital Camera World, What Is a 360 Camera and How Do You Use Them?. Available online: https://www.digitalcameraworld.com/features/what-is-a-360-camera-and-how-do-you-use-them (accessed on 15 July 2022).
54. Adams, J. 360 Cameras, How Do 360 Cameras Work?. Available online: https://www.threesixtycameras.com/how-do-360-cameras-work-explained/ (accessed on 15 July 2022).
55. Australian Government Bureau of Meteorology. How Radar Works. Available online: http://www.bom.gov.au/australia/radar/about/what_is_radar.shtml (accessed on 13 July 2022).
56. Collis, R.T.H. Lidar. Appl. Opt.; 1970; 9, pp. 1782-1788. [DOI: https://dx.doi.org/10.1364/AO.9.001782] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/20094138]
57. Raj, T.; Hashim, F.H.; Huddin, A.B.; Ibrahim, M.F.; Hussain, A. A Survey on LiDAR Scanning Mechanisms. Electronics; 2020; 9, 741. [DOI: https://dx.doi.org/10.3390/electronics9050741]
58. How Do Microphones Work?. Available online: https://mynewmicrophone.com/how-do-microphones-work-a-helpful-illustrated-guide/ (accessed on 17 July 2022).
59. Lee, K.; Tang, W. A Fully Wireless Wearable Motion Tracking System with 3D Human Model for Gait Analysis. Sensors; 2021; 21, 4051. [DOI: https://dx.doi.org/10.3390/s21124051]
60. AzoSensors, Using Sensors to Capture Body Movement. Available online: https://www.azosensors.com/article.aspx?ArticleID=429 (accessed on 13 July 2022).
61. Tang, X.; Hu, Q.; Tang, W. A Real-Time QRS Detection System With PR/RT Interval and ST Segment Measurements for Wearable ECG Sensors Using Parallel Delta Modulators. IEEE Trans. Biomed. Circuits Syst.; 2018; 12, pp. 751-761. [DOI: https://dx.doi.org/10.1109/TBCAS.2018.2823275]
62. Tang, X.; Ma, Z.; Hu, Q.; Tang, W. A Real-Time Arrhythmia Heartbeats Classification Algorithm Using Parallel Delta Modulations and Rotated Linear-Kernel Support Vector Machines. IEEE Trans. Biomed. Eng.; 2020; 67, pp. 978-986. [DOI: https://dx.doi.org/10.1109/TBME.2019.2926104]
63. Tang, X.; Tang, W. A 151nW Second-Order Ternary Delta Modulator for ECG Slope Variation Measurement with Baseline Wandering Resilience. Proceedings of the 2020 IEEE Custom Integrated Circuits Conference (CICC); Boston, MA, USA, 22–25 March 2020; pp. 1-4.
64. Farnsworth, B. What Is ECG and How Does It Work? imotions. Available online: https://imotions.com/blog/learning/research-fundamentals/what-is-ecg/ (accessed on 28 July 2022).
65. Tang, X.; Tang, W. An ECG Delineation and Arrhythmia Classification System Using Slope Variation Measurement by Ternary Second-Order Delta Modulators for Wearable ECG Sensors. IEEE Trans. Biomed. Circuits Syst.; 2021; 15, pp. 1053-1065. [DOI: https://dx.doi.org/10.1109/TBCAS.2021.3113665]
66. Tang, X.; Liu, S.; Reviriego, P.; Lombardi, F.; Tang, W. A Near-Sensor ECG Delineation and Arrhythmia Classification System. IEEE Sens. J.; 2022; 22, pp. 14217-14227. [DOI: https://dx.doi.org/10.1109/JSEN.2022.3183136]
67. Mayo Clinic, EEG (electroencephalogram). Available online: https://www.mayoclinic.org/tests-procedures/eeg/about/pac-20393875 (accessed on 22 July 2022).
68. Tang, X.; Liu, S.; Che, W.; Tang, W. Tampering Attack Detection in Analog to Feature Converter for Wearable Biosensor. Proceedings of the 2022 IEEE International Symposium on Circuits and Systems (ISCAS); Austin, TX, USA, 27 May–1 June 2022; pp. 1150-1154. [DOI: https://dx.doi.org/10.1109/ISCAS48785.2022.9937707]
69. Marquez Chavez, J.; Tang, W. A Vision-Based System for Stage Classification of Parkinsonian Gait Using Machine Learning and Synthetic Data. Sensors; 2022; 22, 4463. [DOI: https://dx.doi.org/10.3390/s22124463]
70. Gresham, B.; Torres, J.; Britton, J.; Ma, Z.; Parada, A.B.; Gutierrez, M.L.; Lawrence, M.; Tang, W. High-dimensional Time-series Gait Analysis using a Full-body Wireless Wearable Motion Sensing System and Convolutional Neural Network. Proceedings of the 2022 IEEE Biomedical Circuits and Systems Conference (BioCAS); Taipei, Taiwan, 13–15 October 2022; pp. 389-393. [DOI: https://dx.doi.org/10.1109/BioCAS54905.2022.9948596]
71. Alkobi, J. Percepto, The Evolution of Drones: From Military to Hobby & Commercial. Available online: https://percepto.co/the-evolution-of-drones-from-military-to-hobby-commercial/ (accessed on 29 July 2022).
72. Stuckey, H.; Al-Radaideh, A.; Escamilla, L.; Sun, L.; Carrillo, L.G.; Tang, W. An Optical Spatial Localization System for Tracking Unmanned Aerial Vehicles Using a Single Dynamic Vision Sensor. Proceedings of the 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS); Prague, Czech Republic, 27 September–1 October 2021; pp. 3093-3100. [DOI: https://dx.doi.org/10.1109/IROS51168.2021.9636665]
73. Stuckey, H.; Al-Radaideh, A.; Sun, L.; Tang, W. A Spatial Localization and Attitude Estimation System for Unmanned Aerial Vehicles Using a Single Dynamic Vision Sensor. IEEE Sens. J.; 2022; 22, pp. 15497-15507. [DOI: https://dx.doi.org/10.1109/JSEN.2022.3187423]
74. Varghese, R.; Sharma, S. Affordable Smart Farming Using IoT and Machine Learning. Proceedings of the 2018 Second International Conference on Intelligent Computing and Control Systems (ICICCS); Madurai, India, 14–15 June 2018; pp. 645-650. [DOI: https://dx.doi.org/10.1109/ICCONS.2018.8663044]
75. Dunn, J.; Tron, R. Temporal Siamese Networks for Clutter Mitigation Applied to Vision-Based Quadcopter Formation Control. IEEE Robot. Autom. Lett.; 2021; 6, pp. 32-39. [DOI: https://dx.doi.org/10.1109/LRA.2020.3028056]
76. Yang, M.D.; Tseng, H.H.; Hsu, Y.C.; Tseng, W.C. Real-time Crop Classification Using Edge Computing and Deep Learning. Proceedings of the 2020 IEEE 17th Annual Consumer Communications & Networking Conference (CCNC); Las Vegas, NV, USA, 10–13 January 2020; pp. 1-4. [DOI: https://dx.doi.org/10.1109/CCNC46108.2020.9045498]
77. Aguiar, A.S.; Santos, F.N.D.; De Sousa, A.J.M.; Oliveira, P.M.; Santos, L.C. Visual Trunk Detection Using Transfer Learning and a Deep Learning-Based Coprocessor. IEEE Access; 2020; 8, pp. 77308-77320. [DOI: https://dx.doi.org/10.1109/ACCESS.2020.2989052]
78. Adami, D.; Ojo, M.O.; Giordano, S. Design, Development and Evaluation of an Intelligent Animal Repelling System for Crop Protection Based on Embedded Edge-AI. IEEE Access; 2021; 9, pp. 132125-132139. [DOI: https://dx.doi.org/10.1109/ACCESS.2021.3114503]
79. Beegam, K.S.; Shenoy, M.V.; Chaturvedi, N. Hybrid Consensus and Recovery Block-Based Detection of Ripe Coffee Cherry Bunches Using RGB-D Sensor. IEEE Sens. J.; 2022; 22, pp. 732-740. [DOI: https://dx.doi.org/10.1109/JSEN.2021.3130747]
80. Li, N.; Zhang, X.; Zhang, C.; Guo, H.; Sun, Z.; Wu, X. Real-Time Crop Recognition in Transplanted Fields With Prominent Weed Growth: A Visual-Attention-Based Approach. IEEE Access; 2019; 7, pp. 185310-185321. [DOI: https://dx.doi.org/10.1109/ACCESS.2019.2942158]
81. Sa, I.; Chen, Z.; Popović, M.; Khanna, R.; Liebisch, F.; Nieto, J.; Siegwart, R. weedNet: Dense Semantic Weed Classification Using Multispectral Images and MAV for Smart Farming. IEEE Robot. Autom. Lett.; 2018; 3, pp. 588-595. [DOI: https://dx.doi.org/10.1109/LRA.2017.2774979]
82. Tufail, M.; Iqbal, J.; Tiwana, M.I.; Alam, M.S.; Khan, Z.A.; Khan, M.T. Identification of Tobacco Crop Based on Machine Learning for a Precision Agricultural Sprayer. IEEE Access; 2021; 9, pp. 23814-23825. [DOI: https://dx.doi.org/10.1109/ACCESS.2021.3056577]
83. Xiang, A.J.; Huddin, A.B.; Ibrahim, M.F.; Hashim, F.H. An Oil Palm Loose Fruits Image Detection System using Faster R -CNN and Jetson TX2. Proceedings of the 2021 International Conference on Electrical Engineering and Informatics (ICEEI); Kuala Terengganu, Malaysia, 12–13 October 2021; pp. 1-6. [DOI: https://dx.doi.org/10.1109/ICEEI52609.2021.9611111]
84. Chen, C.J.; Huang, Y.Y.; Li, Y.S.; Chen, Y.C.; Chang, C.Y.; Huang, Y.M. Identification of Fruit Tree Pests With Deep Learning on Embedded Drone to Achieve Accurate Pesticide Spraying. IEEE Access; 2021; 9, pp. 21986-21997. [DOI: https://dx.doi.org/10.1109/ACCESS.2021.3056082]
85. Jarraya, I.; Ouarda, W.; Alimi, A.M. A Preliminary Investigation on Horses Recognition Using Facial Texture Features. Proceedings of the 2015 IEEE International Conference on Systems, Man, and Cybernetics; Hong Kong, China, 9–12 October 2015; pp. 2803-2808. [DOI: https://dx.doi.org/10.1109/SMC.2015.489]
86. Basu, A.; Dasgupta, A.; Thyagharajan, A.; Routray, A.; Guha, R.; Mitra, P. A Portable Personality Recognizer Based on Affective State Classification Using Spectral Fusion of Features. IEEE Trans. Affect. Comput.; 2018; 9, pp. 330-342. [DOI: https://dx.doi.org/10.1109/TAFFC.2018.2828845]
87. Chakraborty, S.; Singh, S.K.; Kumar, K. Facial Biometric System for Recognition Using Extended LGHP Algorithm on Raspberry Pi. IEEE Sens. J.; 2020; 20, pp. 8117-8127. [DOI: https://dx.doi.org/10.1109/JSEN.2020.2979907]
88. Papa, L.; Alati, E.; Russo, P.; Amerini, I. SPEED: Separable Pyramidal Pooling EncodEr-Decoder for Real-Time Monocular Depth Estimation on Low-Resource Settings. IEEE Access; 2022; 10, pp. 44881-44890. [DOI: https://dx.doi.org/10.1109/ACCESS.2022.3170425]
89. Koubaa, A.; Ammar, A.; Kanhouch, A.; AlHabashi, Y. Cloud Versus Edge Deployment Strategies of Real-Time Face Recognition Inference. IEEE Trans. Netw. Sci. Eng.; 2022; 9, pp. 143-160. [DOI: https://dx.doi.org/10.1109/TNSE.2021.3055835]
90. Nguyen, D.L.; Putro, M.D.; Jo, K.H. Facemask Wearing Alert System Based on Simple Architecture with Low-Computing Devices. IEEE Access; 2022; 10, pp. 29972-29981. [DOI: https://dx.doi.org/10.1109/ACCESS.2022.3158304]
91. Ab Wahab, M.N.; Nazir, A.; Zhen Ren, A.T.; Mohd Noor, M.H.; Akbar, M.F.; Mohamed, A.S.A. Efficientnet-Lite and Hybrid CNN-KNN Implementation for Facial Expression Recognition on Raspberry Pi. IEEE Access; 2021; 9, pp. 134065-134080. [DOI: https://dx.doi.org/10.1109/ACCESS.2021.3113337]
92. Zarif, N.E.; Montazeri, L.; Leduc-Primeau, F.; Sawan, M. Mobile-Optimized Facial Expression Recognition Techniques. IEEE Access; 2021; 9, pp. 101172-101185. [DOI: https://dx.doi.org/10.1109/ACCESS.2021.3095844]
93. Gaikwad, B.; Prakash, P.; Karmakar, A. Edge-based real-time face logging system for security applications. Proceedings of the 2021 12th International Conference on Computing Communication and Networking Technologies (ICCCNT); Kharagpur, India, 6–8 July 2021; pp. 1-6. [DOI: https://dx.doi.org/10.1109/ICCCNT51525.2021.9579758]
94. Yang, J.; Qian, T.; Zhang, F.; Khan, S.U. Real-Time Facial Expression Recognition Based on Edge Computing. IEEE Access; 2021; 9, pp. 76178-76190. [DOI: https://dx.doi.org/10.1109/ACCESS.2021.3082641]
95. Bucki, N.; Lee, J.; Mueller, M.W. Rectangular Pyramid Partitioning Using Integrated Depth Sensors (RAPPIDS): A Fast Planner for Multicopter Navigation. IEEE Robot. Autom. Lett.; 2020; 5, pp. 4626-4633. [DOI: https://dx.doi.org/10.1109/LRA.2020.3003277]
96. Dao, T.T.; Pham, Q.V.; Hwang, W.J. FastMDE: A Fast CNN Architecture for Monocular Depth Estimation at High Resolution. IEEE Access; 2022; 10, pp. 16111-16122. [DOI: https://dx.doi.org/10.1109/ACCESS.2022.3145969]
97. Tsai, T.H.; Hsu, C.W. Implementation of Fall Detection System Based on 3D Skeleton for Deep Learning Technique. IEEE Access; 2019; 7, pp. 153049-153059. [DOI: https://dx.doi.org/10.1109/ACCESS.2019.2947518]
98. Nowosielski, A.; Małecki, K.; Forczmański, P.; Smoliński, A.; Krzywicki, K. Embedded Night-Vision System for Pedestrian Detection. IEEE Sens. J.; 2020; 20, pp. 9293-9304. [DOI: https://dx.doi.org/10.1109/JSEN.2020.2986855]
99. Han, Y.; Oruklu, E. Traffic sign recognition based on the NVIDIA Jetson TX1 embedded system using convolutional neural networks. Proceedings of the 2017 IEEE 60th International Midwest Symposium on Circuits and Systems (MWSCAS); Boston, MA, USA, 6–9 August 2017; pp. 184-187. [DOI: https://dx.doi.org/10.1109/MWSCAS.2017.8052891]
100. Liu, Y.; Cao, S.; Lasang, P.; Shen, S. Modular Lightweight Network for Road Object Detection Using a Feature Fusion Approach. IEEE Trans. Syst. Man, Cybern. Syst.; 2021; 51, pp. 4716-4728. [DOI: https://dx.doi.org/10.1109/TSMC.2019.2945053]
101. Lai, C.Y.; Wu, B.X.; Shivanna, V.M.; Guo, J.I. MTSAN: Multi-Task Semantic Attention Network for ADAS Applications. IEEE Access; 2021; 9, pp. 50700-50714. [DOI: https://dx.doi.org/10.1109/ACCESS.2021.3068991]
102. Li, Z.; Zhou, A.; Pu, J.; Yu, J. Multi-Modal Neural Feature Fusion for Automatic Driving Through Perception-Aware Path Planning. IEEE Access; 2021; 9, pp. 142782-142794. [DOI: https://dx.doi.org/10.1109/ACCESS.2021.3120720]
103. Farooq, M.A.; Corcoran, P.; Rotariu, C.; Shariff, W. Object Detection in Thermal Spectrum for Advanced Driver-Assistance Systems (ADAS). IEEE Access; 2021; 9, pp. 156465-156481. [DOI: https://dx.doi.org/10.1109/ACCESS.2021.3129150]
104. Sun, T.; Pan, W.; Wang, Y.; Liu, Y. Region of Interest Constrained Negative Obstacle Detection and Tracking With a Stereo Camera. IEEE Sens. J.; 2022; 22, pp. 3616-3625. [DOI: https://dx.doi.org/10.1109/JSEN.2022.3142024]
105. Tang, X.; Chen, J.; Yang, K.; Toyoda, M.; Liu, T.; Hu, X. Visual Detection and Deep Reinforcement Learning-Based Car Following and Energy Management for Hybrid Electric Vehicles. IEEE Trans. Transp. Electrif.; 2022; 8, pp. 2501-2515. [DOI: https://dx.doi.org/10.1109/TTE.2022.3141780]
106. Sajjad, M.; Irfan, M.; Muhammad, K.; Ser, J.D.; Sanchez-Medina, J.; Andreev, S.; Ding, W.; Lee, J.W. An Efficient and Scalable Simulation Model for Autonomous Vehicles With Economical Hardware. IEEE Trans. Intell. Transp. Syst.; 2021; 22, pp. 1718-1732. [DOI: https://dx.doi.org/10.1109/TITS.2020.2980855]
107. Vijitkunsawat, W.; Chantngarm, P. Comparison of Machine Learning Algorithm’s on Self-Driving Car Navigation using Nvidia Jetson Nano. Proceedings of the 2020 17th International Conference on Electrical Engineering/Electronics, Computer, Telecommunications and Information Technology (ECTI-CON); Phuket, Thailand, 24–27 June 2020; pp. 201-204. [DOI: https://dx.doi.org/10.1109/ECTI-CON49241.2020.9158311]
108. Nguyen, H.H.; Tran, D.N.N.; Jeon, J.W. Towards Real-Time Vehicle Detection on Edge Devices with Nvidia Jetson TX2. Proceedings of the 2020 IEEE International Conference on Consumer Electronics-Asia (ICCE-Asia); Seoul, Republic of Korea, 1–3 November 2020; pp. 1-4. [DOI: https://dx.doi.org/10.1109/ICCE-Asia49877.2020.9277463]
109. Choi, J.; Chun, D.; Lee, H.J.; Kim, H. Uncertainty-Based Object Detector for Autonomous Driving Embedded Platforms. Proceedings of the 2020 2nd IEEE International Conference on Artificial Intelligence Circuits and Systems (AICAS); Phuket, Thailand, 24–27 June 2020; pp. 16-20. [DOI: https://dx.doi.org/10.1109/AICAS48895.2020.9073907]
110. Díaz, S.; Krohmer, T.; Moreira, Á.; Godoy, S.E.; Figueroa, M. An Instrument for Accurate and Non-Invasive Screening of Skin Cancer Based on Multimodal Imaging. IEEE Access; 2019; 7, pp. 176646-176657. [DOI: https://dx.doi.org/10.1109/ACCESS.2019.2956898]
111. Prabhu, M.S.; Verma, S. A Deep Learning framework and its Implementation for Diabetic Foot Ulcer Classification. Proceedings of the 2021 9th International Conference on Reliability, Infocom Technologies and Optimization (Trends and Future Directions) (ICRITO); Noida, India, 3–4 September 2021; pp. 1-5. [DOI: https://dx.doi.org/10.1109/ICRITO51393.2021.9596380]
112. Chang, C.Y.; Liou, S.H. A Blind Aid System based on Jetson TX2 Embedded System and Deep Learning Technique. Proceedings of the 2019 8th International Conference on Innovation, Communication and Engineering (ICICE); Zhengzhou, China, 25–30 October 2019; pp. 25-29. [DOI: https://dx.doi.org/10.1109/ICICE49024.2019.9117526]
113. Dong, J.; Ota, K.; Dong, M. UAV-Based Real-Time Survivor Detection System in Post-Disaster Search and Rescue Operations. IEEE J. Miniaturization Air Space Syst.; 2021; 2, pp. 209-219. [DOI: https://dx.doi.org/10.1109/JMASS.2021.3083659]
114. Tang, C.; Xia, S.; Qian, M.; Wang, B. Deep Learning-Based Vein Localization on Embedded System. IEEE Access; 2021; 9, pp. 27916-27927. [DOI: https://dx.doi.org/10.1109/ACCESS.2021.3058014]
115. Paluru, N.; Dayal, A.; Jenssen, H.B.; Sakinis, T.; Cenkeramaddi, L.R.; Prakash, J.; Yalavarthy, P.K. Anam-Net: Anamorphic Depth Embedding-Based Lightweight CNN for Segmentation of Anomalies in COVID-19 Chest CT Images. IEEE Trans. Neural Networks Learn. Syst.; 2021; 32, pp. 932-946. [DOI: https://dx.doi.org/10.1109/TNNLS.2021.3054746] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/33544680]
116. Nguyen Huu, P.; Nguyen Thi, N.; Ngoc, T.P. Proposing Posture Recognition System Combining MobilenetV2 and LSTM for Medical Surveillance. IEEE Access; 2022; 10, pp. 1839-1849. [DOI: https://dx.doi.org/10.1109/ACCESS.2021.3138778]
117. Goyal, M.; Reeves, N.D.; Rajbhandari, S.; Yap, M.H. Robust Methods for Real-Time Diabetic Foot Ulcer Detection and Localization on Mobile Devices. IEEE J. Biomed. Health Inform.; 2019; 23, pp. 1730-1741. [DOI: https://dx.doi.org/10.1109/JBHI.2018.2868656] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/30188841]
118. Khan, M.A.; Paul, P.; Rashid, M.; Hossain, M.; Ahad, M.A.R. An AI-Based Visual Aid With Integrated Reading Assistant for the Completely Blind. IEEE Trans. Hum.-Mach. Syst.; 2020; 50, pp. 507-517. [DOI: https://dx.doi.org/10.1109/THMS.2020.3027534]
119. Parra, S.; Carranza, E.; Coole, J.; Hunt, B.; Smith, C.; Keahey, P.; Maza, M.; Schmeler, K.; Richards-Kortum, R. Development of Low-Cost Point-of-Care Technologies for Cervical Cancer Prevention Based on a Single-Board Computer. IEEE J. Transl. Eng. Health Med.; 2020; 8, pp. 1-10. [DOI: https://dx.doi.org/10.1109/JTEHM.2020.2970694] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/32190430]
120. Tsai, M.F.; Huang, J.Y. Predicting Canine Posture with Smart Camera Networks Powered by the Artificial Intelligence of Things. IEEE Access; 2020; 8, pp. 220848-220857. [DOI: https://dx.doi.org/10.1109/ACCESS.2020.3042539]
121. Ciobanu, A.; Luca, M.; Barbu, T.; Drug, V.; Olteanu, A.; Vulpoi, R. Experimental Deep Learning Object Detection in Real-time Colonoscopies. Proceedings of the 2021 International Conference on e-Health and Bioengineering (EHB); Iasi, Romania, 18–19 November 2021; pp. 1-4. [DOI: https://dx.doi.org/10.1109/EHB52898.2021.9657740]
122. Joshi, R.; Tripathi, M.; Kumar, A.; Gaur, M.S. Object Recognition and Classification System for Visually Impaired. Proceedings of the 2020 International Conference on Communication and Signal Processing (ICCSP); Chennai, India, 28–30 July 2020; pp. 1568-1572. [DOI: https://dx.doi.org/10.1109/ICCSP48568.2020.9182077]
123. Wang, X.; Zhang, L.; Huang, W.; Wang, S.; Wu, H.; He, J.; Song, A. Deep Convolutional Networks with Tunable Speed–Accuracy Tradeoff for Human Activity Recognition Using Wearables. IEEE Trans. Instrum. Meas.; 2022; 71, pp. 1-12. [DOI: https://dx.doi.org/10.1109/TIM.2021.3132088]
124. Breland, D.S.; Skriubakken, S.B.; Dayal, A.; Jha, A.; Yalavarthy, P.K.; Cenkeramaddi, L.R. Deep Learning-Based Sign Language Digits Recognition From Thermal Images With Edge Computing System. IEEE Sens. J.; 2021; 21, pp. 10445-10453. [DOI: https://dx.doi.org/10.1109/JSEN.2021.3061608]
125. Liu, M.; Li, Z.; Li, Y.; Liu, Y. A Fast and Accurate Method of Power Line Intelligent Inspection Based on Edge Computing. IEEE Trans. Instrum. Meas.; 2022; 71, pp. 1-12. [DOI: https://dx.doi.org/10.1109/TIM.2022.3152855]
126. Saeed, K.; Adamski, M.; Klimowicz, A.; Lupinska-Dubicka, A.; Omieljanowicz, M.; Rubin, G.; Rybnik, M.; Szymkowski, M.; Tabedzki, M.; Zienkiewicz, L. A Novel Extension for e-Safety Initiative Based on Developed Fusion of Biometric Traits. IEEE Access; 2020; 8, pp. 149887-149898. [DOI: https://dx.doi.org/10.1109/ACCESS.2020.3015503]
127. Kamal, R.; Chemmanam, A.J.; Jose, B.A.; Mathews, S.; Varghese, E. Construction Safety Surveillance Using Machine Learning. Proceedings of the 2020 International Symposium on Networks, Computers and Communications (ISNCC); Montreal, QC, Canada, 20–22 October 2020; pp. 1-6. [DOI: https://dx.doi.org/10.1109/ISNCC49221.2020.9297198]
128. Vu, H.N.; Pham, C.; Dung, N.M.; Ro, S. Detecting and Tracking Sinkholes Using Multi-Level Convolutional Neural Networks and Data Association. IEEE Access; 2020; 8, pp. 132625-132641. [DOI: https://dx.doi.org/10.1109/ACCESS.2020.3010885]
129. Kumar, P.; Batchu, S.; Swamy S., N.; Kota, S.R. Real-Time Concrete Damage Detection Using Deep Learning for High Rise Structures. IEEE Access; 2021; 9, pp. 112312-112331. [DOI: https://dx.doi.org/10.1109/ACCESS.2021.3102647]
130. Tu, Z.; Wu, S.; Kang, G.; Lin, J. Real-Time Defect Detection of Track Components: Considering Class Imbalance and Subtle Difference Between Classes. IEEE Trans. Instrum. Meas.; 2021; 70, pp. 1-12. [DOI: https://dx.doi.org/10.1109/TIM.2021.3117357]
131. Bhattacharya, S.; Ranjan, A.; Reza, M. A Portable Biometrics System Based on Forehead Subcutaneous Vein Pattern and Periocular Biometric Pattern. IEEE Sens. J.; 2022; 22, pp. 7022-7033. [DOI: https://dx.doi.org/10.1109/JSEN.2022.3149286]
132. Altowaijri, A.H.; Alfaifi, M.S.; Alshawi, T.A.; Ibrahim, A.B.; Alshebeili, S.A. A Privacy-Preserving Iot-Based Fire Detector. IEEE Access; 2021; 9, pp. 51393-51402. [DOI: https://dx.doi.org/10.1109/ACCESS.2021.3069588]
133. Ahmed, A.A.; Echi, M. Hawk-Eye: An AI-Powered Threat Detector for Intelligent Surveillance Cameras. IEEE Access; 2021; 9, pp. 63283-63293. [DOI: https://dx.doi.org/10.1109/ACCESS.2021.3074319]
134. Huu, N.N.T.; Mai, L.; Minh, T.V. Detecting Abnormal and Dangerous Activities Using Artificial Intelligence on The Edge for Smart City Application. Proceedings of the 2021 15th International Conference on Advanced Computing and Applications (ACOMP); Ho Chi Minh City, Vietnam, 24–26 November 2021; pp. 85-92. [DOI: https://dx.doi.org/10.1109/ACOMP53746.2021.00018]
135. Adam, M.; Ramachandran, P.; Alex, Z.C. Human Irregularity Detection Based on Posture and Behavioral Analysis. Proceedings of the 2021 Innovations in Power and Advanced Computing Technologies (i-PACT); Chennai, India, 28–30 July 2021; pp. 1-6. [DOI: https://dx.doi.org/10.1109/i-PACT52855.2021.9697052]
136. Chen, Y.C.; Fathoni, H.; Yang, C.T. Implementation of Fire and Smoke Detection using DeepStream and Edge Computing Approachs. Proceedings of the 2020 International Conference on Pervasive Artificial Intelligence (ICPAI); Taipei, Taiwan, 3–5 December 2020; pp. 272-275. [DOI: https://dx.doi.org/10.1109/ICPAI51961.2020.00058]
137. Zhou, C.; Li, J. A Real-time Driver Fatigue Monitoring System Based on Lightweight Convolutional Neural Network. Proceedings of the 2021 33rd Chinese Control and Decision Conference (CCDC); Kunming, China, 22–24 May 2021; pp. 1548-1553. [DOI: https://dx.doi.org/10.1109/CCDC52312.2021.9602058]
138. Benito-Picazo, J.; Domínguez, E.; Palomo, E.J.; Ramos-Jiménez, G.; López-Rubio, E. Deep learning-based anomalous object detection system for panoramic cameras managed by a Jetson TX2 board. Proceedings of the 2021 International Joint Conference on Neural Networks (IJCNN); Shenzhen, China, 18–22 July 2021; pp. 1-7. [DOI: https://dx.doi.org/10.1109/IJCNN52387.2021.9534053]
139. Rawat, P.; Misra, T.; Mitra, S.; Sinha, A. Designing of an Amphibian Hexapod with Computer Vision for Rescue Operations. Proceedings of the 2020 6th International Conference on Control, Automation and Robotics (ICCAR); Singapore, 20–23 April 2020; pp. 662-668. [DOI: https://dx.doi.org/10.1109/ICCAR49639.2020.9108039]
140. Wang, N.; Li, J.Y. Efficient Multi-Channel Thermal Monitoring and Temperature Prediction Based on Improved Linear Regression. IEEE Trans. Instrum. Meas.; 2022; 71, pp. 1-9. [DOI: https://dx.doi.org/10.1109/TIM.2021.3139659]
141. Jabłoński, B.; Makowski, D.; Perek, P. Evaluation of NVIDIA Xavier NX Platform for Real-Time Image Processing for Fusion Diagnostics. Proceedings of the 2021 28th International Conference on Mixed Design of Integrated Circuits and System; Lodz, Poland, 24–26 June 2021; pp. 63-68. [DOI: https://dx.doi.org/10.23919/MIXDES52406.2021.9497542]
142. Yang, R.; Yu, S.; Yu, X.; Huang, J. The Realization of Automobile Fog Lamp Intelligent Control System Based on Jetson Nano. Proceedings of the 2021 5th International Conference on Automation, Control and Robots (ICACR); Nanning, China, 25–27 September 2021; pp. 108-114. [DOI: https://dx.doi.org/10.1109/ICACR53472.2021.9605140]
143. Hong, W.C.; Huang, D.R.; Chen, C.L.; Lee, J.S. Towards Accurate and Efficient Classification of Power System Contingencies and Cyber-Attacks Using Recurrent Neural Networks. IEEE Access; 2020; 8, pp. 123297-123309. [DOI: https://dx.doi.org/10.1109/ACCESS.2020.3007609]
144. Baghezza, R.; Bouchard, K.; Bouzouane, A.; Gouin-Vallerand, C. Profile Recognition for Accessibility and Inclusivity in Smart Cities Using a Thermal Imaging Sensor in an Embedded System. IEEE Internet Things J.; 2022; 9, pp. 7491-7509. [DOI: https://dx.doi.org/10.1109/JIOT.2021.3127137]
145. Dolezel, P.; Stursa, D.; Kopecky, D.; Jecha, J. Memory Efficient Grasping Point Detection of Nontrivial Objects. IEEE Access; 2021; 9, pp. 82130-82145. [DOI: https://dx.doi.org/10.1109/ACCESS.2021.3086417]
146. Lee, J.; Jang, J.; Lee, J.; Chun, D.; Kim, H. CNN-Based Mask-Pose Fusion for Detecting Specific Persons on Heterogeneous Embedded Systems. IEEE Access; 2021; 9, pp. 120358-120366. [DOI: https://dx.doi.org/10.1109/ACCESS.2021.3108776]
147. Zheng, Z.; Liu, W.; Wang, H.; Fan, G.; Dai, Y. Real-Time Enumeration of Metro Passenger Volume Using Anchor-Free Object Detection Network on Edge Devices. IEEE Access; 2021; 9, pp. 21593-21603. [DOI: https://dx.doi.org/10.1109/ACCESS.2021.3054938]
148. Sallang, N.C.A.; Islam, M.T.; Islam, M.S.; Arshad, H. A CNN-Based Smart Waste Management System Using TensorFlow Lite and LoRa-GPS Shield in Internet of Things Environment. IEEE Access; 2021; 9, pp. 153560-153574. [DOI: https://dx.doi.org/10.1109/ACCESS.2021.3128314]
149. Fu, B.; Li, S.; Wei, J.; Li, Q.; Wang, Q.; Tu, J. A Novel Intelligent Garbage Classification System Based on Deep Learning and an Embedded Linux System. IEEE Access; 2021; 9, pp. 131134-131146. [DOI: https://dx.doi.org/10.1109/ACCESS.2021.3114496]
150. Othman, N.A.; Saleh, Z.Z.; Ibrahim, B.R. A Low-Cost Embedded Car Counter System by using Jetson Nano Based on Computer Vision and Internet of Things. Proceedings of the 2022 International Conference on Decision Aid Sciences and Applications (DASA); Chiangrai, Thailand, 23–25 March 2022; pp. 698-701. [DOI: https://dx.doi.org/10.1109/DASA54658.2022.9765087]
151. Minh, H.T.; Mai, L.; Minh, T.V. Performance Evaluation of Deep Learning Models on Embedded Platform for Edge AI-Based Real time Traffic Tracking and Detecting Applications. Proceedings of the 2021 15th International Conference on Advanced Computing and Applications (ACOMP); Ho Chi Minh City, Vietnam, 24–26 November 2021; pp. 128-135. [DOI: https://dx.doi.org/10.1109/ACOMP53746.2021.00024]
152. Han, W. A YOLOV3 System for Garbage Detection Based on MobileNetV3Lite as Backbone. Proceedings of the 2021 International Conference on Electronics, Circuits and Information Engineering (ECIE); Zhengzhou, China, 22–24 January 2021; pp. 254-258. [DOI: https://dx.doi.org/10.1109/ECIE52353.2021.00061]
153. Uddin, M.I.; Alamgir, M.S.; Rahman, M.M.; Bhuiyan, M.S.; Moral, M.A. AI Traffic Control System Based on Deepstream and IoT Using NVIDIA Jetson Nano. Proceedings of the 2021 2nd International Conference on Robotics, Electrical and Signal Processing Techniques (ICREST); Dhaka, Bangladesh, 5–7 January 2021; pp. 115-119. [DOI: https://dx.doi.org/10.1109/ICREST51555.2021.9331256]
154. Zhao, X.; Pu, F.; Wang, Z.; Chen, H.; Xu, Z. Detection, Tracking, and Geolocation of Moving Vehicle From UAV Using Monocular Camera. IEEE Access; 2019; 7, pp. 101160-101170. [DOI: https://dx.doi.org/10.1109/ACCESS.2019.2929760]
155. Wei Xun, D.T.; Lim, Y.L.; Srigrarom, S. Drone detection using YOLOv3 with transfer learning on NVIDIA Jetson TX2. Proceedings of the 2021 Second International Symposium on Instrumentation, Control, Artificial Intelligence, and Robotics (ICA-SYMP); Bangkok, Thailand, 20–22 January 2021; pp. 1-6. [DOI: https://dx.doi.org/10.1109/ICA-SYMP50206.2021.9358449]
156. Mao, Y.; He, Z.; Ma, Z.; Tang, X.; Wang, Z. Efficient Convolution Neural Networks for Object Tracking Using Separable Convolution and Filter Pruning. IEEE Access; 2019; 7, pp. 106466-106474. [DOI: https://dx.doi.org/10.1109/ACCESS.2019.2932733]
157. Rabah, M.; Rohan, A.; Haghbayan, M.H.; Plosila, J.; Kim, S.H. Heterogeneous Parallelization for Object Detection and Tracking in UAVs. IEEE Access; 2020; 8, pp. 42784-42793. [DOI: https://dx.doi.org/10.1109/ACCESS.2020.2977120]
158. Jung, S.; Hwang, S.; Shin, H.; Shim, D.H. Perception, Guidance, and Navigation for Indoor Autonomous Drone Racing Using Deep Learning. IEEE Robot. Autom. Lett.; 2018; 3, pp. 2539-2544. [DOI: https://dx.doi.org/10.1109/LRA.2018.2808368]
159. Basulto-Lantsova, A.; Padilla-Medina, J.A.; Perez-Pinal, F.J.; Barranco-Gutierrez, A.I. Performance comparative of OpenCV Template Matching method on Jetson TX2 and Jetson Nano developer kits. Proceedings of the 2020 10th Annual Computing and Communication Workshop and Conference (CCWC); Las Vegas, NV, USA, 6–8 January 2020; pp. 0812-0816. [DOI: https://dx.doi.org/10.1109/CCWC47524.2020.9031166]
160. Masnavi, H.; Adajania, V.K.; Kruusamäe, K.; Singh, A.K. Real-Time Multi-Convex Model Predictive Control for Occlusion-Free Target Tracking with Quadrotors. IEEE Access; 2022; 10, pp. 29009-29031. [DOI: https://dx.doi.org/10.1109/ACCESS.2022.3157977]
161. Wang, Y.; Tang, C.; Cai, M.; Yin, J.; Wang, S.; Cheng, L.; Wang, R.; Tan, M. Real-Time Underwater Onboard Vision Sensing System for Robotic Gripping. IEEE Trans. Instrum. Meas.; 2021; 70, pp. 1-11. [DOI: https://dx.doi.org/10.1109/TIM.2020.3028400]
162. Zhang, F.; Fan, H.; Wang, K.; Zhao, Y.; Zhang, X.; Ma, Y. Research on Intelligent Target Recognition Integrated With Knowledge. IEEE Access; 2021; 9, pp. 137107-137115. [DOI: https://dx.doi.org/10.1109/ACCESS.2021.3116866]
163. Cheng, L.; Deng, B.; Yang, Y.; Lyu, J.; Zhao, J.; Zhou, K.; Yang, C.; Wang, L.; Yang, S.; He, Y. Water Target Recognition Method and Application for Unmanned Surface Vessels. IEEE Access; 2022; 10, pp. 421-434. [DOI: https://dx.doi.org/10.1109/ACCESS.2021.3138983]
164. Demirhan, M.; Premachandra, C. Development of an Automated Camera-Based Drone Landing System. IEEE Access; 2020; 8, pp. 202111-202121. [DOI: https://dx.doi.org/10.1109/ACCESS.2020.3034948]
165. Kumar, A.; Sharma, A.; Bharti, V.; Singh, A.K.; Singh, S.K.; Saxena, S. MobiHisNet: A Lightweight CNN in Mobile Edge Computing for Histopathological Image Classification. IEEE Internet Things J.; 2021; 8, pp. 17778-17789. [DOI: https://dx.doi.org/10.1109/JIOT.2021.3119520]
166. Parthornratt, T.; Burapanonte, N.; Gunjarueg, W. People identification and counting system using raspberry Pi (AU-PiCC: Raspberry Pi customer counter). Proceedings of the 2016 International Conference on Electronics, Information, and Communications (ICEIC); Danang, Vietnam, 27–30 January 2016; pp. 1-5. [DOI: https://dx.doi.org/10.1109/ELINFOCOM.2016.7563020]
167. Meng, L.; Hirayama, T.; Oyanagi, S. Underwater-Drone With Panoramic Camera for Automatic Fish Recognition Based on Deep Learning. IEEE Access; 2018; 6, pp. 17880-17886. [DOI: https://dx.doi.org/10.1109/ACCESS.2018.2820326]
168. Chavan, S.; Ford, J.; Yu, X.; Saniie, J. Plant Species Image Recognition using Artificial Intelligence on Jetson Nano Computational Platform. Proceedings of the 2021 IEEE International Conference on Electro Information Technology (EIT); Mt. Pleasant, MI, USA, 14–15 May 2021; pp. 350-354. [DOI: https://dx.doi.org/10.1109/EIT51626.2021.9491893]
169. Venkataswamy, P.; Ahmad, M.O.; Swamy, M. Real-time Image Aesthetic Score Prediction for Portable Devices. Proceedings of the 2020 IEEE 63rd International Midwest Symposium on Circuits and Systems (MWSCAS); Springfield, MA, USA, 9–12 August 2020; pp. 570-573. [DOI: https://dx.doi.org/10.1109/MWSCAS48704.2020.9184491]
170. Wang, L.; Ye, X.; Xing, H.; Wang, Z.; Li, P. YOLO Nano Underwater: A Fast and Compact Object Detector for Embedded Device. Proceedings of the Global Oceans 2020: Singapore–U.S. Gulf Coast; Biloxi, MS, USA, 9–12 August 2020; pp. 1-4. [DOI: https://dx.doi.org/10.1109/IEEECONF38699.2020.9389213]
171. Kulathunga, G.; Hamed, H.; Devitt, D.; Klimchik, A. Optimization-Based Trajectory Tracking Approach for Multi-Rotor Aerial Vehicles in Unknown Environments. IEEE Robot. Autom. Lett.; 2022; 7, pp. 4598-4605. [DOI: https://dx.doi.org/10.1109/LRA.2022.3151157]
172. Zhou, Z.; Xu, L.; Wang, C.; Xie, W.; Wang, S.; Ge, S.; Zhang, Y. An Image Captioning Model Based on Bidirectional Depth Residuals and its Application. IEEE Access; 2021; 9, pp. 25360-25370. [DOI: https://dx.doi.org/10.1109/ACCESS.2021.3057091]
173. Yu, F.; Cui, L.; Wang, P.; Han, C.; Huang, R.; Huang, X. EasiEdge: A Novel Global Deep Neural Networks Pruning Method for Efficient Edge Computing. IEEE Internet Things J.; 2021; 8, pp. 1259-1271. [DOI: https://dx.doi.org/10.1109/JIOT.2020.3034925]
174. Park, Y.; Han, S.H.; Byun, W.; Kim, J.H.; Lee, H.C.; Kim, S.J. A Real-Time Depth of Anesthesia Monitoring System Based on Deep Neural Network With Large EDO Tolerant EEG Analog Front-End. IEEE Trans. Biomed. Circuits Syst.; 2020; 14, pp. 825-837. [DOI: https://dx.doi.org/10.1109/TBCAS.2020.2998172] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/32746339]
175. Mascret, Q.; Gagnon-Turcotte, G.; Bielmann, M.; Fall, C.L.; Bouyer, L.J.; Gosselin, B. A Wearable Sensor Network With Embedded Machine Learning for Real-Time Motion Analysis and Complex Posture Detection. IEEE Sens. J.; 2022; 22, pp. 7868-7876. [DOI: https://dx.doi.org/10.1109/JSEN.2021.3139588]
176. Baghersalimi, S.; Teijeiro, T.; Atienza, D.; Aminifar, A. Personalized Real-Time Federated Learning for Epileptic Seizure Detection. IEEE J. Biomed. Health Inform.; 2022; 26, pp. 898-909. [DOI: https://dx.doi.org/10.1109/JBHI.2021.3096127]
177. Jafari, A.; Ganesan, A.; Thalisetty, C.S.K.; Sivasubramanian, V.; Oates, T.; Mohsenin, T. SensorNet: A Scalable and Low-Power Deep Convolutional Neural Network for Multimodal Data Classification. IEEE Trans. Circuits Syst. I Regul. Pap.; 2019; 66, pp. 274-287. [DOI: https://dx.doi.org/10.1109/TCSI.2018.2848647]
178. Alamri, A.; Gumaei, A.; Al-Rakhami, M.; Hassan, M.M.; Alhussein, M.; Fortino, G. An Effective Bio-Signal-Based Driver Behavior Monitoring System Using a Generalized Deep Learning Approach. IEEE Access; 2020; 8, pp. 135037-135049. [DOI: https://dx.doi.org/10.1109/ACCESS.2020.3011003]
179. Sheng, T.J.; Islam, M.S.; Misran, N.; Baharuddin, M.H.; Arshad, H.; Islam, M.R.; Chowdhury, M.E.H.; Rmili, H.; Islam, M.T. An Internet of Things Based Smart Waste Management System Using LoRa and Tensorflow Deep Learning Model. IEEE Access; 2020; 8, pp. 148793-148811. [DOI: https://dx.doi.org/10.1109/ACCESS.2020.3016255]
180. Wang, Y.; Hou, L.; Paul, K.C.; Ban, Y.; Chen, C.; Zhao, T. ArcNet: Series AC Arc Fault Detection Based on Raw Current and Convolutional Neural Network. IEEE Trans. Ind. Inform.; 2022; 18, pp. 77-86. [DOI: https://dx.doi.org/10.1109/TII.2021.3069849]
181. Rizik, A.; Tavanti, E.; Chible, H.; Caviglia, D.D.; Randazzo, A. Cost-Efficient FMCW Radar for Multi-Target Classification in Security Gate Monitoring. IEEE Sens. J.; 2021; 21, pp. 20447-20461. [DOI: https://dx.doi.org/10.1109/JSEN.2021.3095674]
182. Xu, S.; Zhang, L.; Huang, W.; Wu, H.; Song, A. Deformable Convolutional Networks for Multimodal Human Activity Recognition Using Wearable Sensors. IEEE Trans. Instrum. Meas.; 2022; 71, pp. 1-14. [DOI: https://dx.doi.org/10.1109/TIM.2022.3158427]
183. Yang, S.; Gong, Z.; Ye, K.; Wei, Y.; Huang, Z.; Huang, Z. EdgeRNN: A Compact Speech Recognition Network With Spatio-Temporal Features for Edge Computing. IEEE Access; 2020; 8, pp. 81468-81478. [DOI: https://dx.doi.org/10.1109/ACCESS.2020.2990974]
184. Lu, S.; Qian, G.; He, Q.; Liu, F.; Liu, Y.; Wang, Q. In Situ Motor Fault Diagnosis Using Enhanced Convolutional Neural Network in an Embedded System. IEEE Sens. J.; 2020; 20, pp. 8287-8296. [DOI: https://dx.doi.org/10.1109/JSEN.2019.2911299]
185. Mukherjee, I.; Tallur, S. Light-Weight CNN Enabled Edge-Based Framework for Machine Health Diagnosis. IEEE Access; 2021; 9, pp. 84375-84386. [DOI: https://dx.doi.org/10.1109/ACCESS.2021.3088237]
186. Bhat, G.S.; Shankar, N.; Kim, D.; Song, D.J.; Seo, S.; Panahi, I.M.S.; Tamil, L. Machine Learning-Based Asthma Risk Prediction Using IoT and Smartphone Applications. IEEE Access; 2021; 9, pp. 118708-118715. [DOI: https://dx.doi.org/10.1109/ACCESS.2021.3103897]
187. Hantono, B.S.; Cahyadi, A.I.; Putu Pratama, G.N. LSTM for State of Charge Estimation of Lithium Polymer Battery on Jetson Nano. Proceedings of the 2021 13th International Conference on Information Technology and Electrical Engineering (ICITEE); Chiang Mai, Thailand, 14–15 October 2021; pp. 80-85. [DOI: https://dx.doi.org/10.1109/ICITEE53064.2021.9611940]
188. Buzura, L.; Budileanu, M.L.; Potarniche, A.; Galatus, R. Python based portable system for fast characterisation of foods based on spectral analysis. Proceedings of the 2021 IEEE 27th International Symposium for Design and Technology in Electronic Packaging (SIITME); Timisoara, Romania, 27–30 October 2021; pp. 275-280. [DOI: https://dx.doi.org/10.1109/SIITME53254.2021.9663677]
189. Vadlamani, R.; Kramer, V.; Schmidt, K. Automatic watering of plants in a pot using plant recognition with CNN. Proceedings of the 2021 5th International Conference on Electronics, Communication and Aerospace Technology (ICECA); Coimbatore, India, 2–4 December 2021; pp. 911-919. [DOI: https://dx.doi.org/10.1109/ICECA52323.2021.9675994]
190. Zheng, Y.; Zhao, C.; Lei, Y.; Chen, L. Embedded Radio Frequency Fingerprint Recognition Based on A Lightweight Network. Proceedings of the 2020 IEEE 6th International Conference on Computer and Communications (ICCC); Chengdu, China, 11–14 December 2020; pp. 1386-1392. [DOI: https://dx.doi.org/10.1109/ICCC51575.2020.9345140]
191. Lechner, M.; Jantsch, A. Blackthorn: Latency Estimation Framework for CNNs on Embedded Nvidia Platforms. IEEE Access; 2021; 9, pp. 110074-110084. [DOI: https://dx.doi.org/10.1109/ACCESS.2021.3101936]
192. Kim, J.H.; Kim, N.; Won, C.S. Deep Edge Computing for Videos. IEEE Access; 2021; 9, pp. 123348-123357. [DOI: https://dx.doi.org/10.1109/ACCESS.2021.3109904]
193. Blanco-Filgueira, B.; García-Lesta, D.; Fernández-Sanjurjo, M.; Brea, V.M.; López, P. Deep Learning-Based Multiple Object Visual Tracking on Embedded System for IoT and Mobile Edge Computing Applications. IEEE Internet Things J.; 2019; 6, pp. 5423-5431. [DOI: https://dx.doi.org/10.1109/JIOT.2019.2902141]
194. Kim, B.; Lee, S.; Trivedi, A.R.; Song, W.J. Energy-Efficient Acceleration of Deep Neural Networks on Realtime-Constrained Embedded Edge Devices. IEEE Access; 2020; 8, pp. 216259-216270. [DOI: https://dx.doi.org/10.1109/ACCESS.2020.3038908]
195. Romera, E.; Álvarez, J.M.; Bergasa, L.M.; Arroyo, R. ERFNet: Efficient Residual Factorized ConvNet for Real-Time Semantic Segmentation. IEEE Trans. Intell. Transp. Syst.; 2018; 19, pp. 263-272. [DOI: https://dx.doi.org/10.1109/TITS.2017.2750080]
196. Kim, D.S.; Arsalan, M.; Owais, M.; Park, K.R. ESSN: Enhanced Semantic Segmentation Network by Residual Concatenation of Feature Maps. IEEE Access; 2020; 8, pp. 21363-21379. [DOI: https://dx.doi.org/10.1109/ACCESS.2020.2969442]
197. Li, G.; Ma, X.; Wang, X.; Liu, L.; Xue, J.; Feng, X. Fusion-Catalyzed Pruning for Optimizing Deep Learning on Intelligent Edge Devices. IEEE Trans. Comput. Aided Des. Integr. Circuits Syst.; 2020; 39, pp. 3614-3626. [DOI: https://dx.doi.org/10.1109/TCAD.2020.3013050]
198. Ma, X.; Ji, K.; Xiong, B.; Zhang, L.; Feng, S.; Kuang, G. Light-YOLOv4: An Edge-Device Oriented Target Detection Method for Remote Sensing Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote. Sens.; 2021; 14, pp. 10808-10820. [DOI: https://dx.doi.org/10.1109/JSTARS.2021.3120009]
199. Haut, J.M.; Bernabé, S.; Paoletti, M.E.; Fernandez-Beltran, R.; Plaza, A.; Plaza, J. Low–High-Power Consumption Architectures for Deep-Learning Models Applied to Hyperspectral Image Classification. IEEE Geosci. Remote. Sens. Lett.; 2019; 16, pp. 776-780. [DOI: https://dx.doi.org/10.1109/LGRS.2018.2881045]
200. Lim, C.; Kim, M. ODMDEF: On-Device Multi-DNN Execution Framework Utilizing Adaptive Layer-Allocation on General Purpose Cores and Accelerators. IEEE Access; 2021; 9, pp. 85403-85417. [DOI: https://dx.doi.org/10.1109/ACCESS.2021.3088861]
201. Fang, W.; Wang, L.; Ren, P. Tinier-YOLO: A Real-Time Object Detection Method for Constrained Environments. IEEE Access; 2020; 8, pp. 1935-1944. [DOI: https://dx.doi.org/10.1109/ACCESS.2019.2961959]
202. Lin, J.; Gan, C.; Wang, K.; Han, S. TSM: Temporal Shift Module for Efficient and Scalable Video Understanding on Edge Devices. IEEE Trans. Pattern Anal. Mach. Intell.; 2022; 44, pp. 2760-2774. [DOI: https://dx.doi.org/10.1109/TPAMI.2020.3029799]
203. Borrego-Carazo, J.; Castells-Rufas, D.; Biempica, E.; Carrabina, J. Resource-Constrained Machine Learning for ADAS: A Systematic Review. IEEE Access; 2020; 8, pp. 40573-40598. [DOI: https://dx.doi.org/10.1109/ACCESS.2020.2976513]
204. Matsubara, Y.; Callegaro, D.; Baidya, S.; Levorato, M.; Singh, S. Head Network Distillation: Splitting Distilled Deep Neural Networks for Resource-Constrained Edge Computing Systems. IEEE Access; 2020; 8, pp. 212177-212193. [DOI: https://dx.doi.org/10.1109/ACCESS.2020.3039714]
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
Machine learning is an expanding field with an ever-increasing role in everyday life, with its utility in the industrial, agricultural, and medical sectors being undeniable. Recently, this utility has come in the form of machine learning implementation on embedded system devices. While there have been steady advances in the performance, memory, and power consumption of embedded devices, most machine learning algorithms still have a very high power consumption and computational demand, making the implementation of embedded machine learning somewhat difficult. However, different devices can be implemented for different applications based on their overall processing power and performance. This paper presents an overview of several different implementations of machine learning on embedded systems divided by their specific device, application, specific machine learning algorithm, and sensors. We will mainly focus on NVIDIA Jetson and Raspberry Pi devices with a few different less utilized embedded computers, as well as which of these devices were more commonly used for specific applications in different fields. We will also briefly analyze the specific ML models most commonly implemented on the devices and the specific sensors that were used to gather input from the field. All of the papers included in this review were selected using Google Scholar and published papers in the IEEExplore database. The selection criterion for these papers was the usage of embedded computing systems in either a theoretical study or practical implementation of machine learning models. The papers needed to have provided either one or, preferably, all of the following results in their studies—the overall accuracy of the models on the system, the overall power consumption of the embedded machine learning system, and the inference time of their models on the embedded system. Embedded machine learning is experiencing an explosion in both scale and scope, both due to advances in system performance and machine learning models, as well as greater affordability and accessibility of both. Improvements are noted in quality, power usage, and effectiveness.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer





