Article
Written by
Sachin Mishra
Published on
Monday, Aug, 18, 2025
Reading Time
8 Minutes

Artificial Intelligence and the Internet of Things, have changed the way connected devices operate, analyzed, and communicated within a multitude of industries. AIoT is generated from billions of interconnected sensors, all collecting data. Humans are now largely removed from these self-operating ecosystems, where intelligence is decentralized, and billions of endpoints making decisions, optimizing processes, and increasing real-time responsiveness. The article explores the convergence of AI and IoT, highlighting advancements in edge AI, hardware innovations, and their transformative impact across industries.

AI and IoT are in a virtuous cycle: IoT devices collect data on their environment and operations while AI is using this data to make decisions and predictions. This rapidly growing space is referred to as AIoT (Artificial Intelligence of Things), as noted by WTA LLP, and analysts predict the global AIoT marketplace could exceed $16.2 billion by 2025 (with a CAGR of ~28%), as AI algorithms are integrated into devices. This trend is fueled by the scale of IoT: an estimated 30+ billion connected things are online now and this is projected to grow toward 40 billion by 2030. Every sensor, meter or camera provides a data source for AI that can be put to use for use cases, from predictive maintenance in factories to personalized services in smart homes. AIoT involves more than merely connecting devices; ; it involves integrating intelligence directly at the edge for quicker decision-making. Contemporary systems are migrating "from centralized cloud to distributed edge." As an instance, the developing industrial IoT platforms currently offer various software capabilities such as persistent data layers and unified software stacks to manage device data both in the local system and in the cloud. This hybrid approach embraces on-device inference (ultra-low latency, privacy-friendly) to manage time-critical activities while aggregating insights at a global level, all while maintaining surveillance. There are now several methods, like federated learning and distributed inference, that utilize the edge computing power of devices to train models via device-to-device coordination, not through exchanging raw data. To sum things up, AIoT systems consider hardware and software as one ecosystem: sensors, actuators, network, and intelligence are designed together to meet application needs.

Recent advancements in hardware are crucial enablers of AIoT’s surge. High-performance edge AI chips and accelerators now enable more complex models to run directly on devices, on-site. For instance, accelerators delivering 275 TOPS now power portable industrial inspection systems. MAINGEAR’s new AI workstation, equipped with 4× RTX 5000 Ada GPUs, exceeds 1,000 TFLOPS of compute power - enabling it to train large models on-premises. Leading companies likes NXP, MediaTek and Qualcomm are integrating ARM Ethos or Cadence NPUs into their SoCs, enabling advanced AI capabilities in edge devices. These drastically accelerate inference, making it possible to run machine learning model on the edge that previously required cloud resources. At Embedded World 2024, NXP demonstrated its MCX N-Series MCU, achieving 42 x faster machine learning inference compared to CPU-only performance. ARM also showcased its Ethos-U65 NPU, which offloads 70% of inference from the CPU, delivering up to 11x faster performance. As Silicon continues to leap ahead of traditional compute, this translates into smarter edge devices - cameras, robots, gateways and more – capable of performing real-time analytics (vision, audio, and pattern recognition) directly on the edge, reducing reliance on the cloud. (Source – IOT Analytics)

At the extreme low end, TinyML enables microcontrollers to run neural networks on microwatts of power. “TinyDL” models (e.g., MCUNet, EfficientNet-lite, or compressed transformers) can now be deployed on MCUs with only 32-512 KB of RAM. This means that sub-20 ms inference can now be performed on battery-powered sensors, which is essential for wearables or remote monitors. A 100 μW CNN accelerator is a real example of this trend. Neuromorphic computing also comes into play: brain-like chips (using memristors as synapses) allow for parallel, ultra-low-power processing. Neuromorphic sensors could eventually detect patterns with a fraction of the energy consumed today , advancing AI further into constrained IoT nodes. People recognize that both sensors and connectivity are both improving at simultaneously. Camera modules now often have on-chip AI inference engine. LiDAR and higher-resolution sensors produce rich data (i.e., point clouds for precision agriculture). Meanwhile, connectivity hardware has improved significantly; the capabilities of Wi-Fi 6/7 Bluetooth LE, LPWAN modems (e.g., NBIoT, LoRaWAN), and 5G modems are now persistent. Importantly, in 2024, 5G RedCap (Reduces Capability) was introduced for IoT, providing roughly 150 Mbps downlink at sub-100 ms latency for lower-level IoT devices. This enables high-quality video and AR from wearables cameras, or drones. Satellite IoT is also maturing - low-earth-orbit (LEO) networks like Starlink, OneWeb, and Iridium now provide near real-time connectivity in remote areas, creating global IoT coverage for the first time. Overall, the common hardware trends (from SoCs to sensors to radios) are collectively laying the groundwork for "smarter" devices everywhere.
On the software side, a new generation of platforms and frameworks is converging AI and IoT. Edge Computing stacks (e.g. AWS Greengrass, Azure IoT Edge, Google Coral, etc.) allow for data preprocessing and ML inference close to the devices. Many IoT platforms now offer AI toolkits natively as part of their solution: for example, ClearBlade's Edge IoT platform includes a rules engine and an ONNX-compatible runtime that enables ML model deployments on gateways. Developers are also adopting "low-code" IoT SDKs, containerized Edge apps, and tools for AI model management that facilitate deployment. Frameworks like TensorFlow Lite, PyTorch Mobile, and MATLAB's Embedded Coder have set a standard for machine learning on microcontrollers and edge boards. These frameworks have been integrated with AutoML and neural architecture search (NAS) to optimize the size for ML models. Data architecture is also changing. Many new AIoT systems have persistent data layers that work seamlessly across cloud and edge. Middleware services and unified data buses (such as MQTT and OPC UA) are making it possible for sensor data to inform AI at the edge while also streaming back to more centralized analytics. Digital twin platforms are being developed: prototype virtual models that replicate physical assets (machines, cars, structures) in real time using IoT data. The detail data collected can facilitate 'what-if" simulations, as well as AI-enhanced optimizations (for instance, maximizing a factory's productive hours of operation, or traffic light coordination). In practice, the standardization of IoT products is incredibly important. Interoperability frameworks such as Matter (for smart home devices) and OPC UA (for industrial interoperability) are enhancing the ability of multiple IoT vendors to communicate with ease. Generative AI is an emerging area within this space. While LLMs (Large Language Models) are generally cloud-based, companies are exploring edge-scale generative models for smart assistant applications in IoT. Small transformer models running on IoT gateways could drive voice assistants for factories or vehicles. Industry is already investing in GenAI for IoT use cases: in 2024, Siemens launched an "Industrial Copilot" (a generative AI assistant for better factory operations). As hardware supports more parameters, we expect to see more on-device GenAI (e.g., for human-machine interfaces).
As we look forward, AIoT is poised for the next level of implementation. Edge devices will be capable of running more complex AI including multi-modal analytics and miniaturized LLMs. For example, companies are already starting to test on-device chatbots and voice assistants based on generative models. From a networking standpoint, 6G research trends suggest we can expect sub-millisecond latency with integrated space-air-ground connectivity and trillions of IoT endpoints. Additionally, we can also look forward to self-healing IoT networks: AI algorithms will understand system conditions and autonomously reconfigure routes or isolate failures to maintain uptime. On the hardware front, rapid evolution will continue: neuromorphic chips may transition from labs to products, enabling brain-like perception in wearables and drones. Edge AI accelerators will become even more efficient and lighter, possibly integrating tiny photonic or spintronic components. On the software side, AI frameworks will improve in model compression and adaptation, potentially even auto-designing and adapting network architecture to optimize the use of device’s hardware constraints. The digital twin concept will be much more mature significantly: we might have city-scale twins simulating entire power grids or even real-time traffic systems.
Together AI and IoT are creating proactive cyber-physical systems. These systems will not only sense their environment, but will intelligently predict and adapt to it – such as adjusting a supply chain in real time, or a traffic system responding in real time or traffic system responding dynamically to congestion. To realize this vision, the industry must continue to develop hardware-software co-design, establish robust architectures, and fostering collaboration and standardization.
Device Manufacturers
Software & Platform Providers
Enterprises & Industries
Telecom & Connectivity Providers
Policy Makers & Regulators