Wireless IoT Challenges in the Super-Connected World

 

One area that is growing in importance for wireless IoT is the quality of service. Christian Koehler, Manager of Product Management for RF Solutions at TE Connectivity, spoke about the various challenges facing wireless IoT in this super-connected world. Costs, scalability, security and ecosystem are all areas of concern for wireless IoT providers, Koehler said.

“There is also the complexity of antenna design,” Koehler said. “Antennas are different from regular passive components. There is a tendency to ignore this complexity and treat antennas like any other component. And the next challenge is the understanding of RF requirements. How do I know my antenna is good enough today, and more importantly, how do I know it’s good enough tomorrow?”

With the exponential growth in wireless IoT, so grows interference issues. How are today’s wireless engineers navigating these choppy waters to provide solutions with so many different concerns?

“The most advanced engineering groups are in the area of smartphones and wireless handhelds like tablets and even laptops,” Koehler said.

Proper antenna development for those devices is crucial to get right. And Koehler mentioned three vital elements for a design cycle for the new development of a device. There needs to be a clear definition of RF requirements, an understanding of how the device will operate under the worst-case field conditions and an overall understanding of its conditions.

These elements will determine the type of antenna the device will require.

“And this makes clear,” Koehler added, “the antenna cannot be designed at the end of a design cycle. It should be considered at the beginning.”

Twitter – @MarketScale
Facebook – facebook.com/marketscale
LinkedIn – linkedin.com/company/marketscale

Follow us on social media for the latest updates in B2B!

Image

Latest

custom AI chips
Custom AI Chips Signal Segmentation for AI Teams, While NVIDIA Sets the Performance Ceiling for Cutting-Edge AI
February 18, 2026

Microsoft’s introduction of the Maia 200 adds to a growing list of hyperscaler-developed processors, alongside offerings from AWS and Google. These custom AI chips are largely designed to improve inference efficiency and optimize internal cost structures, though some platforms also support large-scale training. Google’s offering is currently the most mature, with a longer production…

Read More
GPUs
OpenAI–Cerebras Deal Signals Selective Inference Optimization, Not Replacement of GPUs
February 18, 2026

OpenAI’s partnership with Cerebras has raised questions about the future of GPUs in inference workloads. Cerebras uses a wafer-scale architecture that places an entire cluster onto a single silicon chip. This design reduces communication overhead and is built to improve latency and throughput for large-scale inference. Mark Jackson, Senior Product Manager at QumulusAI, says…

Read More
nvidia rubin
NVIDIA Rubin Brings 5x Inference Gains for Video and Large Context AI, Not Everyday Workloads
February 18, 2026

NVIDIA’s Rubin GPUs are expected to deliver a substantial increase in inference performance in 2026. The company claims up to 5 times the performance of B200s and B300s systems. These gains signal a major step forward in raw inference capability. Mark Jackson, Senior Product Manager at QumulusAI, explains that this level of performance is…

Read More
autonomous trucking
Autonomous Trucking Can Shrink Coast-to-Coast Delivery Times and Increase Fleet Productivity
February 18, 2026

The idea of a self-driving 80,000-pound truck barreling down the interstate once felt like science fiction. Now, it’s operating on real freight lanes in Texas. After years of hype and recalibration, autonomous trucking is entering its proving ground. Persistent driver shortages and rising freight demand have forced the industry to look beyond incremental improvements. The…

Read More