Skip to main content

The Hottest Things in AI: 3 Emerging Trends every AI Engineer needs to Explore

(Source: Omelchenko/Shutterstock.com)

In late 2019, IDC predicted that global spending for AI systems will grow to $97.9 billion by 2023, almost three times 2019’s $37.5 billion. Even the disruption the COVID-19 pandemic brought to 2020 has not slowed the inflection point for AI.

As an Artificial Intelligence (AI) specialist, it’s a good idea to be on the lookout for emerging trends and opportunities for skill diversification so that you can take advantage of these new trends. We have identified three areas for your consideration. Each has the potential to better prepare you for the growth explosion experts predict.

AIoT Embedded Systems

Until recently, developers only considered doing serious processing in a data center. Processing locally was not an option. But now, supercomputer performance has crossed important size, cost and power requirement thresholds. NVIDIA’s Jetson Xavier NX development kit, released in May 2020, costs under $400, and can run using just 10 Watts of power. It delivers up to 21 TOPS of computing, real power capable of running modern neural networks in parallel. and processing data from multiple high-resolution sensors, exactly what is needed for full AI system. Other competitors are not far behind.

 

These advancements greatly simplify the development of cloud-connected AI and ML multi-modal applications (Figure 1). Imagine building an advanced live video and thermal image processing system that detects potential threats from visitors with a high fever, or individuals who have been placed on a “do not admit” list. Even with 5G, it’s not practical to upload high-resolution live video and thermal sensor data to the cloud for processing in a remote server farm, then waiting for the facial-recognition and risk-assessment scoring algorithms to do their work, and results to be transmitted back, perhaps minutes later. The only solution is to process the live data locally so that the system can quickly assess the threat level.

Figure 1: Cloud-connected AI applications are greatly advancing (Source: Phonlamai Photo/Shutterstock.com)

 

This theoretical system is now possible with standard cameras, thermal imaging devices, and this 8-core NVIDIA device that sells as a complete development kit for under $400. The common AI application building blocks like object detection, text-to-speech, classification, recommender engines, language translation, sentiment analysis, and more can now be found in pre-trained models. Developing applications with these capabilities is much faster when you start with these models and then tune them for specific use cases. The NVIDIA GPU Cloud (NGC) library offers hundreds of these models to get a jump-start on common AI tasks, and they are already optimized for the Jetson Xavier and the company’s other offerings. And once you get your project running, you can easily re-train these models by updating just a few layers, saving hours.

 

These power-packed supercomputers are perfect for smart camera projects like the one described above but can also be leveraged to build an unlimited number of other high-performance AI systems like medical instruments, automated optical inspection, robots for commercial applications, and more.

Natural Language Generation/Conversational AI

Significant advances in natural language generation, specifically Microsoft’s Turing Natural Language Generation (T-NLG), have opened some doors previously inaccessible for AI developers. But this was only possible with some resourceful engineering.

 

Even with 32Gb of RAM, a single GPU cannot fit models with over 1.3 billion parameters. The Microsoft team solved this problem by taking advantage of several hardware and software breakthroughs. They leveraged a hardware setup from NVIDIA with superfast communication between GPUs, and applied tensor slicing to shard the model across four GPUs. The model-parallelism degree was able to be reduced from 16 to 4, increasing the batch size per node by 400 percent, and increasing training speed threefold. The result, DeepSpeed, trains at a batch size of 512 with only 256 GPUs compared to 1024 GPUs needed using previous configurations. The great thing is that you don’t have to understand this past paragraph to be able to start using the technology.

 

The February 2020 beta release of T-NLG makes it possible to answer search queries with complete sentences, generating language that is conversational in real-time. Imagine a chatbot that is intelligent enough to speak your language and respond in a way that makes complete sense.

 

T-NLG is the largest model ever published at 17 billion parameters making it possible to create applications like those that can assist authors with composing their content or summarizing a long text excerpt. Or, imagine improving customer experience with a digital assistant that is significantly “smarter” than the ones we are all so used to. We believe that exploring natural language generation tools will help any AI developer cultivate new skills that will be highly marketable in the next few years.

 

Getting started with T-NLG is easier than you may think, especially if you are already using Python. Head over to GitHub and explore the DeepSpeed library repo (compatible with PyTorch) and dive right in. This API library makes distributed training easy, efficient, and effective with deep-learning models that are ten times larger and five times faster, letting you leverage a hundred billion parameters to train models in record time.

Neural and Evolutionary Computing

A database is typically the bottleneck in AI systems and applications. It’s a problem that can’t easily be solved by applying more processing power or memory. One trend to watch is the application of neural networks to database system design.

 

An experimental data management system design by MIT and Google swaps in a neural network for the core components, increasing performance over cache-optimized B-Trees by up to 70% while reducing memory requirements.

 

The “learned indexes” quickly learn the structure of the database’s lookup keys, using them to predict the position of records. Neural nets have the potential to significantly impact future systems designs, so we feel it’s a topic worth exploring. So pick a topic: AIoT, NLG, or Neural Computing; or pick them all! And prepare to ride what is sure to be an exciting wave of AI innovation.

About the Author

Stephen is often invited to advise Fortune 100 companies in overall Product Strategies and Architecture Design especially as it pertains to Workflow Management, eCommerce, Artificial Intelligence and Machine Learning, bringing with him an objective view of current processes and recommending small shifts in strategy that yield big long term results and immediate ROI. As CTO / Chief System Architect, Stephen brings an in-depth knowledge of what it takes to build successful Software-as-a-Service platforms, often combining multiple legacy systems to achieve a secure, unified view of complex datasets, through scalable cloud-based architectures.

Profile Photo of Stephan Cassar