The future of tech: 16 trends for 2017 through 2022

Analytics, machine learning and AI will influence healthcare over the next five years.


The future of tech: 16 trends for 2017 through 2022

What will the New Year bring to analytics, machine learning and artificial intelligence? The IEEE Computer Society has released its predictions for the “Future of Tech for 2017 and the Next Five Years.”

Contributors to these predictions include: Paolo Faraboschi, fellow at Hewlett Packard Enterprise; Eitan Frachtenberg, data scientist; Hironori Kasahara, president-elect with the IEEE Computer Society; Phil Laplante, professor at Penn State University; Dejan Milojicic, distinguished technologist at Hewlett Packard Enterprise, and IEEE Computer Society past president; and John Walz, IEEE Computer Society past president.



Industrial IoT

With many millions of IoT sensors deployed in dozens of industrial-strength, real-world applications, this is one of the largest and most impactful arenas for big data analytics in 2017, says the IEEE.



Artificial intelligence, machine learning, cognitive computing

These overlapping areas are a fundamental requirement for big data analytics and for other areas of control and management, IEEE explains. “Machine learning, and deep learning in particular, are quickly transitioning from research lab to commodity products. On the software side, advanced engines and libraries from industry leaders, such as Facebook and Google, are making it to open source. On the hardware side, we see continually improving performance and scalability from existing technologies (CPUs and GPUs), as well as emerging accelerators. Consequently, writing domain-specific applications that can learn, adapt, and process complex and noisy inputs in near real time is easier than ever, and a wide range of new applications is emerging.”



5G (Fifth Generation mobile networks)

While it is unlikely that 5G will have immediate adoption in the next year, its roadmaps and standards are being developed, influencing the applications that will eventually evolve. Also, some early-use cases of deployment are being pursued, IEEE notes.



Accelerators

While looking at the long term, “the ending of Moore's Law is being addressed by novel technologies such as those covered by rebooting computing (see bullet 1 in 5 Year Trends below), heterogeneous computing founded on accelerators enables the stretching of performance boundaries in today's technologies,” IEEE says.



Disaggregated memory

“Fabric-attached Nonvolatile Memory (NVM): While NVM has achieved mixed success in productization in the past year, the number of companies working in this arena, be it on materials, architecture or software, makes it a certain candidate for imminent adoption,” IEEE explains. Fast, nonvolatile storage bridges the gap between RAM and SSDs, with a performance-cost ratio lying somewhere in between. This fast, nonvolatile storage will be initially configured either as a disc, accessed by the OS like any other permanent storage device, or as RAM in DIMM slots, accessed by the OS as memory. But once the hardware and OS support is fully figured out, this technology will open the door to new applications that aren't currently available.



Sensors everywhere and edge compute offerings

From smart transportation and smart homes, to retail innovations, surveillance, sports and entertainment, and industrial IoT, we are starting to see intelligence being aggressively deployed at the edge, IEEE notes. With intelligence comes the need to compute at the edge, and a variety of edge compute offerings are opening up new disruptive opportunities.



Blockchain (beyond Bitcoin)

While known as the technology behind Bitcoin, Blockchain has far more disruptive uses, potentially changing the way in which processes are implemented like voting, financial transactions, title and ownership, anti-counterfeiting, and digital rights managements, securing these processes without the need (and bottleneck) of a central authority, IEEE suggests.



Hyper-converged systems

Also known as “software-defined everything,” hyper-converged systems are bundles of hardware and software that contain elements of compute, storage and networking together with an orchestration system that lets IT administrators manage them using cloud tools and dev/ops practices, IEEE explains. “While they have been on the roadmap for major IT players for the last three to five years, we see major adoption trends that may cause their growth to explode in 2017.”



Technology trends that will reach adoption in 5 years:

The IEEE predicts the following trends for 2017 through 2022.



Rebooting computing (includes quantum computing)

The end of Moore's Law has resulted in the end of the ITRS (International Technology Roadmap for Semiconductors) and its transformation into IRDS (International Roadmap for Devices and Systems), focusing on new technologies, such as quantum computing, neuromorphic, adiabatic and many others, IEEE says.



Human brain interface

There are many types of interfaces developed, but the one that can be most impactful is human brain interface that can drive and control machines directly, the IEEE experts say. “This will be enabled by the rebooting computing technologies above, but will also require separate innovation to connect the human brain to hardware.”



Capabilities—hardware protection

Protecting data at rest and flight requires more sophisticated security technologies based on more robust hardware protection, such as capabilities, IEEE explains. Capabilities had been popular in the 1960s but were abandoned in favor of paging, which was sufficient when physical memory was small. Rapid advances in memory, interconnects and processors, as well as requirements of big data applications, open up new opportunities for capabilities.



The year of Exascale

The scientific community is starting to converge on 2022 being the year where they can expect the first wave of Exascale systems to be deployed, IEEE predicts. An Exascale machine would almost double the performance of all of 2016's top 500 supercomputers put together, enabling breakthroughs in scientific fields such as weather, genomics, life sciences, energy and manufacturing.”



NVM reaches maturity

There are indicators that the long-predicted adoption of NVM is coming and, by 2022, we'll be at least in the second or third generation of true nonvolatile memory devices that will change the entire memory-storage hierarchy, and associated software stack, across the IT industry, IEEE says.



Silicon photonics becomes a reality

While bridging technologies (such as VCSEL-based photonics) may be sufficient to address the needs for the next five years, IEEE sees 2022 as the pivot point where highly integrated silicon photonics components will be necessary to meet the combined cost, energy and performance requirements of Exascale systems.”



Smart NICs

Networking equipment, such as the kind seeing explosive growth in data centers, is becoming more commoditized and open, IEEE says. Ever more sophisticated chips in network interface cards (NICs) enable more offloading of traditional networking tasks from the CPU to the NIC, including encryption, compression, package management and more. “We've seen this trend before with graphics cards: commodity-specialized hardware mated with good library support enabled an explosion of applications and libraries in domains far from graphics, earning the nickname "GPGPU." Similarly, GPNICs may enable newly accelerated software to take advantage of the unique hardware properties of NICs, both within classical network applications, such as key-value stores, and in new domains, such as text processing.



Power conservative multicores

Integrated processor cores on a chip “will go over hundreds and thousands for top 500 and green 500 HPC machines,” IEEE predicts. “With more processors on a chip, memory architectures and data transfer will become key technologies in hardware. In software, a parallelizing compiler that allows users to employ the many cores efficiently and easily will reduce rapidly increasing software development costs. Automatic power reduction with the collaboration of the architecture and compiler will become crucial to apply clock or power gating or frequency and voltage lowering to idle processor cores.”



More for you

Loading data for hdm_tax_topic #better-outcomes...