Nand Kishor Contributor

Nand Kishor is the Product Manager of House of Bots. After finishing his studies in computer science, he ideated & re-launched Real Estate Business Intelligence Tool, where he created one of the leading Business Intelligence Tool for property price analysis in 2012. He also writes, research and sharing knowledge about Artificial Intelligence (AI), Machine Learning (ML), Data Science, Big Data, Python Language etc... ...

Full Bio 
Follow on

Nand Kishor is the Product Manager of House of Bots. After finishing his studies in computer science, he ideated & re-launched Real Estate Business Intelligence Tool, where he created one of the leading Business Intelligence Tool for property price analysis in 2012. He also writes, research and sharing knowledge about Artificial Intelligence (AI), Machine Learning (ML), Data Science, Big Data, Python Language etc...

3 Best Programming Languages For Internet of Things Development In 2018
431 days ago

Data science is the big draw in business schools
604 days ago

7 Effective Methods for Fitting a Liner
614 days ago

3 Thoughts on Why Deep Learning Works So Well
614 days ago

3 million at risk from the rise of robots
614 days ago

Top 10 Hot Artificial Intelligence (AI) Technologies
317847 views

Here's why so many data scientists are leaving their jobs
82302 views

2018 Data Science Interview Questions for Top Tech Companies
80358 views

Want to be a millionaire before you turn 25? Study artificial intelligence or machine learning
78243 views

Google announces scholarship program to train 1.3 lakh Indian developers in emerging technologies
62907 views

Research Team Wins Award for Machine Learning Diagnostic

By Nand Kishor |Email | Nov 17, 2017 | 5112 Views

A team of scientists hailing from the Sandia National Laboratories and Boston University developed an experimental algorithm that could automatically diagnose problems in supercomputers.

There is an array of internal and external issues that could arise with these powerful machines. For instance, factors like physical parts breaking can occur or previous programs performing "zombie processes" that prevent the computer from functioning properly.

Furthermore, the repair process for these devices can take an extended period of time, which raises another issue since these computers perform critical tasks like forecasting the weather and ensuring the U.S. nuclear arsenal is safe and reliable without needing to do underground testing.

To develop the algorithm, the team took a multi-step approach.

First, the engineers created a suite of issues they became familiar with over the time they spent working on various supercomputers, which was then followed by them writing specific codes to re-create these anomalies. 

Two supercomputers, one residing at Sandia and a public cloud system that Boston University helps operate, ran a variety of programs with and without the anomaly codes. A large quantity of data points were collected in this process including how much energy, processor power, and memory was used in each node.

Next, this trove of information was programmed into several machine learning algorithms which were able to detect anomalies by comparing data from normal program runs and those with anomalies.

In addition, these specialized programs were given additional training to determine which one was the best at diagnosing these problems.

One technique that was highlighted is called Random Forest. It was adept at analyzing vast quantities of monitoring data, identifying which metrics are important, and then determining if the supercomputer was being affected by anomaly.

Ultimately, the analysis process was further streamlined by incorporating calculations of various statistics for each metric including values like average, fifth percentile, and 95th percentile, along with more complex indications like noisiness as well as trends over time and symmetry that help suggest abnormal behavior.

The end result was a trained machine learning program that could use less than one percent of the systemÔ??s processing power to analyze data and find these complexities.

Future work on this prototype would entail more work with artificial anomalies while also finding ways to validate these diagnostics to gauge their performance in finding real anomalies during normal runs on these supercomputers.

Source: RD Mag