Scientific Sense Podcast

Saturday, December 31, 2016

A new spin on Artificial Intelligence

New research from Tohoku University (1) demonstrating pattern finding using low energy solid state devices, representing synapses (spintronics), has potential to reduce the hype of contemporary artificial intelligence and move the field forward incrementally. Computer scientists have been wasting time with conventional computers and inefficient software solutions on what they hope to be a replication of intelligence. However, it has been clear from the inception of the field that engineering processes and know-how fall significantly short of its intended goals. The problem has always been hardware design and the fact that there are more software engineers in the world than those who focus on hardware, has acted as a brake on progress.

The brain has always been a bad model for artificial intelligence. A massive energy hog that has to prop itself up on a large and fat storing gut just to survive, has always been an inefficient design to create intelligence. Largely designed to keep track of routine systems, the brain accidently took on a foreign role that allowed abstract thinking. The over design of the system meant that it could do so with relatively small incremental cost. Computer scientists' attempts to replicate the energy inefficient organ, designed primarily for routine and repeating tasks, on the promise of intelligence have left many skeletons in the long and unsuccessful path to artificial intelligence. The fact that there is unabated noise in the universe of millennials about artificial intelligence is symptomatic of a lack of understanding of what could be possible.

Practical mathematicians and engineers are a bad combination for effecting ground breaking innovation. In the 60s, this potent combination of technologists designed the neural nets - to simulate what they felt was happening inside the funny looking organ. For decades, their attempts to "train," their nets met with failure with the artificial constructs taking too long to learn anything or spontaneously becoming unstable. They continued with the brute force method as the cost of computers and memory started to decline rapidly. Lately, they have found some short cuts that allows faster training. However, natural language processing, clever video games and autonomous cars are not examples of artificial intelligence by any stretch of the imagination.

To make artificial intelligence happen, technologists have to turn to fundamental innovation in hardware. And, they may be well advised to lose some ego and seek help from very different disciplines such as philosophy, economics and music. After all, the massive development of the human brain came when they started to think abstractly and not when they could create fire and stone tools at will.

  1. William A. Borders, Hisanao Akima, Shunsuke Fukami, Satoshi Moriya, Shouta Kurihara, Yoshihiko Horio, Shigeo Sato, Hideo Ohno. Analogue spin–orbit torque device for artificial-neural-network-based associative memory operationApplied Physics Express, 2017; 10 (1): 013007 DOI: 10.7567/APEX.10.013007

Thursday, December 29, 2016

Coding errors

A recent publication in Nature Communications (1) seems to confirm that DNA damage due to ionizing radiation is a cause of cancer in humans. The coding engine in humans has been fragile, prone to mistakes even in the absence of such exogenous effects. As humans attempt interplanetary travel, their biggest challenge is going to be keeping their biological machinery, error free. Perhaps what humans need is an error correction mechanism that implicitly assumes that errors are going to be a way of life. Rather than attempting to avoid it, they have to correct it optimally.

Error detection and correction have been important aspects of electronic communication. Humans do have some experience with it, albeit in crude electronic systems. The human system appears to be a haphazard combination of mistakes made over a few million years. They have been selected for horrible and debilitating diseases and every time they step out into the sunlight, their hardware appears to be at risk. It is an ironic outcome for homosapiens who spent most of their history naked under the tropical sun. Now ionized radiation from beyond the heavens render them paralyzed and ephemeral.

Perhaps it is time we have taken a mechanistic and computing view of humans. The clever arrangement of $26 worth of chemicals seem to last a very short period of time, stricken down by powerful bugs or her own immune system. Now that bugs have been kept at a safe distance, it is really about whether the system can code and replicate optimally. The immediate challenge is error detection and correction at a molecular level. If some of the top minds, engaged in such pointless activities as investing, death curing and artificial intelligence, could focus on more practical matters, humans can certainly come out ahead.


Saturday, December 17, 2016

Does life matter?

Philosophical, ethical and religious considerations have prevented humans from defining the value of life. Short sighted financial analysis that defined the value of life as the NPV of the future utility stream, is faulty. Additionally, there is a distinct difference between personal utility and societal utility that do not coincide. The more important deficiency in the approach is that it does not account for uncertainty in future possibilities and the flexibility held by the individual in altering future decisions. And in a regime of accelerating technologies that could substantially change the expected life horizon, the value of life is increasing every day, provided expected aggregate personal or societal utility is non-negative.

The present value of human life is an important metric for policy. It is certainly not infinite and there is a distinct trade-off between the cost of sustenance and expected future benefits, both to the individual and society. A natural end to life, a random and catastrophic outcome that is imposed by exogenous factors, is highly unlikely to be optimal. The individual has the most information to assess the trade-off between the cost of sustenance and future benefits. If one is able to ignore the excited technologists, attempting to cure death by Silicon, data and an abundance of ignorance, one could find that there is a subtle and gentle slope upward for the human’s ability to perpetuate her badly designed infrastructure. The cost of sustenance of the human body, regardless of the expanding time-span of use, is not trivial. One complication in this trade-off decision is that the individual may perceive personal (and possibly societal) utility, higher than what is true.  Society, prevented from the forceful termination of the individual on philosophical grounds, yields the decision to the individual, who may not be adept enough to do so.

Humans are entering a tricky transition period. It is conceivable that creative manipulation of genes may allow them to sustain copies of themselves for a time-span, perhaps higher by a factor of 10 in less than 100 years. However, in transition, they will struggle, trying to bridge the status-quo with what is possible. This is an optimization problem that may have to expand beyond the individual, if humanity were to perpetuate itself. On the other hand, there appears to be no compelling reasons to do so.

Wednesday, December 14, 2016

Milking data

Milk, a new computer language created by MIT's Computer Science and Artificial Intelligence Laboratory (CSAIL) promises a four fold increase in the speed of analytics on big data problems. Although true big data problems are still rare, albeit the term is freely used for anything from large excel sheets to relational data tables, Milk is in the right direction. Computer chip architecture designs have been stagnant, still looking to double speed every 18 months, by packing silicon ever closer with little innovation.

Efficient use of memory has been a perennial problem for analytics, dealing with sparse and noisy data. Rigid hardware designs shuttle unwanted information based on archaic design concepts never asking the question if the data transport is necessary or timely. With hardware and even memory costs in a precipitous decline, there has not been sufficient force behind seeking changes to age old designs. Now that exponentially increasing data is beginning to challenge available hardware again and the need for speed to sift through the proverbial haystack of noise to find the golden needle is in demand, we may need to innovate again. And, Milk paves the path for possible software solutions.

Using just enough data at the right time to make decisions is a good habit, not only in computing but also in every other arena. In the past two decades, computer companies and database vendors sought to sell the biggest steel to all their customers on the future promise of analytics once they collect all the garbage and store it in warehouses. Now that analytics has "arrived," reducing the garbage into usable insights has become a major problem for companies.

Extracting insights from sparse and noisy data is not easy. Perhaps academic institutions can lend a helping hand to jump start innovation at computer behemoths, as they get stuck in the status-quo.

Monday, December 12, 2016

Democracy's event horizon

Recent results from a survey (1) of 2200 Americans showing over 1 in 4 believe that the sun goes around the earth is problematic for democracy. The system, that reflects the aggregate opinion of all participants, has served humanity well in recent years. However, the same characteristic could be its Achilles' heel as its leaders will have to reflect its population. If aggregate knowledge present in a democratic society falls below a threshold value, it can act like the event horizon of a black hole. Once through it, there is no turning back as it will spiral down to a singularity.

There have been telltale signs in many democratic societies for some time. In the world's largest democracy, elections were decided by last names and not policy choices. In Southern Europe, star power has been more dominant. More recently, powerful democratic countries have opted for less optimal outcomes. All of these may imply that democracy, as a system, is running out of its originally intended use - assure optimum outcomes for society in the long run. Instead, it is now more likely to reinforce low knowledge content, if it is dominant.

One democracy appears to have resisted the race to the bottom. Down under, where penalties are imposed for those not bothering to vote, high turn-out has assured that knowledge content of the voters is above the democratic event horizon. It appears that the prescription for ailing democracies returning sub-optimal results is to enhance voter turnout, possibly by the imposition of penalties. The biased selection in the non-voter cohort may be just enough to keep the system from the plunge to the unknown.