Saturday, January 21, 2017

It's time for universal basic income

Finland's grand experiment, albeit in small scale, in providing a Universal Basic Income (UBI) without preconditions, ushers in a new dawn in modern societal design. The idea is already late for many countries as accelerating technology makes routine jobs irrelevant and any education less than college, nearly valueless. It is a regime change in such a short time that disallows gradual adjustments and it affects large swaths of populations across the world. Finely tuned welfare programs that create a disincentive for the poor to seek work and policies such as minimum wages that curb opportunities for the young to gain experience, has been creating stress in the social fabric for many decades. UBI will not only correct such disincentives but also remove the cost and inefficiencies associated with the bureaucracies that manage such programs.

The objective function for a modern society is clear - maximize aggregate happiness. Most research on happiness indicate an inverted U relationship with significant disutility in the absence of basic necessities or the fear of not having them in the future. UBI will remove such fear but avoid any disincentive effects. More importantly, UBI could provide optionality for each individual with private utility functions to select optimal pathways to maximize own happiness. If each individual has the flexibility to design such pathways, then society will unambiguously maximize aggregate happiness. What's missing from the status-quo of centrally administered myriad of welfare programs is flexibility for the individual to maximize own utility, unencumbered by the lack of basic necessities - food, shelter, health and information. UBI could provide that at a lower cost than current programs.

Universal Basic Income is conceptually and practically elegant. But to implement it, politicians have to acquire a desire to do something good during the course of their long and uninterrupted careers.

Monday, January 16, 2017

Change resistant society

As India copes with demonetization - a minor perturbation to a cash dominated society - it is clear that its long history and comparatively low diversity may be contributing to significant resistance to change - any change. A democratic system, largely driven by a few personalities since independence, has been trying to break out for ever but India remains to be a "country with unrealized but great potential." In spite of the legend, not many tortoises win races as they succumb to watching laggards zoom past them. To propel the country to the next level, it has to substantially change its attitude, and it is unlikely with the current generation, steeped in pride, history and an unwavering ability to cling to the status-quo.

Even though it is one of the largest economies on a purchasing power parity basis, it is highly insular thanks to "strategic policies," pursued by its non-capitalist leaders from inception. It does not figure in the top 100 countries on exports or imports as measured as a percent of GDP. It is also symptomatic of its lack of understanding as to how economies grow. Specializing in its own comparative advantages and freely trading with others who do other things better, is a simple economic principle, something the leaders in India never seem to have understood. No country is good at everything, India included, but this could be anathematic to the Indian diaspora, let alone its leaders.

Blindly following what seems to be successful elsewhere is an equally dangerous proposition, especially if they are pushed by policy makers enamoured by what they see on trips abroad. And, India's leaders appear to be very prone to this disease. The wisdom of a few has never been shown to be effective in understanding and implementing optimal policies. What India needs is globalization not transplantation - and the confidence to free trade and implement free markets. If so, there appears to be no stopping it but it is a tall order.


Saturday, December 31, 2016

A new spin on Artificial Intelligence


New research from Tohoku University (1) demonstrating pattern finding using low energy solid state devices, representing synapses (spintronics), has potential to reduce the hype of contemporary artificial intelligence and move the field forward incrementally. Computer scientists have been wasting time with conventional computers and inefficient software solutions on what they hope to be a replication of intelligence. However, it has been clear from the inception of the field that engineering processes and know-how fall significantly short of its intended goals. The problem has always been hardware design and the fact that there are more software engineers in the world than those who focus on hardware, has acted as a brake on progress.

The brain has always been a bad model for artificial intelligence. A massive energy hog that has to prop itself up on a large and fat storing gut just to survive, has always been an inefficient design to create intelligence. Largely designed to keep track of routine systems, the brain accidently took on a foreign role that allowed abstract thinking. The over design of the system meant that it could do so with relatively small incremental cost. Computer scientists' attempts to replicate the energy inefficient organ, designed primarily for routine and repeating tasks, on the promise of intelligence have left many skeletons in the long and unsuccessful path to artificial intelligence. The fact that there is unabated noise in the universe of millennials about artificial intelligence is symptomatic of a lack of understanding of what could be possible.

Practical mathematicians and engineers are a bad combination for effecting ground breaking innovation. In the 60s, this potent combination of technologists designed the neural nets - to simulate what they felt was happening inside the funny looking organ. For decades, their attempts to "train," their nets met with failure with the artificial constructs taking too long to learn anything or spontaneously becoming unstable. They continued with the brute force method as the cost of computers and memory started to decline rapidly. Lately, they have found some short cuts that allows faster training. However, natural language processing, clever video games and autonomous cars are not examples of artificial intelligence by any stretch of the imagination.

To make artificial intelligence happen, technologists have to turn to fundamental innovation in hardware. And, they may be well advised to lose some ego and seek help from very different disciplines such as philosophy, economics and music. After all, the massive development of the human brain came when they started to think abstractly and not when they could create fire and stone tools at will.


  1. William A. Borders, Hisanao Akima, Shunsuke Fukami, Satoshi Moriya, Shouta Kurihara, Yoshihiko Horio, Shigeo Sato, Hideo Ohno. Analogue spin–orbit torque device for artificial-neural-network-based associative memory operationApplied Physics Express, 2017; 10 (1): 013007 DOI: 10.7567/APEX.10.013007

Thursday, December 29, 2016

Coding errors

A recent publication in Nature Communications (1) seems to confirm that DNA damage due to ionizing radiation is a cause of cancer in humans. The coding engine in humans has been fragile, prone to mistakes even in the absence of such exogenous effects. As humans attempt interplanetary travel, their biggest challenge is going to be keeping their biological machinery, error free. Perhaps what humans need is an error correction mechanism that implicitly assumes that errors are going to be a way of life. Rather than attempting to avoid it, they have to correct it optimally.

Error detection and correction have been important aspects of electronic communication. Humans do have some experience with it, albeit in crude electronic systems. The human system appears to be a haphazard combination of mistakes made over a few million years. They have been selected for horrible and debilitating diseases and every time they step out into the sunlight, their hardware appears to be at risk. It is an ironic outcome for homosapiens who spent most of their history naked under the tropical sun. Now ionized radiation from beyond the heavens render them paralyzed and ephemeral.

Perhaps it is time we have taken a mechanistic and computing view of humans. The clever arrangement of $26 worth of chemicals seem to last a very short period of time, stricken down by powerful bugs or her own immune system. Now that bugs have been kept at a safe distance, it is really about whether the system can code and replicate optimally. The immediate challenge is error detection and correction at a molecular level. If some of the top minds, engaged in such pointless activities as investing, death curing and artificial intelligence, could focus on more practical matters, humans can certainly come out ahead.


(1) http://esciencenews.com/articles/2016/09/13/study.reveals.how.ionising.radiation.damages.dna.and.causes.cancer

Saturday, December 17, 2016

Does life matter?

Philosophical, ethical and religious considerations have prevented humans from defining the value of life. Short sighted financial analysis that defined the value of life as the NPV of the future utility stream, is faulty. Additionally, there is a distinct difference between personal utility and societal utility that do not coincide. The more important deficiency in the approach is that it does not account for uncertainty in future possibilities and the flexibility held by the individual in altering future decisions. And in a regime of accelerating technologies that could substantially change the expected life horizon, the value of life is increasing every day, provided expected aggregate personal or societal utility is non-negative.

The present value of human life is an important metric for policy. It is certainly not infinite and there is a distinct trade-off between the cost of sustenance and expected future benefits, both to the individual and society. A natural end to life, a random and catastrophic outcome that is imposed by exogenous factors, is highly unlikely to be optimal. The individual has the most information to assess the trade-off between the cost of sustenance and future benefits. If one is able to ignore the excited technologists, attempting to cure death by Silicon, data and an abundance of ignorance, one could find that there is a subtle and gentle slope upward for the human’s ability to perpetuate her badly designed infrastructure. The cost of sustenance of the human body, regardless of the expanding time-span of use, is not trivial. One complication in this trade-off decision is that the individual may perceive personal (and possibly societal) utility, higher than what is true.  Society, prevented from the forceful termination of the individual on philosophical grounds, yields the decision to the individual, who may not be adept enough to do so.

Humans are entering a tricky transition period. It is conceivable that creative manipulation of genes may allow them to sustain copies of themselves for a time-span, perhaps higher by a factor of 10 in less than 100 years. However, in transition, they will struggle, trying to bridge the status-quo with what is possible. This is an optimization problem that may have to expand beyond the individual, if humanity were to perpetuate itself. On the other hand, there appears to be no compelling reasons to do so.

Wednesday, December 14, 2016

Milking data

Milk, a new computer language created by MIT's Computer Science and Artificial Intelligence Laboratory (CSAIL) promises a four fold increase in the speed of analytics on big data problems. Although true big data problems are still rare, albeit the term is freely used for anything from large excel sheets to relational data tables, Milk is in the right direction. Computer chip architecture designs have been stagnant, still looking to double speed every 18 months, by packing silicon ever closer with little innovation.

Efficient use of memory has been a perennial problem for analytics, dealing with sparse and noisy data. Rigid hardware designs shuttle unwanted information based on archaic design concepts never asking the question if the data transport is necessary or timely. With hardware and even memory costs in a precipitous decline, there has not been sufficient force behind seeking changes to age old designs. Now that exponentially increasing data is beginning to challenge available hardware again and the need for speed to sift through the proverbial haystack of noise to find the golden needle is in demand, we may need to innovate again. And, Milk paves the path for possible software solutions.

Using just enough data at the right time to make decisions is a good habit, not only in computing but also in every other arena. In the past two decades, computer companies and database vendors sought to sell the biggest steel to all their customers on the future promise of analytics once they collect all the garbage and store it in warehouses. Now that analytics has "arrived," reducing the garbage into usable insights has become a major problem for companies.

Extracting insights from sparse and noisy data is not easy. Perhaps academic institutions can lend a helping hand to jump start innovation at computer behemoths, as they get stuck in the status-quo.

Monday, December 12, 2016

Democracy's event horizon

Recent results from a survey (1) of 2200 Americans showing over 1 in 4 believe that the sun goes around the earth is problematic for democracy. The system, that reflects the aggregate opinion of all participants, has served humanity well in recent years. However, the same characteristic could be its Achilles' heel as its leaders will have to reflect its population. If aggregate knowledge present in a democratic society falls below a threshold value, it can act like the event horizon of a black hole. Once through it, there is no turning back as it will spiral down to a singularity.

There have been telltale signs in many democratic societies for some time. In the world's largest democracy, elections were decided by last names and not policy choices. In Southern Europe, star power has been more dominant. More recently, powerful democratic countries have opted for less optimal outcomes. All of these may imply that democracy, as a system, is running out of its originally intended use - assure optimum outcomes for society in the long run. Instead, it is now more likely to reinforce low knowledge content, if it is dominant.

One democracy appears to have resisted the race to the bottom. Down under, where penalties are imposed for those not bothering to vote, high turn-out has assured that knowledge content of the voters is above the democratic event horizon. It appears that the prescription for ailing democracies returning sub-optimal results is to enhance voter turnout, possibly by the imposition of penalties. The biased selection in the non-voter cohort may be just enough to keep the system from the plunge to the unknown.

(1) http://www.npr.org/sections/thetwo-way/2014/02/14/277058739/1-in-4-americans-think-the-sun-goes-around-the-earth-survey-says

Monday, November 28, 2016

The irony of Artificial Intelligence

The indiscriminate use of the term "Artificial Intelligence," by analysts and consultants, demean what it was originally meant to be and is still supposed to imply. Those who have worked in this area for decades, much before the internet, search and autonomous vehicles existed, know fully well that there is nothing "artificial" about it. A lack of definition for "consciousness," has constrained computer scientists and philosophers alike from conceptualizing intelligence. Faster computers, memory and unlimited capital are unlikely to advance the field any further unless we return to thoughtfully studying the underlying issues. Let's do that before a few gamers think that they have accidently invented the "artificial mind."

Intelligence, something solely attributed to humans till very recently, is now shown to be present in a plethora of biological systems. Biological systems, however, are not an infinite array of on-off switches nor do they process information as computer systems do. So the idea that one can network a large number of silicon based idiot boxes, using mathematics from the 60s, to replicate the brain is fraught with issues. The recent fad of "artificial general intelligence," - the ability to teach a computer virtually anything such that it can not only replicate a human but also become a lot better, is a nice fantasy. What the technologists may be missing is that it is not a software problem. The millennials have become good at software but projecting what one is good at onto hard problems may not the best path forward.

Nature had nearly 3.8 billion years to perfect the underlying hardware to deliver intelligence. It was a slow and laborious process and it was never a software problem. Once the hardware was perfected, it was able to run virtually any software. This may give a clue to those plunging head first into teaching machines to "think," using software and age old mathematics. More importantly, the current architecture of computing representing calculators is not amenable to modeling intelligence. Just as there is a distinct difference between Arithmetic and Mathematics, conventional computers differ significantly from a true quantum computer. Adding and subtracting numbers fast is one thing but conceptualizing what it means is quite another.

The unfortunate term, "Artificial Intelligence," has led many mediocre lives astray. And God, with an immense sense of humor, seem to still lead them down blind alleys.