Shoumen Palit Austin Datta1,2,3,*, Tausifa Jan Saleem4, Molood Barati5, MarÃa Victoria López López6, Marie-Laure Furgala7, Diana C. Vanegas8, Gérald Santucci9, Pramod P. Khargonekar10, and Eric S. McLamore11
1MIT Auto-ID Labs, Department of Mechanical Engineering, Massachusetts Institute of Technology, 77 Massachusetts Avenue, Cambridge, MA 02139, USA
2MDPnP Interoperability and Cybersecurity Labs, Biomedical Engineering Program, Department of Anesthesiology, Massachusetts General Hospital, Harvard Medical School, 65 Landsdowne Street, Cambridge, MA 02139, USA
3NSF Center for Robots and Sensors for Human Well-Being, Collaborative Robotics Lab, School of Engineering Technology, Purdue University, 193 Knoy Hall, West Lafayette, IN 47907, USA
4Department of Computer Science and Engineering, National Institute of Technology Srinagar, Jammu & Kashmir 190006, India
5School of Engineering, Computer and Mathematical Sciences Auckland University of Technology, Auckland 1010, New Zealand
6Facultad de Informática, Deparmento Arquitectura de Computadores y Automática, Universidad Complutense de Madrid, Calle Profesore Santesmases 9, 28040 Madrid, Spain
7Director, Institut Supérieur de Logistique Industrielle, KEDGE Business School, 680 Cours de la Libération, 33405 Talence, France
8Biosystems Engineering, Department of Environmental Engineering and Earth Sciences, Clemson University, Clemson, SC 29631, USA
9Former Head of the Unit, Knowledge Sharing, European Commission (EU) Directorate General for Communications Networks, Content and Technology (DG CONNECT); Former Head of the Unit Networked Enterprise & Radio Frequency Identification (RFID), European Commission; Former Chair of the Internet of Things (IoT) Expert Group, European Commission (EU); INTEROP-VLab, Bureau Nouvelle Région Aquitaine Europe, 21 rue Montoyer, 1000 Brussels, Belgium
10Vice Chancellor for Research, University of California, Irvine and Distinguished Professor of Electrical Engineering and Computer Science, University of California, Irvine, California 92697
11Department of Agricultural Sciences, Clemson University, Clemson, SC 29634, USA
Since 1999, the concept of the Internet of Things (IoT) was nurtured as a marketing term [2] which may have succinctly captured the idea of data about objects stored on the Internet [3] in the networked physical world. The idea evolved while transforming the use of radio frequency identification (RFID) where an alphanumeric unique identifier (64‐bit EPC [4] or electronic product code) was stored on the chip (tag [5]) but the voluminous raw data were stored on the Internet, yet inextricably and uniquely linked via the EPC, in a manner resembling the structure of internet protocols [6] (64‐bit IPv4 and 128‐bit IPv6 [7]). IoT and, later, cloud of data [8] were metaphors for ubiquitous connectivity and concepts originating from ubiquitous computing, a term introduced by Mark Weiser [9] in 1998. The underlying importance of data from connected objects and processes usurped the term big data [10] and then twisted the sound bites to create the artificial myth of “Big Data” sponsored and accelerated by consulting companies. The global drive to get ahead of the “Big Data” tsunami, flooded both businesses and governments, big and small. The chatter about big data garnished with dollops of fake AI became parlor talk among fish mongers [11] and gold miners, inviting the sardonicism of doublespeak, which is peppered throughout this essay.
Much to the chagrin of the thinkers, the laissez‐faire approach to IoT percolated by the tinkerers overshadowed hard facts. The “quick & dirty” anti‐intellectual chaos adumbrated the artifact‐fueled exploding frenzy for new revenue from “IoT Practice” which spawned greed in the consulting [12] world. The cacophony of IoT in the market [13] is a result of that unstoppable transmutation of disingenuous tabloid fodder to veritable truth, catalyzed by pseudo‐science hacks, social gurus, and glib publicity campaigns to drum up draconian “dollar‐sign‐dangling” predictions [14] about “trillions of things connected to the internet” to feed mass hysteria, to bolster consumption. Few ventured to correct the facts and point out that connectivity without discovery is a diabolical tragedy of egregious errors. Even fewer recognized that the idea of IoT is not a point but an ecosystem, where collaboration adds value.
The corporate orchestration of the digital by design metaphor of IoT was warped solely to create demand for sales by falsely amplifying the lure of increasing performance, productivity, and profit, far beyond the potential digital transformation could deliver by embracing the rational principles of IoT (Figures 2.1–2.4).
Ubiquitous connectivity is associated with high cost of products (capex or capital expense) but extraction of “value” to generate return on investment (ROI) rests on the ability to implement SARA, a derivative of the PEAS paradigm (see Figures 2.7 and 2.8). SARA – Sense, Analyze, Respond, Actuate – is not a linear concept. Data and decisions necessary for SARA make the conceptual illustration more akin to The Sara Cycle, perhaps best illustrated by the analogy to the Krebs [28] Cycle, an instance of bio‐mimicry. Data and decisions constantly influence, optimize, reconfigure, and change the parameters associated with, when to sense, what to analyze, how to respond, and where to actuate or auto‐actuate. Combining SARA with the metaphor of IoT by design may help to ask these questions, with precision and accuracy.
It is hardly necessary to overemphasize the value of the correct questions for each element of SARA in a matrix of connected objects, relevant entities which can be discovered, distributed nodes, related processes, and desired outcomes. Strategic inclusion of SARA guides key performance indicators (KPI). Lucidity and clarity of thoughtful integration of digital by design idea is key to reconfiguring operations management. Execution and embedding SARA is not a systems integration task but rather a fine‐tuned synergistic integration based on the weighted combination of dependencies in the SARA matrix. Failure to grasp the role of data and semantics of queries, in the context of KPI may increase transaction costs, reduce the value proposition for customers, and obliterate ROI or profitability.
This essay meanders, not always aimlessly, around discussions involving data and decision. It also oscillates, albeit asynchronously, between a broad spectrum of haphazard realities or “dots” which may be more about esoteric analysis rather than focusing on delivering real‐world value. In part, this discussion questions the barriers to the rate of diffusion of technologies in underserved communities. Can implementing simple tools act as affordable catalysts? Can it lift the quality of life, in less affluent societies, by enabling meaningful use of data, perhaps small data, at the right time, at the lowest cost?
The extremely nonlinear business of delivering tools and technologies makes it imperative to consider the trinity of systems’ integration, standards, and interoperability. We advocate that businesses may wish to gradually disengage with the product mindset (sensors, hardware, and software) and engage in the ecosystem necessary to deliver services to communities. The delivery of service to the end‐user must be synergized. Hence, system integration may be a subset of synergistic integration. But, before we can view this “whole,” it is better to understand the coalition of cyber (data) with the physical (parts). In many ways, this discussion is about cyberphysical systems (CPS) but not for lofty purposes, such as landing on Mars, but for simple living, on Earth.
Because it may be difficult to grasp the whole, we tend to focus on the part, and parts, closest to our comfort zone, in our area of interest. This reductionist approach may be necessary ab initio but rarely yields a solution, per se. Reconstruction requires synthesis and synergy, the global glue which underlies mass adoption and diffusion, of tools, in an age of integration, which, itself, is a khichuri [29] of parts, some known (industrial age, information age, and systems age) and others, parts unknown.
Divide and conquer still remains a robust adage. It may be the philosophical foundation of reductionism. The latter has rewarded us with immense gains in knowledge and the wisdom as to why this modus operandi is sine qua non. For example, the pea plant (Pisum sativum) unleashed the cryptic principles of genetics [30] and unicellular bacteria shed light on normal physiological underpinnings of feedback control [31] common in genetic circuits as well as regulatory networks for maintenance and optimization of biological homeostasis, quintessential for health and healthcare in humans and animals. Cancer biology was transformed by Renato Dulbecco [32] by reducing the multifactorial complexity of human cancer research to focus on a single gene (the SV40 large T‐antigen) from Papova viruses.
Biomimicry also inspired the creation of better machines and systems [17], using the principles and practice of control theory borrowed from science, strengthened by mathematics and successfully integrated with design and manufacturing, by engineers. An early convergence [33] of control theory with communication may be found in the 1948 treatise “Cybernetics” by Norbert Wiener [34] (who may have borrowed [35] the word “cybernétique” proposed by the French physicist and mathematician André‐Marie Ampère [18] to design the then nonexistent science of process control).
In other examples of “divide and conquer,” the theoretical duo “Alice and Bob” is at the core [36] of cryptography [37] as well as the game theoretic [38] approach [39] to “prisoner’s dilemma” which has influenced business strategies [40] and now it is spilling over to knowledge graph (KG) [41] databases. The simple concept of a lone travelling salesman proposed by Euler in 1759 appears to have evolved [42] as the bread and butter of most optimization engines, which, when considered together with data and information, continues to improve decision support systems (DSS) in manufacturing, retail, transportation, logistics [43], and omnipresent supply chain [44] networks, almost in every vertical which uses DSS.
The purpose of these disparate examples are to emphasize the notion that there are fundamental units of activity or models or set(s) of patterns or certain basic behavioral criteria (for lack of a better descriptive term) that underlie most actions and reactions. When taken apart or sufficiently reduced, we may observe these as isolated units or patterns or models of rudimentary entities. When combined, these simple models/units/patterns/elements can generate an almost unlimited variety of system behaviors observed on grand scales. When viewing the massive scale of systems from the “top,” it may be quite counterintuitive to imagine that the observed manifestations are due to a few or a relatively small group of universal “truths” which we refer to as models, units, rules, logic, patterns, elements, or behaviors. To further illustrate this perspective, consider petals (flowers), pineapple (fruit), and pyramids. The variation between and within these three very different examples may boil down to Fibonacci [45] numbers, fractal [46] dimensions, and the Golden [47] Ratio [48] in some form, or the other. In another vein, the number, eight, seems to be central to atoms (octet) and an integral part of the Standard Model in physics (octonions [49]). Number 8 is revered by the Chinese due to its link with words synonymous with wealth and fortune (fa).
If one is still unconvinced and remain skeptical that small sets of underlying elements, generally, may be responsible, albeit in part, for the “big things” we consider diverse, then the “killer” example is that of nucleic acids, deoxyribonucleic acid (DNA) and ribonucleic acid (RNA), made up of only five subunits or molecules (adenine, guanine, cytosine, thymine, and uracil). DNA and RNA serve as the blueprint for all humans, animals, plants, bacteria, and viruses that may ever exist. The infinite diversity of multicellular [50] and unicellular organisms, whose creation is instructed by a combination of these five molecules in DNA and RNA, may vastly exceed 5 × 1030 (5,000,000,000,000,000,000,000,000,000,000 [51]). The known exception to the DNA–RNA dogma may be the case for prions [52] which uses proteins [53] as the transmissible macromolecule.
Parallel examples can be drawn from physical sciences. Large‐scale system behaviors can be reduced and mapped to simple models. Combination of these simple models, with widely different microscopic details, applies to, and generates, a large set of possible systems [54] and system of systems. Another example of “hidden complementarities” emerged from cryptic mathematical bridge embedded in natural sciences. It is now established that eigenvectors may be computed [55] using information about eigenvalues. Students are still taught that eigenvectors and eigenvalues are independent and must be calculated separately starting from rows and columns of the matrix. Mathematicians authored papers in related fields [56] yet none “connected the dots” between eigenvectors and eigenvalues. The insight that eigenvalues of the minor matrix encode hidden information may not be entirely new [57] but was neither understood nor articulated. The relationship of centuries‐old mathematical objects [58] ultimately came from physicists. Nature inspires mathematical thinking because mathematics thrives when connected to nature. Grasping these connections enables humans to create tools to mimic nature (bio‐mimicry).
The lengthy and winding preface is presented to substantiate the opinion that there may be a disconnect between the volume of data we have generated as a result of the “information age” versus the lackluster gains in performance, as estimated by the productivity [59] index. We may have 2.7 zettabytes [20] (2.7 billion terabytes) of data, but some estimates claim as much as 33 zettabytes [60] of data, at hand (2018). It is projected to reach 175 zettabytes circa 2025.
The deluge of data as a result of “information technology” is far greater in magnitude than the diffusion of electricity [61] a century ago. Productivity increases due to the introduction of electricity and IT offers economic parallels [62] but based on the magnitude of change, the shortfall (in productivity) cannot be brushed aside by attributing the blame to mismeasurement explanations [63] for the sluggish [64] pace. Extrapolating measurements using the tools of classical productivity [65] to determine the impact of IT and influence of data is certainly fraught with problems [66], yet the incongruencies alone cannot explain the shrinkage. In socioeconomic terms, there is a growing chasm between IT and data/information versus productivity, improvement in quality of life, labor, compensation [67], and standard of living.
Despite trillions of dollars invested in data, digital transformation and other IT tools [68] (big data, AI, blockchain), the perforated ROI [69] increasingly points to massive [70] waste. One reason for this “waste” may be due to use of models of data where errors are aggregated under a generalized [71] form or variations [72] of the normal (homoskedastic) distribution. Heteroskedasticity was addressed [73] using ARCH [74] (autoregressive conditional heteroskedasticity [75]) and GARCH [76] models [77] (generalized ARCH). The use [78] of these proven techniques [79] for time series data (for example, sensor data showing water temperature in marine aquaponics [80] or cold chain [81] temperature log of vaccine package during transportation) in financial [82] econometrics [83] may be extended. Applications in predictive [84] modeling and forecasting [85] techniques may wish to adopt these econometric tools (GARCH) as a standard, whenever time series data are used (for example, supply chain [86] management, sensor data in health), but only if there is sufficient data (volume) to meet the statistical rigor necessary for successful error correction.
Perhaps, it is best to limit the postmortem analysis of IT failures, snake‐oil sales of AI [87], and other debacles. Let us observe from this discussion that in the domain of data, and extraction of value from data to inform decisions and the tools necessary for meaningful transformation of data to inform decisions may benefit from re‐viewing the processes and technologies with “new” eyes. We must ask, often, if we are pursuing the correct questions, if the tools are appropriate and rigorous. The productivity gap and reports of corporate waste are “sign‐posts” on the road ahead, except that the signage is in the incorrect direction, with respect to the intended destination, that is, profit and performance.
There are no novel proposed solutions in this essay, only new commentary about approaches to solutions. The violent discord between volume of data versus veracity of decisions appears to be one prominent reason why the productivity gap may widen to form a chasm. The “background” section discussed how the reductionist approach points to simple models or underlying units or key elements, which, when combined, in some form, by some rules or logic, may generate large‐scale systems.
Data models [88] for DBMS are very different from models in data. Pattern mining [89] from data [90] is a time‐tested tool. What new features can we uncover or learn about data, from patterns? What simpler models or elements are cryptic in data? Are these the correct questions? If there are simpler models or patterns in some types of data, can we justify extrapolating these models and patterns as a general feature of the data? The failure to accept and curate data which may be void of information is of critical importance. The contextual understanding of this issue appears to be uncommon and tools for semantic data curation are nonexistent. Although we have been mining for patterns and models (clustering, classification, categorization, and principal component analysis) for decades, why have not we found simpler models or patterns, yet? Are we using the wrong tools or wrong approaches or looking at wrong places? How rational are we in our search for these general/simple models in view of the fact that models of data from retail or manufacturing or health clinics should be quite different? Is model building by humans an irrational approach since humans are innate, irrational organisms endowed with sweeping bias?
Thus, the lowest common denominator of general models/patterns may not be an ingredient for building that experimental “thought” bridge. Increasing volume of data could help GARCH tools but it is a slippery slope in terms of data quality with respect to informing DSS and/or the veracity of decisions (output). Data models/patterns as denominators from grocery shopping or dry wall manufacturing or mental health clinics are different. In lieu of “universal” common denominators, we may create repertoires of domain‐specific common denominators. A comparative analysis between common denominators of retail grocery shopping model from Boston vs. Beijing may reveal the spectrum of nutritional behaviors. If linked to eating habits, perhaps we can extrapolate its influence on health/mental health. As this suggestion reveals, we may be able to explore very tiny subsets of models.
Domain‐specific denominator models (DSDM) are not new. It requires an infrastructure approach to data analytics which needs multitalented teams to explore almost every cross section and combination of very large volumes of data, from specific domains, to identify obvious correlations as well as unknown/nonobvious relationships. If there is any doubt about the quality of the raw data, then quality control may mandate data curation. The latter alone, makes the task exponentially complex. Curation may introduce reasonable doubt in evaluating any outcome because the possibility exists that curation algorithms and associated processes were error‐prone or untrustworthy (post‐curation jitters).
Another demerit for DSDM and the idea of denominator models, in general, may be rooted in the “apples vs oranges” dilemma. Denominator models that underlie science and engineering systems are guided by natural laws, deemed rational. The quest for denominator models in data (retail, finance, supply chain, health, and agriculture) are influenced, infected, and corrupted by irrational [91] human behavior. Rational models of irrational behavior [92] may coexist elsewhere but remains elusive for data science due to volatility and the vast spectrum of irrationality that may be introduced in data by human interference.
Perhaps, the concept of DSDM, ignoring its obvious caveats, may be applied to select domains for specific purposes, for example, healthcare, where deliberate human interference to introduce errors in data is a criminal offense. Case‐specific model building, and pattern recognition, may benefit from machine learning (ML) approaches. The latter fueled a plethora of false [93] claims but real success is still a work in progress because the bridge between data and decisions will be perpetually under construction. Productivity gap and corporate waste are indicators that existing approaches (see Figure 2.5) are flawed, failing, or have [94] failed. We need new roads. The boundary of our thought horizon “map” is in Figure 2.5. The tools are incremental variations [95] garnished with gobbledygook alphabet soup. Unable [96] to create any breakthrough, the return of seasonal “winters of AI” indicates the struggle to shed new light in this field since the grand edification [97] during the 1950s. Unable to cope with data challenges, hard facts [85], and difficult progress, the field offered a perfect segue for con artists and hustlers to inculcate falsehoods and deceive [98] the market. ML was substituted [99] by mindless drivel from ephemeral captains of industry and generated hype [100] from corporate [101] marketing machines as well as greedy academics.
Data consumers have been led astray by vacuous buzz words manufactured mostly by consulting groups. Part of the productivity gap may be due to fake news, propaganda [94], and glib strategy from smug consultants to coerce large contracts with cryptic “billable hours” to help “monetize” false promises due to “big” data, fabricated [102] claims [103] of “intelligence” in artificial intelligence (AI) [104], and deliberately conniving misrepresentations [105] of “blockchain” as a panacea [106] for all problems [107] including basic food safety and security. Callous and myopic funding agencies invested billions in academic [108] industry partnerships to fuel banal R&D efforts orchestrated by corporate collusion [109] and perhaps [110] criminal [111] practices. Abominable predatory practices on display in Africa are disguised under the “smart cities” marketing campaign to mayors of African cities, which cannot even provide clean drinking water to its residents. Vultures from the industry [112] are selling mayors of African cities surveillance technology and AI in the name of cameras for smart city safety and security. These behemoths are cognizant as to how autocrats use data as an ammunition to plan and justify abuse of its citizens, through algorithms of repression.
Uploading data from nodes along a variety of supply chains is an enormous undertaking given trillions of interconnected processes and billions of nodes with extraordinarily diverse categories of potential data streams, with different security mandates, for example, (i) sensor data about heavy metal (mercury) contamination in water used for irrigation, (ii) near real‐time respiratory rate (RR) of patient with chronic obstructive pulmonary disease (COPD) under remote monitoring telemedicine in rural nursing home, and (iii) automated check‐out scan data from retail grocery store sales, of fast‐moving consumer goods, contracted for replenishment (penalty for out of stock) under vendor managed inventory (VMI). The e‐tail revolution is creative supply chain optimization and reducing retail information asymmetry.
Transforming data and data analytics to inform decision support for small cross‐sections of examples cited, here, may be theoretically easy in “power point” diagrams which “connects” nodes and integrate decision feedback to optimize processes, using pixels. The reality may be different. Aggregating data from various nodes, sub‐nodes, devices, and processes, on a platform, to enable collective evaluation of dependencies, which could influence outcomes/decisions, may be beneficial, or germane for certain domains, for example, healthcare [113] and clinical [114] environments where patient safety [115] must be of paramount importance.
Agreement on any one standard platform is unlikely to succeed. But an anastomosis of platforms is probably rational, if interoperable. An open platform of platforms with secure, selective, interoperable data exchange, between platforms, may be valuable. Synthesis and convergence of data acquisition and analytics begins to catalyze information flow, decision support, and meaningful [116] use of data [117]. This suggestion is a few decades old, but still far from practice. The drive to connect data was accelerated by the introduction of the concept of the IoT [118]. Platform [119] efforts [120] are addressing data [121] upload from devices and sensors [122] but nowhere near a turn‐key implementation (Figure 2.6).
There are not any silver bullets and one shoe does not fit [124] all. If we focus on the data to decision process, alone, in any vertical or domain, the variations of analysis and analytics may be astronomical. Initial investments necessary for these endeavors almost guarantee that the extracted value from data (and relevant information) may not be democratized or made functionally available to those who cannot pay the high cost. In principle, the outcome from data to decisions, when appropriate, may be sufficiently distributed and democratized to provide value for communities under economic constraints. Any meaningful solution, therefore, is not a scientific or engineering outcome, alone, but must be combined with the economics of technology [125] which must be a catalyst for implementation and adoption by the masses, if transaction costs [126] can be sustained by the community of users, in less affluent geographies.
The economic principle for impoverished environments may be rooted in microfinance [127] and micro‐payments [128] with low transaction costs (the downside: misinformation [129] can be propagated and disseminated at low cost, too). By eliminating classical “product sales,” the focus shifts to delivery of “service” which is a package of the product plus other resources (retail mobile banking, infrastructure, telecommunications, cybersecurity [130], security [131], and customer service). Users pay (pennies) only when they use the service. Pay‐a‐penny‐per‐use or pay‐a‐price‐per‐unit [132] (PAPPU) is a metaphor for economic instruments which may lower the barrier to entry into markets with billions of users.
The economic incentive for democratization of data is the potential to unleash and create new markets for data, information, and decision support, for billions of new consumers (users). The reward in the lucrative service economy model depends on harvesting the economies of scale where each user (market of billions) may pay one or more “pennies” (micro‐payment for pay‐per‐use services). The risk in the service economy is the collection of that “penny” (per use) at the last step of the seamless service delivery process, if the user is satisfied with the quality of service (QoS) metrics. The plethora of partners necessary to create and sustain the ecosystem to deliver the seamless service is a herculean task. Sharing a fraction of that “penny” with the partners in the ecosystem is not a trivial challenge. If the QoS delivery metrics suffer due to poor performance of any one partner (component), the end‐user “penny” may be unpaid if the QoS metric fails to reach a predetermined value (time, duration, speed, rate, and volume). The inability of one provider (weakest link) in the service supply chain can be financially detrimental to all other supply chain partners due to loss of that penny, albeit, only for that transaction (unless the partner has a chronic problem, then, it must be excluded from the ecosystem and the entire value network [133]). Delivery of service is a real‐time convergence of operations management which includes (but is not limited to) multiple value chains which must integrate [134] the physical supply chain and the financial supply chain with the service supply chain and customer relationship management (brand expectation).
Determining the cost of execution, to deploy the example in Figure 2.6, may be one way to study feasibility. Simulating models to explore financial engineering of “what if” scenarios, may project the potential for adoption of services in the context of various economies of scale and PAPPU models. The reward for unchaining the economics of technology is in adoption, by the next billion users.
Decision scientists must build a compass to help extract value from data. One compass will not suffice to guide domain‐specificity. Existing tools may limp along with snail‐ish advances (Figure 2.5) yet it may remain inaccessible to the masses because the tools may not be feasible for mass adoption. The struggle to transform data into information is still in quest of a Renaissance.
The path from data‐informed to information‐informed to knowledge‐informed decision remains amorphous. Transforming information to knowledge is in the realm of unknown unknowns. Making sense of data is handicapped due to (i) an apparently insurmountable semantic barrier, (ii) scarcity of tools to facilitate location‐aware and context‐aware discovery of data at the edge or point of use, and (iii) lack of standards and interoperability between objects, platforms, and devices for data and analytics sharing.
Users in less affluent nations may not want to idle away while the architects of Renaissance are still in short supply. In the near term, it is necessary that we continue to work on dissemination of data which can deliver at least some value, sooner, rather than later. Decision support based on sensor data analytics may provide economic benefits [135] and incentives, if we can share the digital dividends with the masses, for example, in health [26] and agriculture, including every facet of food, required daily, globally.
Tangle, a tool [136] to share sensor data using masked authenticated messaging [137] (MAM) may offer hope. Can nano‐payments for sensor data address some of the feasibility challenges [138] and pave the way for human‐centric economy of things [139] using IoT as a design metaphor? SNAPS [140] is a tiny step in that general direction: distributing low‐cost tools to enable data‐informed decision support (DIDAS) for less complex problems. Assuming the Pareto principle holds true, perhaps 80% of the problems may be addressed, and even resolved, with simple tools to deliver solutions as a service, at the right‐time, at the point of use.
The inflated view of the sensor‐based economy [141] is carefully [142] crafted [143] to create [144] new markets [145] and momentum [146] for sales [147] of sensors and data services [148] aimed to amplify the IoT [149] hype to fortify the deception game. It is promoting the desired effect by spawning mass hysteria and skillfully obfuscating the hard facts which then paves the ground for hordes of consultants to act as “trusted advisors” to make sense of this “revolution” which is supposedly going to change the future of work, life, and living. One glaring outcome of delusional [150] propaganda [151] is the near‐trillion dollar [48] waste related to investment in technology with a failed ROI. Trillions of sensors and devices that could connect to the Internet (basis for the cosmic scale of IoT) is due to the scale of unique identification [152] made possible by adopting a 128‐bit structure in the internet protocol. The unique address spaces in IPv6 [153] is 29 orders of magnitude higher than IPv4 if one compares [154] 4.3 × 109 address spaces for the 64‐bit IPv4 versus 3.4 × 1038 unique address spaces for the 128‐bit IPv6. New possibilities [155] and applications [156] may arise due to the flexibility of IPv6 to directly connect to the Internet (rather than sub‐nesting under/via gateway nodes).
The difference between promise and perils in deploying the concept of IoT as a design metaphor is rooted in grasping the difference between connectivity, discovery, and actionable insight. Just because something is connected does not mean value emerges, automatically, without a connected ecosystem. If a visitor’s tablet can discover the printer in an office and use it to print a meeting agenda, then we have extracted some type of value between the connectivity of the tablet and the printer, which were able to “discover” each other, and that discovery enabled the gain in efficiency (printing the agenda). In its basic form, this is an example of very simple data fusion which leads to an actionable output and provides “information” for the meeting attendees in terms of the printed agenda. Connecting trillions of entities to the Internet is futile unless discovery and data fusion enable semantically meaningful extraction of data to move up the DIKW [157] value chain [158] where data precede information, knowledge, and wisdom.
The PEAS paradigm resembles OODA (Figure 2.8) because “observations” refer to scanning (sensing) the environment and “orientation” informs the image of the environment by encapsulating both descriptive and predictive analytics (“decide” includes prescriptive analytics). The integration of data fusion and analytics with ABS (Agent‐Based System) is critical in the era of IoT. The networked society faces a deluge [159] of data, yet the human ability to deal with data, analytics, and synthesis of information may be inefficient. How can devices discover data and facilitate processes without intervention by humans? Automated on/off action taken by a domestic thermostat and HVAC based on temperature sensors may be quite primitive when considering autonomous objects in air (UAV), land, and water.
Raw sensor data unless discovered and combined with “perceptions” from the environment, may be context‐deprived and over/under utilized, which lowers the value of the data with respect to the desired goals. The perception from the environment is not unique but a “learning” task for the system. It may reuse the experience (learning), when relevant and appropriate, at a different instance (Figure 2.7). Can this “learning” become mobility‐enabled and “teach” other devices, for example, by transmitting a tutor virion to another computer or drive or system? Can this device communicate in natural language (NL) and/or respond/understand the semantics in human queries?
Taken together, unleashing the value of data may require coordination of ABS in every facet of our interaction with machines, objects, and processes which may benefit from feedback. ABS is an old [160] concept [161] but resistant to succinct definition [162] because agent activity must remain agile and adapt to the operating objective (PEAS) and problem context (OODA). Equation‐based models (EBM) create rigid, hard‐coded software. ABS design induces agility, may enable “drag & drop” variant configuration to adjust (on‐demand) to volatility, uncertainty, and ambiguity, inherent in most environments. In the context of democratization of data and benefits for the masses, agents can be highly personalized and “belong” to people, for example, personal agents, as discussed [163] elsewhere, with respect to cybersecurity. A similar modus operandi can be adopted for other use‐cases where data fusion [164] can be dynamic and composable (composed when necessary, depends on context) not only for use‐cases but also for individual user‐specific case/application (healthcare treatment plan) (Table 2.1).
The role of software agents to “discover” and then determine which data and/or data fusion may be meaningful or relevant (user‐specific), is an old idea, still waiting to be effectively applied. Connecting data must be contextual. The established contextual relationship must be discovered and “understood” by agents or group of agents. Another old idea is to pre‐establish the context based on KGs (Knowledge Graphs). The thinking that W3C standard resource description framework [166] (RDF) triples are the solution for KGs is incomplete. This myopia, is, in part, one reason why the semantic web [167] failed to flourish. The brilliant idea of representing subject‐predicate‐object (SPO) as a relational RDF graph is certainly useful and applicable in many instances but the approach bites the dust when the reductionist 1:1 granular relationship fails to represent reality. The latter is painfully obvious, especially in medicine and healthcare, where the rigidity of the RDF standard structure and RDF schema may be an anathema. The “force‐fitting” of RDF to healthcare applications [168] oversimplifies scenarios to the point where it may, inadvertently, introduce errors, simply due to exclusion, which may prove to be fatal.
Table 2.1 Generally, agents are computational entities (software) designed to perform specific tasks, autonomously.
Source: Based on Rafferty, E.R.S., et al. (2019). Seeking the optimal schedule for chickenpox vaccination in Canada: using an agent‐based model to explore the impact of dose timing, coverage and waning of immunity on disease outcomes. Vaccine (November 2019).
Agent characteristics | Definition |
---|---|
Autonomy | Operates without the direct intervention of humans or others |
Sociability | Interacts with other agents, that is, communicates with external environment such as sensors, fusion systems, and human operators |
Reactivity | Perceives its environment and responds in a timely fashion |
Pro‐activity | Exhibits goal‐directed behavior by taking the initiative |
Learnability | Learns from the environment over time to adjust knowledge and beliefs |
Mobility | Moves with code to a node where data reside |
Anthromorphicity | Externally behaves like human |
Agents embedded in devices (sensors) may have logic capabilities to perform artificial reasoning tasks (ART) and/or optimization [165] in multi‐agent systems (MAS).
One proposal suggests adapting [169] RDF by creating relationships between sets/subsets (rather than points and vertices as in classical SPO) using the set theoretic [170] approach. It is easy to grasp why “set” of symptoms and potential set of causes may make more sense in medicine and healthcare. The overlapping (Venn diagram) subset of relationships may be indicative of likely causes for symptoms. Generic symptoms, for example, fever, can be due to a plethora of causes and why a rigid 1:1 relationship in RDF could turn lethal in healthcare applications. The finer granularity of RDF is a disadvantage, yet it is key to merging attribute lists about an entity sourced from different data sources. The latter enables better search and discovery across diverse domains, the hallmark of globalization of enterprise systems.
An even older idea [171] which is recently [172] enjoying scientific [173] as well as public attention [174] is the labeled property graph (LPG). It is suitable for use‐cases which may be focused on providing stores for single applications and single organizations, such as, DSDM. LPG proponents are less committed to standardization, interoperability, and sharing. It is in contrast to the W3C ethos and RDF which favors standardization, interoperability, and sharing, which makes it useful in discovery using graph pattern searches. Optimization of local (domain specific) searches [175] using graph‐traversal algorithms [176] are better suited for property graph (PG) databases. KG networks embedded between sets/subsets may give rise to amorphous “linked” clouds, which could be industry [177] specific and may be domain specific [178] as well as user specific. Imagine if data from each patient could be used by an automated knowledge graph engine to create precision, patient‐specific, personalized KGs. Extracting relationships and contextualizing the relevance of symptoms may improve the accuracy of diagnosis. When viewing KGs in a population study (epidemiology), it may be easier to detect outlier events or cases that did not fit the expected patterns.
Therefore, DSDM may be represented as domain‐/user‐specific KG networks. Agents may be invaluable in working within this environment to discover relationships and contexts (specificity reduces search space), as well as discover data sources, and perhaps, based on embedded logic, decide whether the features or attributes calls for data fusion.
For any agent‐based approach to succeed, it is critical that the agent framework and standards are interoperable with the KG network and the data domains where the agent is searching. The opposing tendencies of RDF vs LPG in terms of standardization, interoperability, and sharing may limit agent‐mediated “cross‐investigation” of domains, discovery, and data. Therefore, it begs to question the expectation that one agent must perform in all domains. Perhaps, the success of agent search and discovery depends on semantically annotated structured data. The latter depends on ontological structure. W3C proposed [179] OWL standard web ontology language [180] and recent variations (VOWL [181]) may contribute to interoperability. The old idea of Internationalized Resource Identifier (IRI) as a complement to the Uniform Resource Identifier (URI) [182] to identify resources (to facilitate discovery) is a valid principle but yet to be adopted in practice. The plethora of old ideas (referred here) suggests that the value of these ideas may have to be revisited. We need new “blood” and new “eyes” to reimagine new ways to address interoperability. However, in reality, today, on top of this wobbly incompatible infrastructure, we are layering the “snake oil AI” and unleashing an incorrigible torrent of half‐truths.
Africa (>1 billion) and Asia (>4 billion), if combined, may soon represent 80% of the world’s population. Global corporations view this “80%” world as a “market” which promises new markets, new customers, and new wave of consumerism. It has little to do with lifting the lives of people. Discussion about the physics and mathematics of data, therefore, is a tempest in a tea cup. For ~7 billion people, the trials and tribulations of data and data analytics, we have discussed here, can be dismissed with an eye roll. It is useless for pressing daily applications for ~7 billion of the ~7.8 billion people in the world.
Thus far, what we have discussed, on one hand, may be an exploration of the tessellated facets in our search for meaning, and on the other hand, it is a discussion which may find parallels with the “six blind men and the elephant” syndrome [185] apparently divorced from complementarity [186] or synergy. It is as if the “commerce” from 20% of the global population, relevant or not, is thrust upon the remainder of the world market. In 80% of the cases for 80% of the global population, the daily decisions about FEWSH (bare necessities of life: food, energy, water, sanitation, healthcare) do not require AI, ML algorithms, or optimization of “state space” for hundreds of variables (Figures 2.9 and 2.10).
In 80% of the cases for 80% of the global population, the daily decisions about FEWSH require data, small data, data in near real time, and data that impact and enhance the user experience. In that context, the clamor about democratization of data is tantamount to chest‐thumping. The data in these use‐cases may be related to a subset of FEWSH (food, farm, agriculture, water, healthcare). If the tools are there to acquire these data, then the data are available. Therefore, is democratization really an issue? Is it a politically correct word that the 20% world prefers to use as a hand‐waving advocacy of problems that are divorced from reality on the ground? Is “democratization” a “theme” song for advocacy groups in OECD nations who are displaying the symptoms of the six blind men and elephant syndrome?
Table 2.2 offers a glimpse of one problem in healthcare. The tools to acquire the data are in short supply. Measuring the risk of osteoporosis is a prerequisite for prevention and treatment, if affordable. Arm‐chair “scenarios” of medical IoT will want to connect DEXA (DXA) scan data with sales of milk and exposure to sunlight as a “wellness” indicator. From the “double A” perspective, it may be a futile “power point” exercise because milk may not be available for the age group [200] generally at high risk of osteoporosis in the AA nations. In most parts of Africa and Asia, there is an opulence of sunlight.
Just because there is an “IoT” scenario, does not mean it is worthwhile or valid for users in “double A” nations. Just because there are data, does not mean there is information. Can we reduce incidence rates of osteoporosis simply by adding more DXA machines per capita? A recent (2013) study using seven national electronic healthcare records (EHR) databases revealed that Denmark (14.2 DXA units per million) showed age‐ and sex‐standardized incidence rates (IRs) of hip/femur fractures 2× higher than those observed in the United Kingdom (8.2 DXA units/million), Netherlands (10.7 DXA units/million), and Spain (8.4 DXA units/million), while Germany (21.1 DXA units/million) yielded IRs in the middle range (Table 2.3).
On one extreme we have presented sophisticated ideas for making sense of data. On the other hand, we doubt whether the toothless call for democratization of data from the affluent 20% of the world can help to lift the lives of people on the other side of the porous Pareto partition (80% of the world). It is not a true “Pareto” scenario, but the 80/20 nature of this problem evokes the Pareto principle as an analogy, hence, Pareto partition. The R&D outcome of the 20% may contribute certain elements to the 80% side. Thus, the “partition” is a metaphorical porous membrane, with bidirectional porosity.
But operations must be contextual. For example, is it necessary to deal with data and data models in this scholastic [203] manner (Figure 2.11) for all problems? The 80/20 global partition may be prominent in agriculture, healthcare, and energy. In case of the latter, what is the value of smart metering or load balancing algorithms when there is not enough energy, at an affordable cost, to supply the basic tenets of economic growth? How many farmers in “80%” world can afford to use drone‐on‐demand [208] systems? Why should people from the majority sector (80%) need useless marketing tools [209] when daily healthcare for the less fortunate can solve a myriad of problems with just‐in‐time little bits of data, for example, daily blood glucose level from a diabetic (versus the always‐on real‐time monitoring of blood glucose) or monitoring individuals for silent myocardial ischemia [210], a leading contributor to death. This is the debate where the economics of technology and its relevance to the community are crucial issues which may enable adoption or disable the dissemination of technology, which could have contributed to economic growth, workforce development, and sustainable job creation.
Table 2.2 Availability [193] of DEXA (dual energy X‐ray absorptiometry) scan machines to measure bone mineral density (BMD), a fair prognosticator for osteoporosis.
Source: Hernlund, E., et al. “Osteoporosis in the European Union: medical management, epidemiology and economic Burden: a report prepared in collaboration with the international osteoporosis foundation (IOF) and the European federation of pharmaceutical industry associations (EFPIA).” Archives of Osteoporosis, vol. 8, no. 1–2, 2013, p. 136. © 2013, Springer Nature.
The number and provision of central DXA units available in the EU27 (Data on reimbursement and waiting time [10]) | ||||
---|---|---|---|---|
DXA units/million | Waiting time (d) | Cost (€) | Reimbursement | |
Austria | 28.7 | 14 | 30 [11] | Yes |
Belgium | 53 | 14 | 34 [12] | Partial |
Bulgaria | 1.2 | 0 | 59 [9] | None |
Cyprus | 23.9 | 20 | 75 [9] | Yes (depending on income) |
Czech Republic | 5.2 | 40* | 32 [9] | Yes |
Denmark | 14.6 | 30 | 187 [13] | Yes |
Estonia | 8.9 | 14 | 14 [14] | Yes |
Finland | 16.8 | 1 | 146 [28] | Yes |
France | 29.1 | 14 | 41 [9] | Yes (conditional) |
Germany | 21.1 | 0 | 36 [9] | Yes |
Greece | 37.5 | 11* | 115 [9] | Yes |
Hungary | 6.0 | 15* | 7 [15] | Yes |
Ireland | 10.0 | 140*a | 99 [16] | Yes (conditional) |
Italy | 18.6 | 83* | 81 [9] | Yes (conditional) |
Latvia | 4.9 | 10* | 18 [29] | Yes |
Lithuania | 3.4 | 6* | 28 [30] | No |
Luxembourg | 2.0 | 30 | 59 [31] | Yes |
Malta | 9.7 | 105* | 184 [32] | Yes |
Netherlands | 10.7 | 14* | 84 [17] | Yes |
Poland | 4.3 | 1 | 10 [9] | Yes (conditional) |
Portugal | 26.9 | 8 | 5 [33] | Yes |
Romania | 2.4 | 7 | 5 [34] | Yes |
Slovakia | 10.7 | 18* | 32 [9] | Yes |
Slovenia | 27.1 | 11* | 29 [35] | Yes (conditional) |
Spain | 8.4 | 105* | 109 [9] | Yes |
Sweden | 10.0 | 60 | 152 [18] | Yes |
UK | 8.2 | 11* | 51 [36] | Yesa |
The European standard [194] is 11 (DEXA) DXA units/million. In an updated estimate, the poorest country in EU27 offers four machines/million [195] whereas Bulgaria’s neighbor Greece boasts 37.5 DXA units/million. In comparison, Indonesia [196] has only 0.13, India [197] 0.18, and Morocco [198] 0.6 DXA units/million. For the health of the people in these nations, how can democratization of data lower their risk of osteoporosis? Are we asking the correct questions? Are we pursuing the wrong reasons? Are we arm‐chair analysts helping the (BMD) medical device industry [199] accelerate their sales campaigns to AA nations? Can data provide relevant answers?
*, average of range; a, data; d, days.
Table 2.3 Plague of unethical profitability makes US pharmaceutical [201] business model in healthcare an abomination which is inappropriate for mimicry in any part of the world.
Source: Dr James Nolan [202].
Drug | Prescribed for | UK price | US price | Price rise |
---|---|---|---|---|
NEXIUM per 20 mg tablet | Acid reflux | £0.66 | £7.40 | 1120% |
ACTIMMUNE, 12 vials | Genetic diseases, osteopetrosis | £5,400 | £42,990 | 800% |
DARAPRIM per tab | HIV, cancer, malaria patients | £2.30 | £619 | 26,900% |
NASONEX, 50 mg | Nasal allergies | £7.68 | £224 | 2900% |
CINRYZE, 2 vials | HAE, genetic disorder | £1,336 | £3,645 | 200% |
HARVONI per tab | Hepatitis C virus | £464 | £928£270% | |
SOVALDI per 400 mg tab | Hep C in children under 12 | £416 | £855 | 200% |
DIAZEPAM, per tab | Anxiety, relaxation, muscle spasms | £0.02 | £3.05 | 15,200% |
OVEX, 100 mg tablet | Threadworm parasite | £2.54 | £300 | 11,800% |
LIPITOR, per 10 mg tab | Statin | £0.46 | £4.50 | 980% |
VIAGRA, per 25 mg tab | Male impotency | £4 | £61 | 1500% |
ZOCOR per 10 mg tab | Statin | £0.64 | £4.20 | 656% |
CYMBALTA per 30 mg capsule | Anti‐depressant | £0.80 | £9.48 | 1200% |
EPIPEN, 300 mg | Allergies | £52.90 | £523 | 1000% |
HUMALOG INSULIN | Diabetes | £16.61 | £215.30 | 1300% |
HIP REPLACEMENT OPERATION | £7,313 | £26k‐£37k | 350% | |
KNEE REPLACEMENT OPERATION | £6,315 | £24,801 | 390% | |
CATARACT OPERATION | £803 | £5,780 | 720% |
Worldwide, Africa [212] accounts for 9 out of every 10 child deaths due to malaria, 9 out of every 10 child deaths due to AIDS, and for half of the world’s child deaths due to diarrheal disease and pneumonia. More than one billion children are severely deprived of at least one of the essential goods and services they require to survive, grow, and develop [213] – these include nutrition, water, sanitation facilities, access to basic healthcare services, adequate shelter, education, and information. As a result, almost 9.2 million children under five die every year and 3.3 million babies are stillborn. Most of the 25,000 children under five that die each day are concentrated in the world’s poorest countries in sub‐Saharan Africa and South Asia. There, the child mortality rate is 29 times greater than in industrialized countries: 175 deaths per 1000 children compared with 6 per 1000 in industrialized countries.
These facts (paragraph above) and Figure 2.13 offer a vastly contrasting view to that of data tools and democratization of data as essential for lifting the lives of the people living on the majority side of the porous Pareto partition. Simple forms of small amount of data, sufficiently informing ordinary tasks, may be suitable for delivery of global public goods and services to the majority of the 80% world. It is absolutely ludicrous to think that “big” data, AI/ML, blockchain, or smarmy publicity [214] stunts may help, in this context. What we need is the concept [215] of “bit dribbling” perhaps coupled with PAPPU systems to help people improve their quality of life without the constant quest for charity.
Technology may play a central role to reach the billions who need services but not in the form of business [216] which is stable in the West and copied by the thoughtless Eastern schools, especially in India. Technical tools will generate data. The ability to use that data, judiciously, may be key to the value of data, for impoverished nations. Coupling social need with technical catalysts must be optimized in the context of the community and not according to Wired or MIT Tech Review or HBR. Advanced R&D is the bread and butter of progress, but the application of advanced tools must be contextual to the services that the community can sustain. Just because auto maker Koenigsegg claims the Agera model was built with a “less is more” philosophy does not mean it is a pragmatic standard of transportation suitable for Calcutta, India. In the realm of systems engineering courses and education, dynamic optimization (DO, Figure 2.14) illustrates a similar perspective. The principle is worth teaching, worldwide, but the practice must be relevant to the case. Do we all need DO in everyday life and living? Is it necessary for all types of edge analytics to process data using convolutional neural networks (CNN) on a mobile device or phone?
The conundrum of not applying the tools we think we have mastered is counterintuitive to the problem‐solving ethos in the 20% world. We are ever ready to use the latest and greatest gadgets from the bleeding edge to derive and drive the best possible perfection and performance. The quagmire of lies aside, we do have real tools which offer notable advantages. But, the volume of the 80% of the world and the economic handicap in these communities must be assimilated in order to change our thinking. The acronym KPI is for “performance” which is euphemistic for profit in the affluent world. It may not be in the best interest of the people. For 80% of the world, perhaps KPI should stand for “key people indicator” and ascertain whether a tool or the service improves the life of people.
Improving lives, however, is relative to the life you aim to improve, a life with disabilities [221] versus life with social [222] void are active domains in robotics. Robotics is useful but the robot propaganda, written mostly by hacks [223] and driven by media [224] sales, is a sign of the times. Essential robotics and robots for tasks that are dangerous, dirty, and dull (repetitive) are a welcome relief, for all involved.
The idea of the automated robotic factory was popularized by Philip K. Dick’s fiction “Autofac” published [225] in 1955 (Galaxy magazine). The “lights‐out” automated manufacturing facility FANUC [226] (factor automated numerical control) has been in operation since 2001, in Japan, but it is an exception. Even though “lights‐out” robotics made significant strides in heavy industry, it is far from the Orwellian scenarios promoted through chicanery [227] and buffoonery [228] by discombobulating the masses. The promise of robotics must be balanced with the degree of trust [229]in automated execution and the system must be acutely aware of the perils due to cyberthreats. Cybersecurity is quintessential for automation.
In other instances where human life is at risk (for example, transportation, manufacturing, and mining), the trust in automated action (robot) is as good as the planning for “what ifs” when the auto execution goes awry. But, that is a deterministic perspective where what could go wrong is anticipated, albeit with some degree of uncertainty. However, if a mobile robot crashes with a holonic manufacturing podium, it may generate a cascade of events where the outcome may be nondeterministic. The critical question in such a scenario is the extent to which a nondeterministic outcome can be tolerated and the acceptable cost of risk despite the “open‐ended” uncertainty. Few can even approach to answer this/these questions [230] because it verges on the domain of unknown unknowns.
But, that may not deter simulation aficionados from pursuing stochastic (what if) models to capture distribution of randomness in nondeterministic outcomes. Heuristics approaches may surface to suggest contingency measures. This is “video gaming” of automation [231] which could turn deadly in reality. The executive robot may be suited for “3D” tasks (dull, dirty, and dangerous) but unsuitable for relinquishing human oversight and control if lives are at risk. However, even worse are evil acts perpetrated by humans to bury [232] and ignore [233] the failure of automation, in the pursuit of profit.
Robotic tools in the 20% world are engaged in sophisticated activity which may be subjected to high oversight. In general, the 80% world is not a customer for such implementations in terms of mass consumption. Automation replacing or reorganizing jobs is not a new event (for example, auto industry) because technology [234] shifts the cycle of jobs and with it, the economy. Rapid changes in skill sets and the volatility of job categories influence other domains, namely, K‐16 education, training, skills development, (capital, labor market, and employment) and communication (hopefully, the truthful variety). The rate of change in certain ecosystems is dreadfully slow (for example, education system) whereas the evolution of the job market may resemble the rapid pace of bacterial growth, albeit slower than viral growth rates. The diffusion of robotics will take time and only if the building blocks of automation can be popularized, globally, in a manner that Lego blocks may have inspired young minds to compose, create, and construct.
The 80% world can benefit from robotics, for example, by reducing global disease [235] burden in emerging economies. In India, children are still used to clean sewers (flexible enough to reach cramped spaces, similar to chimney sweepers [236] in seventeenth‐eighteenth century UK). Can robotic tools replace the children? Robotics can improve lives and public health [237] rather than fear‐mongering and flagrant deception [238] how robots will replace human jobs. Similar to enantiomeric profiling of chiral drugs [239] and opioids excreted in sewer water [240], the post‐pandemic world must monitor wastewater for pathogens [241] as a surveillance strategy (prevention tool) or an early warning system prior to the onslaught of detectable clinical [242] symptoms in the general population.
The “partition” suggestion does not disguise the reality of the “ours” vs. “theirs” view of a divided world. It is unfortunate but necessary to serve as a constant thorn in our conscience and sow discomfort. The “partition” thinking originates from the corporate pursuit of developing a smörgåsbord of bleeding edge tools and then coerce 80% of the world to buy such products and services (“next billion users”). To add insult to injury, corporations from the 80% world are salivating to acquire rights to these products and bring it to their market (for example, Tata (TCS), TechMahindra, Wipro, Infosys, and other “body‐shops” in India [243]). There exists a nano‐cosm of people in the 80% world who could be a part of the 20% world. Because they are an influential minority and hold the financial power in the 80% world, they are aligning their astigmatic vision, greed, and “profit” objectives with the 20% world.
This mismatch may be at the heart of this global dilemma and creates the necessity to consider the porous Pareto partition in terms of people and service for the end‐user. People in AA nations are not buying facial recognition software systems. The abuse [244] is perpetrated by governments. People in the 80% world are not seeking quantum computing to process exabytes of data. People are seeking simple information, for example, for their health (data for blood cholesterol level) or from their farm (data about concentration of heavy metal contaminants in irrigation water used for fresh produce, such as, tomatoes). These services help people, the end‐user, the consumer. This discussion is about what science, engineering, and technology businesses can do for people where the KPI is user‐centricity and human‐centric [245] well‐being, a fact which is immensely clear in the post‐pandemic world.
This mismatch between the business to consumer (B2C) services versus the business to business (B2B) services is not new. The 80% world is always looking to the 20% world when planning strategic moves for climbing “up” the supply chain. The fact that the tools from the 80% of the world may not fit in the 20% world is obvious from “frugal innovation” calls [246] by others. Yet, imagination, invention [247], and innovation [248] from R&D in the 20% world are often helpful in the lives of the 80% world. What may be often lost is the translation of the advances from the 20% world, for people‐centric applications, in the 80% world. This discussion is not singing the praises about the investment in research that only the 20% of the world can afford to push forward because we know [249] the facts. The world is indebted for the strides made possible due to entrepreneurial innovation in such havens such as Massachusetts and California. This discussion is about exposing the lies [250] but not slowing the leaps of vision from the 20% world which may help create low‐cost [251] tools to serve the 80% world. The world needs R&D from the 20% world, hence, “porous” partition may facilitate the flow of innovation. Not “as is” but with contextual modifications to better serve communities in the 80% world, at a self‐sustaining cost (for example, the PAPPU model, pay‐a‐penny‐per‐use or pay‐a‐price‐per‐unit).
Mental health is one problem where “porosity” is most welcome because most of the world are affected by generic [252] as well as specific issues, which contributes to economic [253] drain. Inherited bipolar and unipolar disorders [254] do not discriminate on the basis of race, color, religion, or national origin. The neurochemical, neuroendocrine, and autonomic abnormalities associated with these disorders need biomedical research to elucidate the neurobiological basis of these diseases. The latter is not feasible for the 80% world. Harvesting data [255] from external symptoms and pattern analysis [256] may offer a low‐cost substitute, to inform the nature of treatment required. People in the 80% world may find it useful.
However, this discussion is not a to‐do list. It is not a roadmap. It may be a compass, oscillating asynchronously from esoteric thoughts to bare necessities. We are immersed in this duality. One cannot exist without the other. The role of the “partition” is to help focus on the issues that are unique to the environment and community that we wish to serve. It is not a partition of R&D or people or products but a partition for delivery of service. The idea of democratization of data is a bit buzzy but gimmicks are key to marketing. August institutions, including MIT, are complicit in sponsoring potentially puffy pieces to keep the hype [257] alive. But, the fact remains that enabling data to inform decision is a bedrock of measurement, central to all, irrespective of economic status. The porous Pareto partition is a catalyst to focus on services for the 80% world, where less could be more and serves our sense of égalité.
Dribbling bits of data to inform a person that her RR (Respiratory Rate) is fluctuating, too often, may be a preventative measure (think of the proverb a stitch in time saves nine). Informing the person that her RR data are not copasetic, may reduce future morbidity due to COPD (Chronic Obstructive Pulmonary Disease). Providing data and information may be without impact on the quality of life in the absence of follow‐up (clinic). In terms of data and information, alone, by enabling something simple and even mundane, the people‐centric application of technology and data, preferably at the edge [258] (point of use), may help to do more with less. Unbeknownst to us, we are attempting to use the pillars of science, engineering, data, information, and knowledge to build bridges which may serve as a platform to provide service to billions of users. Rather than gilding the lily, we are offering a “bare bones” bridge which serves a rudimentary purpose and still may exclude a few. The volume and demand for such low‐cost services [PAPPU services] may be, eventually, profitable for the business ecosystem.
Supporting a sustainable effort, to lift the quality of life, will depend on the extent of the product ecosystem and many other “things” in addition to technical and sensor data as well as the cohesion of the service supply chain. Socioeconomic data [259] and related factors are equally significant. Core elements are education of women [260] and trust [261] in women, followed by civic honesty [262], social value [263] as well as inculcating the practice of ethical profitability in social business and entrepreneurial innovation to accelerate the pace of creating pragmatic tools and solutions for remediable [264] injustices.
Neither data nor AI [265] is a panacea. Acquisition of data and analysis of data are not a guarantee that there is information in the data or that the information is actionable in terms of delivering value for the user, at an affordable cost. The COVID‐19 pandemic has made it clear that the global public goods that define “life‐blood” are food, energy, water, sanitation, and healthcare (FEWSH). The 80% world needs contextual, advanced, and affordable array of tools and technologies to leapfrog the conventional practices of FEWSH in the 20% world to vastly improve their crisis response systems.
In this context, energy is one rate‐limiting entity and in a “tie” with food and water, in terms of human existence and life. The “hand‐me‐downs” from the 20% world of energy may not be sustainable. Perhaps the Sahara Desert may be a source of energy for creating a global “battery” field, an idea [266] triggered by an 1877 [267] proposal, in a different context (it was, too, subjected to misrepresentation [268] and mockery [269]). Whether this is a “good” idea or not is not exclusively a matter of technological feasibility of implementation or transaction cost of service delivery. The question is, if it is good for the people. Global public goods are a matter of context for the community as well as the continent. Exploring the cleavage between entrepreneurial engineering innovation and complexities of social egalitarianism requires willingness to recognize, and adapt, among many different conceptions of a sense of the future. It may not be the future deemed appropriate by the 20% world experts. The future is asynchronous and nonlinear (Figures 2.15 and 2.16).
It may be a nonbinary future with multiple paths and unequal connectivity between amorphous nexus of networks representing nonlinear choices, aspirations, and outcomes. “Good” decisions are relative to that mix which defies definition yet works as a catalyst for economic rejuvenation. Even this type of “good” will (must) change with time and culture because no one version of good can fit all the world [271]. A binary outcome, with exceptions, must not be confused with binary decision‐making because a plethora of nonbinary factors can influence the outcome, which may appear as a binary output.
An oversimplified and cherubic example of the latter may resonate with residents of the Boston area. The choice between Mike’s [272] and Modern [273], famed confectioners located almost opposite each other on Hanover Street in Boston’s North End, is far from binary. The filling in the cannoli, taste of java, and the length of the queue are factored in the decision‐making process, which generally presents itself cloaked in a binary‐esque outcome. The choice masquerades a slew of nonbinary experiences.
Thus, creativity, imagination, and knowledge (from science, engineering, technology, medicine, and mathematics) may need to connect a few or many “dots” to inform solution development and delivery. It is true that “porosity” may contribute to solutions in the 80% world, and perhaps, less is more, but it will be remiss to leave the reader with the impression that invention/innovation may have to take a second place in the 80% world. In some cases, we must seek out‐of‐the‐world or counterintuitive ideas [274] and blend it with incisive insight which may be nontraditional. Far‐reaching convergence of bio, nano, info, and eco [275] is not an alternative but an imperative to stitch practical solutions, to satisfy, survive, and surpass the criteria dictated by the economics of technology and technology policy [276], which may be necessary to transform grand visions [270] into reality (and uncover new [277] tools, in the process).
But “grand visions” are often manufactured [278] as incremental mediocrity. Patents for using quantum [279] computing [280] are as absurd as the misuse of the term “cognitive” and the accompanying belligerence [281] in marketing. Those who throw around the term “cognition” may not have consulted a credible expert [282] or explored its meaning/definition {cognition [n] mental action or process of acquiring knowledge and understanding through thought, experience, and the senses}. Neither neurology nor modern computational neuroscience comprehends the combined electrochemical, cellular, and molecular nature of what it may mean to “acquire knowledge” by animals or humans. Any model, equation, algorithm, or hand‐waving “AI” is simply false [283] because it is far beyond our grasp, at this point, to claim anything more than a vague impression of what “acquire knowledge” may mean. The other key words in the definition (thought, experience) are at depths we do not even dare to know how to measure. Deciphering “processes” based on functional nuclear magnetic resonance imaging (fMRI “activation” maps of real‐time blood flow) is mindless drivel due to constraints in spatiotemporal resolution and limited ability of fMRI [284] to reliably detect functional activation. At the current state of instrumentation, resolution is inversely proportional to the ability to detect functional activation. Optimizing both is essential before fMRI data may be even considered precise.
The absurdity cryptic in the claims about cognition, learning, experience, and thought is neither colored by the author’s cognitive dissonance nor a figment of our uninformed imagination. Table 2.4 [285] captures the duration of so‐called “deep learning” training over “days” on a tensor processing unit [286] (TPU) scale with vast amounts of data (GB) which generated undifferentiated [287] rubbish. By comparing row 1 vs. 10 (bottom), the scores of the relevant match (#1, 0.892) between learning (saved query) vs. challenge (new query) is unimpressively different (#10, 0.765). According to Google BERT, after several days of “deep learning” “Blah blah blah blah” it was challenged with the query “Does this integrate with gmail?” which generated value of 0.765, suggesting 76.5% similarity between the two.
The fanfare of GPT‐3 [288] and ballyhooed context‐awareness of ELMo [289], BERT [290], and its cousin ALBERT [291], due to permutations and combinations of including masked language model (MLM) and next sentence prediction (NSP), is utterly devoid of intelligence. Except for fanatics feigning ignorance any rational observer may not be incorrect in thinking that throwing data (please see “Data” row in the upper part of Table 2.4) or using generic high volume of data for training (BERT claims to use “Wikipedia” in column 1, row 4) is ineffective. To be effective, these tools (ELMo, BERT, ALBERT, DILBERT, ROBERTA, and GPT‐3) must use training data relevant to the context of the target (search). Context‐awareness in the absence of data curation is as fake as claiming that a marble bowl is made of gold. Nevertheless, these advances in search techniques are immense strides [292] but the tools still are not “intelligent” but dumb as doorknobs. The doorknob does not turn unless one turns it or actuates it, manually or mechanically. Wikipedia as an experimental control is a plausible idea. Using curated data for training (ANN) may improve accuracy of search and better guide informed users to extract notions of connections and relationships with BERT‐esque tools as supplementary aids (Tshitoyan et al. [293] provides supporting evidence, Table 2.5). These tools are of limited value for non‐mission‐critical applications, for example, recommendations (movies, books, and restaurants), weather for entertainment (IBM’s Weather Channel), and fault‐tolerant uses (open garage door, on/off sprinklers). Nonessential human‐centric uses (congestion routing, temperature control, and voice message to email) may qualify if the outcome is almost correct in 80% of the cases. Actual use with humans‐in‐the‐loop (healthcare, emergency response, and security) may be scuppered if based on any credible risk versus reward analytics, except for offering nonbinding and non‐executable suggestions or alerts for human decision makers.
The paramount significance of curated contextual data in training any model (including artificial neural networks, ANNs) cannot be overemphasized. Individuals and institutions in possession of less than lofty ideals may revert to trickery in an attempt to sow doubt or discombobulate or disqualify the type of outcomes, for example, presented in Table 2.5. It is the age‐old deception due to over‐fitting [295] which can be also applied to ANN during training and the “fit” may be driven to precision using tools such as recursive feature addition [296] (RFA). For readers seeking a simpler analogy may wish to revisit what we discussed as the “force‐fitting” of RDF to healthcare applications (Ref. [168]). The erudition necessary to train ANN with curated data is not easily gleaned from a cursory review. Extensive perusal of scholastic research [297] begins to reveal the minutiae with respect to the nature of the domain‐specific data and the context of data curation (see sections 2.3 and 2.4 in Nandy et al., Ref. [297) that forms the bulk of the preparatory work based on rigor and strength of broad‐spectrum [298] knowledge. In an earlier section, we referred to domain‐specific models in a “macro” sense whereas the domain specificity of this example (Nandy et al.) is at the molecular (atomic and/or subatomic) scale.
Table 2.4 Google BERT thinks “blah blah blah” is 76.5% similar to “Does this integrate with gmail?”
Sources: Suleiman Khan (2019). BERT, RoBERTa, DistilBERT, XLNet — which one to use. © 2019, Medium; FloydHub Blog (2019). When not to choose the best NLP model.
BERT | RoBERTa | DistilBERT | XLNet | |||
---|---|---|---|---|---|---|
Size (millions) | Base: 110 Large: 340 | Base: 110 Large: 340 | Base: 66 | Base: ~110 Large: ~340 | ||
Training time | Base: 8 × V100 × 12 days*Large: 64 TPU Chips × 4 days (or 280 × V100 × 1 days*) | Large: 1024 × V100 × 1 day; 4–5 times more than BERT | Base: 8 × V100 × 3.5 days; four times less than BERT | Large: 512 TPU Chips × 2.5 days; five times more than BERT | ||
Performance | Outperforms state‐of‐the‐art in October 2018 | 2–20% improvement over BERT | 3% degradation from BERT | 2–15% improvement over BERT | ||
Data | 16 GB BERT data (Books Corpus + Wikipedia). 3.3 Billion words | 160 GB (16 GB BERT data + 144 GB additional) | 16 GB BERT data. 3.3 Billion words. | Base: 16 GB BERT data Large: 113 GB (16 GB BERT data + 97 GB additional). 33 Billion words. | ||
Method | BERT (Bidirectional transformer with MLM and NSP) | BERT without NSP** | BERT distillation | Bidirectional transformer with permutation‐based modeling | ||
Saved query | New query | BERT score | USE score | ELMO score | XLNet score | |
1 | How much will this cost? | Is this expensive? | 0.892 | 0.803 | 0.742 | 0.720 |
2 | Where is your data stored? | How secure is your product | 0.890 | 0.625 | 0.765 | 0.705 |
3 | What temperature is it today? | What time is the game on? | 0.880 | 0.677 | 0.746 | 0.730 |
4 | How do I change my password | I cannot find the settings page | 0.868 | 0.671 | 0.717 | 0.706 |
5 | Can I sign up for a free trial | Do I need a credit card to get started? | 0.868 | 0.736 | 0.753 | 0.759 |
6 | Where can I view my settings | Does this integrate with gmail? | 0.866 | 0.620 | 0.717 | 0.691 |
7 | I really do not like this product | I really like this product | 0.865 | 0.747 | 0.864 | 0.870 |
8 | What is the Capital of Ireland? | What time is the film in the cinema | 0.865 | 0.578 | 0.663 | 0.778 |
9 | Hello, is there anyone there? | What time is the game on? | 0.832 | 0.594 | 0.680 | 0.723 |
10 | Blah blah blah blah | Does this integrate with gmail? | 0.765 | 0.519 | 0.585 | 0.668 |
The laughable outcome is not at all surprising despite the hordes of brilliant scientists working to create tools (RNet, XLNet, ELMo, BERT, ALBERT, DILBERT, and ROBERTA) over the past 20 years because it is gnarly to capture semantics of language which has evolved over the past 200,000 years.
Table 2.5 (from Tshitoyan et al. Extended Data Table 2.4) – Analogy scores (%) for materials science versus “grammar” from different sources.
Source: Vahe Tshitoyan, John Dagdelen, Leigh Weston, Alexander Dunn, Ziqin Rong, Olga Kononova, Kristin A. Persson, Gerbrand Ceder and Anubhav Jain (2019) Unsupervised word embeddings capture latent knowledge from materials science literature. Nature571, 95–98. © 2019, Springer Nature.
Text corpus | Materials | Grammar | All | Corpus size |
---|---|---|---|---|
Wikipedia | 2.6 | 72.8 | 51.0 | 2.81B words |
Wikipedia elements | 2.7 | 72.1 | 41.4 | 1.08B words |
Wikipedia materials | 2.2 | 72.8 | 41.3 | 781M words |
All abstracts | 43.3 | 58.3 | 51.0 | 643M words |
Relevant abstracts | 48.9 | 54.9 | 52.0 | 290M words |
Pre‐trained model | 10.4 | 47.1 | 30.8 | 640k papers |
Training using Wikipedia for – metals – is grammatically rich (>72%) but content poor even when using select Wikipedia for materials (2.2% analogy). The smallest corpus (290 M words) used for training using continuous bag of words (CBOW) offers the best performance (48.9%) on materials‐related analogies when curated for “relevant” abstracts. The best performance for grammar may turn out to be profitable by enabling ELMo, BERT, ALBERT, ROBERTA, and DILBERT (DistilBERT) to be the voices of artificial trainers for the standardized twaddle marketed with impunity and known as Test of English as a Foreign Language (TOEFL). The contextual enrichment in this table is similar to the example of enrichment shown in Figure 2.17, suggesting a need for relevance and curation.
(Right) Data show large ANN error (−4.91 eV) with respect to DFT for a quintet [Mn(HNNH)6]3+ transition metal complex. The quintet [Mn(HNNH)6]3+ complex highest occupied molecular orbital (HOMO) level is underestimated by 4.9 eV, which is almost double the mean absolute error (MAE). This ANN was specifically trained using ΔEg data models on a set of 64 octahedral homoleptic complexes (OH64). The discrepancy (ANN error) is significant because frontier molecular orbital energetics provide essential insight into chemical reactivity and dictate optical and electronic properties. Small errors could make an immense difference in terms of chemistry of the transition metal complex. In this illustration, the metals are shown as spheres and coordinating atoms as sticks (C atoms, gray; N atoms, blue; H atoms, white). If your healthcare diagnosis and treatment was based on such an ANN outcome, would you trust, accept, and abide by the direction of the treatment suggested by such results? If this outcome is based on data from your electronic health records (EHR) which is known to be erroneous, would you trust poor data quality to inform a poorly performing ANN engine to design your healthcare?
The third piece of evidence that also dispels the marketing myths of AI in favor of viewing through the lens of artificial reasoning tools (ART, referring to ANN, CNN, RNN, DL, RL), is another variety of neural network [300] with credible capabilities. Message passing neural network [301] (MPNN) for molecules [302] is a tool [303] to unleash data [304] for human‐centric applications in health and medicine. This example centers on uncovering and repurposing a previously known molecule as an antibiotic [305] using a plethora of tools including MPNN and collectively referred to as deep learning (DL). Stokes et al. and the two other papers (Tshitoyan, Nandy) emphasize data curation and learning, without mentioning the term AI or “artificial intelligence” in the scientific papers. Unfortunately, the marketing and MIT news item [306], as expected, did not shy away from fake sensationalism to bolster the false appeal of AI.
The learning that generated the antibiotic (renamed Halicin) is nauseatingly detailed and the training (MPNN) was excruciatingly structured, optimized, and re‐optimized (using hyperparameter [307] optimization). The old idea of ensembling [308] was applied to improve outcomes in silico but predictions were biologically tested through rigorous experiments. Even after repeated steps to minimize errors, the authors remain cognizant of the pitfalls: “It is important to emphasize that machine learning is imperfect. Therefore, the success of deep neural network model‐guided antibiotic discovery rests heavily on coupling these approaches to appropriate experimental designs” (Stokes et al., page 698).
A curated set of 2335 molecules were used as the training set for new antibiotic molecules. The 2335 training dataset included a FDA library of 1760 molecules pre‐selected (curated) based on their ability to inhibit microbial (E. coli BW25113) growth. In other words, molecules with structure and function known to possess antimicrobial activity. Training MPNN with this dataset enables the neural network to learn the structures in order to select similar (or dissimilar) structures from a larger library of structures. The expectation is that when a “challenge” library is presented to the MPNN, the degree of similarity or dissimilarity, in terms of the output from the MPNN, can be tuned by modifying selection parameters. For example, using prediction scores (PS) to categorize molecules from a larger library (in this case, the ZINC database with ~1.5 billion molecules). By selecting higher PS value (>0.7, >0.8, >0.9), the outcome is “enriched” and a subset of molecules (in this case, 107,347,223, reductionism at work) is further subjected to other selection criteria, for example, nearest neighbor analysis (Tanimoto score). Finally, potential molecules (in this case, 23) are biologically screened (microbial assay) to identify the “new” antibiotic candidate(s). One such candidate is Halicin (Stokes et al.), previously identified as the c‐Jun N‐terminal kinase inhibitor SU3327 and rediscovered as a broad‐spectrum antibiotic, renamed Halicin, but still the same molecule as SU3227, albeit repurposed, based on function.
In combination, these three examples offer preliminary evidence that ART such as ANN, MPNN, DL, RL (reinforcement learning), etc., are excellent tools. However, ART and related tools are not intelligent, they do not self‐operate and the outcome is solely due to the skill and sophistication of the human operators. The steps must be designed with cautious intellectual strategy embracing the breadth of diverse knowledge, often dismissed by many institutions. Execution demands depth of erudition and incisive foresight to weigh the pros and cons of the criteria used to assess the quality of curated data prior to training neural networks with such data.
It is essential to learn the meaning of context in order to sufficiently inform the “artificial” part of ART. Models and patterns are like chicken playing tic‐tac‐toe [309] without context and semantics. Human knowledge to equip ART is almost impossible to transfer because we do not have a clue how to abstract continuous knowledge and use discrete processes to build it into an artificial system. Hence, ART may not “possess” an internal model of the external world. The immense variability in terms of features and which features may be relevant in which environment makes it difficult to model a state by claiming that feature selection will address all relevant and discrete contexts that the item or object may experience. Even if feature engineering was automated to levels of precision and continuity that could encapsulate all possible permutations and combinations of the behavior of an entity or object, the model may be still inadequate in the hands of different users due to inherent bias. It is not trivial but may not be impossible to model behavior and optimize for some features within a narrow cross section in a retail environment (for example, who may shop at Whole Foods, who may return to Andronico’s versus Mollie Stone’s).
If reason could inform common sense, then one may prefer ART over AI and champion the value of reasoning in ML techniques using neural networks. Neural networks and ML tools are amplifying, modifying, and regurgitating whatever humans have programmed into the tool. It cannot learn beyond the range of data or information provided, until humans decide to change, adjust, or add/subtract parameters/attributes which will influence the learning and the output from ART. The obstreperous zeal to move away from the misnomer of modern [310] AI [311] and adopt ART as a generic term may be a back to the future moment for rule‐based [312] expert systems [313] and principles [314] but coupled with new ML tools [315]. Marketing panders to creators of unstructured data but zettabytes of data anarchy occasionally may offer value irrespective of the clamor for general AI or ambient AI or intuitive AI or cognitive AI. Isn’t it possible to deliver value using ART?
Does the acronym matter? AI is a false trigger for technology transitions [316] but it is cheaper [317] and cheaper [318] to promote. It is profitable for conference organizers, narcissistic speakers, greedy social gurus, and other forms of eejits. Irresponsible computation may be draining the energy [319] economy, yet the marketing world is oblivious of the socioeconomic incongruencies in terms of thermodynamics [320] of computation, which is absent from daily discussions. Perpetrating the myth of intelligence in AI may be a moral anathema. ART lacks the cachet and panache, but promotes the rational idea of learning tools, which are, and will be, helpful for society.
The state of artificial learning is analogous to receiving a map of the world on a postage stamp and expect the bearer of the map (stamp) to arrive at 77 Massachusetts Avenue, Cambridge, MA, using that postage‐stamp‐sized map as the only guide. Neurologists shudder [321] at AI, the public are ignorant of the evidence of sham (Tables 2.4–2.6 and Figure 2.17) while marketing accelerates the “show” over substance. Sensationalism amplifies attention and siphon funds away from real‐world issues, making it harder for elements of FEWSH [322] to move forward, for the 80% world. The task ahead is to be creative, more than expected, and avoid the oxymoronic implementation of innovation as usual. Dynamic combinations [323] and cross‐pollination of counterintuitive connections may be worth exploring [324] to find many different ways to lift billions of boats, not just a few yachts. Future needs égalitarian resistance to our “default” state of society where we allow, acquiesce, and accept greed [325] as a price and penalty we pay for progress (Figures 2.18).
The tapestry of this discussion touches upon models of data, fake propaganda about AI, use of ART, and the lack of managed sanitation services for at least a billion people in the world ravaged by a pandemic. It presents a tortuously complex series of challenges each with its own bewildering breadth (Figures 2.20 and 2.21).
Table 2.6 Even after extensive training using precision data enriched for features using RFA, it is not surprising when gross errors are found in the outcome (analysis).
Source: Aditya Nandy, Chenru Duan, Jon Paul Janet, Stefan Gugler, and Heather J. Kulik (2018). “Strategies and software for machine learning accelerated discovery in transition metal chemistry.” Industrial and Engineering Chemistry Research, volume 57, number 42, pages 13973–13986. © 2018, American Chemical Society.
[Co(NH2CH3)6]3+ | [Mn(HNNH)6]3+ | |
---|---|---|
DFT | −20.00 eV | −18.64 eV |
ANN | −19.91 eV | −23.55 eV |
Figure 7 (from page 13981 in Nandy et al.) is one example how artificial neural networks (ANN) used in machine learning (ML) exercises and analytics generate erroneous results. ΔEg data (LEFT) shows ANN error (0.09 eV) with respect to DFT (density functional theory [299]) in a singlet [Co(NH2CH3)6]3+ transition metal complex.
More than ever, it is thorny to grasp how to balance efforts to continue creativity of thought considering the grim fact that more than 15 million people are infected by SARS‐CoV‐2 virus and deaths may soon exceed a million [340] people. The pandemic [341] may continue for another few years (2021–2025) and fluctuate in severity (acuity) due to antigenic drift [342] naturally caused by mutations. Lofty, esoteric ideas from any affluent [343] oasis poses moral and ethical dilemma when billions are facing a mirage in their effort to obtain the essentials for survival (FEWSH).
Yet, the pandemic has made molecular epidemiology the third most important job in the world (medical professionals and essential workers are first and second, respectively). Epidemiology is at the heart of public health data. Without metrics to measure performance, we may be forced to rely merely on anecdotes. The plural of anecdotes is not evidence.
We continue exploring the many facets of data but begin to ascend the pyramid how data may optimize the output or performance, an enormously complicated topic with vast gaps of knowledge.
The idea that tools must perform adequately is obvious and simple to grasp. Yet, performance as a scientific and engineering metric is a product of a fabric of dependencies and parameters which must be combined to deliver service, hopefully, useful for the community, for example, public health. In the digital domain, data is the common denominator, for example, data is the central force in epidemiology. The use of data to extract information must be a part of the discussion with respect to science in the service of society. These tools must be ubiquitous, accessible, composable, and modifiable, on demand.
Information depends on raw data but data may not, always, contain information. Even though data are driving the granularity, the curation of the contextual from the granular (data) to extract relevant meaningful information is essential for decision systems including digital transformation. A plethora of real and fake tools and technologies are peddled on this pilgrimage from data to information. On this route, one often finds the snake‐oil of AI [283] which is a fake tool and a marketing gimmick. But, under the bonnet of fake AI lies the very valuable techniques from ML and key statistical tools which are the “bread and butter” for many instances, applications, and information‐informed outcomes.
We return (please rereview Figures 2.7 and 2.8) to the systems world to revisit PEAS, a mnemonic borrowed from agent systems which consists of percepts (P), environment (E), actuators (A), and sensors (S). Performance is the informed outcome from the PEAS process which is a superset of SARA [344], which we mentioned in the beginning of this essay (see Section 2.1). Performance is an overarching goal of PEAS because solutions evolving from science and engineering in the service of society are not only “point solutions” or “tokens” but a fabric or tapestry, which must integrate, analyze, synthesize, and synergize data and information from various inter‐dependent domains to arrive at recommendations or actions that can deliver value to nonexpert users. Point solutions (SNAPS) are useful but “fabrics” may better suit complex cases. PEAS must combine data/information from subsystems (P, E, A, S) for data‐ (informed) DIDAS and in future, knowledge‐informed decision support (KIDS). Data may be inextricably linked to sub‐domains and outcomes may be influenced by interrelationships which are predominantly nonlinear. For specific instances we may need to select cross sections of data domains pertinent to specific applications. Once we determine which segments of data are required, the process of “discovery” is key to link the (contextual) data. Logic tools, agents, and ART (in combination) may extract the information. Metrics must evaluate the (semantically viable) outcome, if we wish to certify it as rigorous, reproducible, and credible for data and/or information‐informed decision support (DIDAS).
Digital transformation depends on making sense of these relationships and the granularity of nearly noise‐free data to support the claim that the relationships in question indeed have dependencies (as opposed to correlation without causation). The ability to discover and connect these interrelated data domains is challenging and yet is the heart of “knowledge” systems. Only those who do not dread the fatiguing climb of the steep path from data to information may reap the harvest of what it is to be information‐informed. Ascension to knowledge is difficult by orders of magnitude due to the necessity to discover the relationships/dependencies and then validate the connectivity with credible (curated) data which may be globally distributed in unstructured or structured (databases) forms.
One “connectivity” tool from graph theory is referred to as KG which is a connected graph of data and associated metadata. KG represents real‐world entities, facts, concepts, and events as well as relationships between them, yielding a comprehensive representation of relationships between data (healthcare data, financial data, and company data). The metadata about the data may catalyze feature integration, access to models, information assets, and data stores. Interoperability between KG tools is quintessential for data discovery and growth of knowledge networks. Although “knowledge” is the term of choice, the outcome of KGs are still far from knowledge but efficient in establishing connectivity between entities. The term “connectivity graph” is most appropriate. But in the data science parlance it is exaggerated to convey the impression that graph‐theoretic tools may approach the “knowledge” level in the DIKW pyramid. It does not. KGs are excellent tools for better connected DIDAS but not KIDS.
The cartoon in Figure 2.22 is a hypothetical example of how KGs and graph networks may serve as a backend for point of care (circle in square) applications, for example, a patient suffering from coronavirus infection who is treated by one or more medical professionals. The physician or nurse may create an ad hoc medical profile of the patient to prepare a treatment plan by combining data and/or information, for example, [1] stored data (medical history) from EHR about patient’s blood group [2], patient’s blood analysis results from pathology lab to determine antibodies (mAbs) in serum [3], available drugs (remdesivir, dexamethasone) for treatment, and [4] whether the ventilator (device) is available, if necessary. Accessing the strands of data in real time at the point of care (on a secure mobile device) is typical of medical decision support. The data/information at the point of care must be relevant to the context of the patient (age, height, weight, symptoms, existing conditions, and allergies). Relationships between data and their dependencies with respect to this specific patient must be “understood” by the data discovery process (the importance of semantic metadata) when the medical professional queries the system (technical metadata). Logic, limits, rates, flows, policies, exclusions, etc., which determines “where/when/how/if this then that” must be a part of this system. Data discovery may be a combination of standard data (for example, the range of values for normal blood count, such as, platelets, hemoglobin, etc.) as well as data discovered specifically for the patient’s treatment plan.
KGs are useful for relationship mapping and mining. KG maps are useful for discovery of data assets using agents, algorithms, and search engines in knowledge graph networks (KGN) which can be accessed and triggered by external queries. KG is just one element of digital transformation. It is not a panacea or a general solution. KGN requires pre‐created KG and generating KG requires deep understanding of the nodes that the graph will connect. Individuals creating KGs for domain‐specific use must know computational aspects as well domain knowledge. Introducing bias by connecting select domains may introduce errors and cause harm. On‐demand ad hoc tools to automate creation of KG and KGNs may suffer from semantic ambiguity of NL‐driven bots.
KG abstractions in the PEAS context suggests an overlap between two strategies, both grounded in a multifactorial approach to decision support based on links, relationships, and dependencies. The difference between cartoons in Figures 2.22 and 2.23 versus illustration in Figure 2.24 is one of domain specificity (healthcare, Figure 2.22 or generic, Figure 2.23) compared to the big picture of PEAS (Figure 2.24). In the context of domains, KGs are connecting sub‐domains and even granular data/information assets which lie far below the surface (for example, A, B, C, D, Figure 2.23).
Taken together, meaningful use of data and selected types of digital transformation tools are applicable to many domains which can help society. The financial instruments based on the paradigm of PAPPU may be key for ethical profitability if we wish to democratize the benefits and dividends from digital transformation for billions of users. Social business ideas to provide services to help reduce food waste and provide access to primary healthcare may use some version of PAPPU to collect micro‐revenue. Suggestions (see items 00, 01, 02, 03, and 04 [345]) about entrepreneurial innovation related to FEWSH may be essential as global public goods but they must work “in concert” to deliver the performance – survival in the post‐pandemic world and improving the quality of life. Performance as the outcome of the PEAS paradigm (Figure 2.24) works by connecting ideas of domains (percept, environment, actuate, and sensors) using KGs. Comparing Figures 2.21–2.24 may help to visualize the big picture as an abstraction of connectivity between PEAS and KG.
These system of systems and multilayer convergence suggested in Figure 2.25 embrace a dynamic broad‐spectrum of options and opportunities which may be beyond the grasp of end‐users, for decades. It is useful to sketch the “big picture” but implementation of these systems and extracting the synergies are beyond civilian reach. The complexity in the cartoon may be analogous to the problems faced by the “omics” tools in medicine and healthcare (metabolomics, genomics, and proteomics). Physiological changes prior to and during a disease state are almost never a binary outcome. The network view in Figure 2.25 is even more intensified in humans and animals because changes in physiology and metabolism are almost always a network effect and genome‐wide associations are only too common. Dissecting data to extract useful information may not be achieved by focusing on one or two aspects. The sluggish pace of progress in harvesting the value of genomics in precision medicine is one example. The herculean task of protein profiling [346] is still a work in progress. Hence, field applications may still need point solutions (SNAPS).
Digital transformation is a buzz word unless contextual data are meaningfully used to deliver information of value. Graph structures (KG) are an infrastructure of relationships to connect relevant granular data to improve total performance. The choice of tools from graph theory is based on their mathematical credibility. The combination of KG with PEAS may offer the rigor we seek for measurable progress of data‐informed digital transformation to help the 80% and the 20% world. We are still in the early stages and changes will accompany digital transformation (for at least another century) not because it is due to digital transformation but because change is a gift of periodicity. Focus, applications, and our demands change. Analyzing, predicting, and understanding these changes may be challenging. Tools in the process of change must evolve, adapt, and serve what is best for our society, industry, and commerce, at that time. With the progress of time, events evolve and unleash their influence on the global ecosystem.
If data cannot deliver performance, then the value of such data is negligible in a world where the “outcome economy” is no longer about the product but the value of the meaningful service (which may include the product). The trainer or sneaker is now a “service” where the consumer pays for duration of the service (life cycle) provided by the trainer or sneaker. It may seem tad obsequious but the service economy is here to stay, especially for the 80% world. The PAPPU model may unlock markets of billions who may have nonzero amounts of disposable income. Wealth of the poor may fuel ethical profitability for social business by tapping into markets of next billion users. Current shades of “western” business ethos, roles, and models are unsuitable as pillars for the future of service economies. Expect resistance from western business behemoths if they fail to control the 80% world.
Nevertheless, data is only one tiny element in the forthcoming global transformation, which will test our knowledge and our greed. Compassion without knowledge is ineffective. Knowledge without compassion is inhuman [349]. In our pursuit of life, we must think about our duties and responsibilities. In our pursuit of ideas, we must strive to think about the correct questions. Wrong questions will generate wrong answers [350]. Humanity needs compassionate dreamers [351] and an ethical sense of higher purpose guided by humility. Leaders with a higher moral fabric are few and far between. Humanity seeks those who can rise above their personal greed and narcissistic needs. To believe in greater good for the greatest number is that pursuit of “what life expects of me” [352] which often presents itself as a braided lifeline of chance, choice, and character.
Les savants des autres nations, à qui nous avons donné l'exemple, ont cru avec raison qu'ils écriraient encore mieux dans leur langue que dans la nôtre. L'Angleterre nous a donc imités; l'Allemagne, où le latin semblait s'être réfugié, commence insensiblement à en perdre l'usage; je ne doute pas qu'elle ne soit bientôt suivie par les Suédois, les Danois et les Russes. Ainsi, avant la fin du dix‐huitième siècle, un philosophe qui voudra s'instruire à fond des découvertes de ses prédécesseurs, sera contraint île charger sa mémoire de sept à huit langues différentes, et. après avoir consume à les apprendre le temps le plus précieux de sa vie, il mourra avant de commencer à s'instruire [353]. | The scholars of other nations, to whom we have provided an example, believed with reason that they would write even better in their language than in ours. England has thus imitated us; Germany, where Latin seems to have taken refuge, begins insensibly to lose the use of it: I do not doubt that it will soon be followed by the Swedes, the Danes, and the Russians. Thus, before the end of the eighteenth century, a philosopher who would like to instruct himself about his predecessor’s discoveries will be required to load his memory with 7–8 different languages; and after having consumed the most precious time of his life in acquiring them, he will die before having begun to instruct himself [354]. |
Critical review by Robert F Curl (https://www.nobelprize.org/prizes/chemistry/1996/curl/biographical/) and editorial comments by Alain Louchez (http://ipat.gatech.edu/people/alain‐louchez) are gratefully acknowledged. Unvarnished opinions and unconventional commentary by the corresponding author may have influenced the timbre of this essay. Some of the opinions of the corresponding author were supported and strengthened by Rodney Brooks (https://www.csail.mit.edu/person/rodney‐brooks) based on an earlier and related essay by the corresponding author (https://arxiv.org/abs/1610.07862).
Market proponents of AI are individuals who blather about neural networks of which they know little, to solve problems using learning tools which they know less, for the society of human beings of whom they know nothing. (Adapted from “Les médecins administrent des médicaments dont ils savent très peu, à des malades dont ils savent moins, pour guérir des maladies dont ils ne savent rien” – Voltaire [355]) (Doctors are men who prescribe medicines of which they know little, to cure diseases of which they know less, in human beings of whom they know nothing.) |
3.143.168.172