The Big Picture: the next century of computing

Home / Uncategorized / The Big Picture: the next century of computing

80 brief predictions for the future

BY:

Eric De Grasse
Chief Technology Officer

PROJECT COUNSEL MEDIA

 

5 October 2022 (San Francisco, California) – Charles Rosenbauer is a long-time mate, master programming language developer, complexity theory expert and a whiz at unraveling algorithmic dark matter, and tons of other stuff. And we like him especially because he’s a “Big Picture” kind of guy.

He recently gave us via his Substack 80 brief predictions on the future of computing and its impact on the broader world. You’ll not find a lot of these elsewhere, and they are certainly NOT an exhaustive list of the ideas percolating in that brilliant mood of his. However, based on much of his theoretical research (he does a lot of reading) and various trends he’s seen playing out, he has put together “places where I see things eventually deviating from common expectations”.

Many of these take the form of a niche that he sees existing now or in the future. And he is quick to add “how fast these predictions come true will be highly dependent on how fast people can find these niches and begin to fill them, but clearly – they are all underway”. They are brief but he will be fleshing out more lengthy explanations in future articles.


  1. Let us start by getting the obvious out of the way. Moore’s law is coming to an end. It is slowing down rather than coming to a grinding halt, but already Dennard scaling has broken down, which eliminates many of the real benefits from scaling further for chips that are not almost entirely memory.

  2. The end of Moore’s Law will quickly result in much more bizarre hardware. The decades ahead of us will be a Cambrian Explosion of bizarre hardware.

  3. Existing architectures will be abandoned. No more x86, ARM, or RISC-V. However, this will go far deeper than people expect. The basic concept of computing as a machine executing a stream of instructions, shuffling data back and forth between processor and memory, will eventually be abandoned in favor of more exotic models. The models we have today will be shown to be largely arbitrary, holding back potential efficiency gains and theoretical insights with models that reflect naive computational theory from the 1940s and 50s that has yet to die far more than any fundamental nature of computing.

  4. Legacy code will still be runnable through emulation. Native hardware support will be abandoned. Modern hardware already spends >99.9% of its complexity and energy on smoke and mirrors to provide the illusion of being an impossibly fast PDP-11. Fancy features like Out-of-Order execution, cache coherency, etc. will be scrapped and will be replaced by simpler, more efficient, more scalable hardware.

  5. General-purpose processors will get multiple orders of magnitude faster even after the end of Moore’s law, but only by exploiting a tradeoff between performance and familiarity.

  6. Hardware will begin to conform less to human models and more to physics. Computers are made of real atoms, use real energy, produce real heat, occupy real space, and take real time to send bits around to different locations in that space. These are things that cannot be abstracted away without paying serious costs, and without Moore’s law providing free efficiency subsidies by other means, real efficiency gains will come from deconstructing these old models and more deeply understanding the relationship between computation and fundamental physics.

  7. Much of the mandatory rethinking of hardware design and emulation-focused compatibility will allow old design mistakes to be corrected. For example, modern OSes are only millions of lines of code today because of arbitrary HW decisions in the 90s that produced a performance advantage for embedding driver code into the OS.

    It will become practical for ordinary people to write operating systems from scratch again.

  8. Silicon compilers will become common place. In response, innovative fabs will scale up wafer-sharing services. It will soon be possible to design a custom CPU or ASIC, upload the files to TSMC’s website, pay $500 and have a batch of 10 chips delivered to your door a couple months later. Hardware engineering will become almost as commonplace as software engineering.

  9. In the short term, the inevitable Javascript chip design frameworks will create massive security problems, baked into unchangeable silicon. Software engineers are not prepared for hardware engineering. The low costs and low volumes around shared wafer prototyping will mitigate problems somewhat, but eventually the powerful formal methods tooling that has already been used by hardware engineers for decades will be forced into the hands of a much wider audience.

  10. The computing industry will get over its irrational fear of Turing. Undecidability is deeply misunderstood; undecideable problems must be solved, and are regularly solved by any useful static analysis tool. Undecidable problems are not incomputable (that’s a separate computability class), but rather are the computational equivalent of irrational numbers; impossible to compute exactly, but rather easy to approximate. A golden age of code analysis and formal methods will follow once this misplaced fear is gone.

  11. As hardware becomes more physics-constrained, it will simultaneously become more bizarre but also more regular. One of the first forms this will take, a form already beginning to appear in some niche, high-performance computing paradigms, is the tiled architecture; large numbers of small, simple cores in a regular grid.

    While the inner workings of individual cores might become harder to immediately understand, computing will become far more geometrically intuitive.

    Tesla Dojo chip, a tiled architecture
    e5_epiphany-1024x569

    Adapteva’s Epiphany-V, an older tiled architecture that closely resembles Dojo, but lacked the resources and software support necessary to make it a successful product

  12. Computing on tiled architectures shares much in common with distributed systems programming, albeit with more predictability and on a much smaller scale. For physics-related reasons, computing on data is cheap, moving data around is relatively expensive.

    The operating system will begin to resemble a server taking up some number of cores on the chip, system calls being replaced by a simple packet-based protocol, rather than relying on context switches. Cores dedicated to specific programs or services will minimize the need to move large amounts of data around, maintaining effectively special-purpose caches.

  13. Far more powerful software diagnostic tools will be built. Debuggers that haven’t improved in any meaningful way since the 1980s will be abandoned for tools that can stream and visualize megabytes or gigabytes of data per second. Many applications will feature a dedicated diagnostics dashboard that future debuggers will be able to access.

  14. Garbage collection for large programs will become impractical, as it relies on many behaviors that will scale absurdly poorly with physics-constrained hardware. GC may still be practical on small scales, or for programs that make little use of hardware. Programming languages and other tools will provide better and more intuitive tools for memory management, creating a middle ground between manual and automatic memory management.

  15. Computer memory will generally make more use of memory arenas and similar techniques. This will not just provide performance advantages when freeing memory, but will also enable easier code analysis and better control of data locality (which will be one of the most important performance factors).

    Combined with the increasing relevance of the physical locality of data, the concept of “heap allocation” – mixing almost all data in a program into a large homogenous region – will give way to a paradigm where memory is divided into many dedicated regions with a defined spatial layout.

  16. More visual and intuitive tooling will make a far larger impact on the ability for normies to participate in software development than any VC-backed nocode project will ever come close to achieving. Not only this, but it will vastly improve the productivity of professional software developers.

  17. Computer science as a mathematical field will become less introverted and will begin to intermingle with other branches of mathematics on a far deeper level. It will also begin to more closely resemble traditional mathematical fields such as geometry and statistical physics. Computer science insights will then be turned around to produce deeper insights into those fields.

    A Rosetta stone between computing and many seemingly unrelated fields of mathematics and physics will be developed, creating countless revolutionary insights.

  18. Much of these insights will be seeded from attempts to reverse-engineer deep learning and the brain. The tools developed there will begin to influence tooling meant for more traditional software development, highlighting deep geometric relationships at the core of computation. This will radically change how we think about computer programming. Our existing programming paradigms and language designs are extremely unlikely to survive this transition, and computing on the other side will be almost unrecognizable to us today.

  19. This will blend nicely with physics-constrained hardware. There are close relationships between data storage, communication latency/bandwidth, and basic geometry. These relationships will become increasingly more relevant over time.

  20. Reverse-engineering efforts in machine learning and neuroscience will bear substantial fruit. Reverse-engineering efforts in other parts of biology will soon follow. In much the same way that physics has directly inspired an enormous deal of the most powerful modern mathematics, completely new computing paradigms will emerge from studying computational systems that were designed by gradient descent or natural selection as opposed to human engineering.

  21. Theoretical neuroscience will provide a new paradigm of error-tolerant computation. Granted, theoretical brain models with these properties have been known of for decades, but they will become more relevant in the future.

  22. The “Singularity” will fail to materialize. The vast power of human ingenuity will prove to not be purely the result of big brains, but rather the product of human language allowing us to stand on the shoulders of giants, giants whom themselves stand upon the shoulders of others. Yes, more powerful brains will be capable of some impressive feats, but will also be prone to making many arguments along the lines of “heavy objects fall faster than light objects” without constant real-world experimentation that is not bottlenecked by compute. Exposure to real-world entropy will become a more relevant bottleneck toward progress.

  23. Human-level AI will materialize, though largely from theoretical neuroscience research and not deep learning. The many subtle deviations from human behavior will create machines that, though intelligent, will have severe social deficits and be utterly incapable of functioning at all in society. It will be many decades before the subtleties of human behavior are understood well enough to replicate a functional human being and the interdependencies between humans will become a sharply more apparent factor.

  24. Neural implants such as those from Neuralink and Synchron will become available and have some interesting capabilities, and will be especially be useful in helping the disabled regain function. However, until major advances are made in software development tooling and theoretical computer science, the translation from human-style computing paradigms to the comparatively-alien computing inside the brain will dramatically slow any progress on more interesting functionality.

    Neuralink brain implant
  25. In the short term, multisensory computing and sensory substitution/augmentation will become a much more effective way of providing the benefits that brain implants have the ability to provide, easily possible without an implant. Humans have a strong desire for multisensory experiences.

  26. Sensory augmentation tools will be developed for software development. It will be possible to listen to sounds that communicate useful information about the structure and behavior of code.

  27. Sensory augmentation and substitution tools will be further developed and may be for a time a more popular alternative for disabled people to regain functionality than invasive implants.

  28. Various forms of non-text programming tools will be developed. Purely audio-based programming will be available for blind programmers as well as someone who may want to just code via a conversation with their computer while going for a walk. Software developers may cease to be purely desk-ridden. Haptic feedback will also likely be involved.

  29. Haptic feedback will become more common and more advanced in devices, similar to the HD Rumble in the Nintendo Switch.

  30. The current wave of VR will generally prove to be misguided. Optimizing for iMmErSiOn will prove to be nothing more than a meme that wastes billions of dollars. A good book is immersive. The real benefit of VR will be in significantly increasing the bandwidth of data that can be extracted from the brain and by increasing the sensory range of human-computer interaction. This lesson is likely to be one that must be learned the hard way, probably bankrupting a certain trillion-dollar tech company that is currently betting the farm on it.

  31. AR will prove to be much more useful than VR. We evolved brains to better navigate the world around us, not to escape into a fantasy world. Any country that sees any large-scale adoption of Metaverse-style tech will face both dystopia and prompt economic ruin. VR job simulators don’t put food on the table, and don’t run the fabs to make the chips that such a technology would be impossible without in the first place.

    AR on the other hand will augment many very complicacted and technical jobs and will eventually prove very economically valuable.

  32. Quantum computing will prove useful in certain domains, but will likely not break any real cryptography system. It is simply far too hard to minimize noise and errors in quantum systems, both from decoherence and from the analog nature of quantum systems.

    Instead, quantum computing will be most useful as a tool for simulating other quantum systems. This will revolutionize materials science.

  33. Weird emergent quantum properties will be discovered and will eventually make their way into computing devices. Exotic quantum sensors of all kinds will hit the market.

  34. The photolithography techniques behind modern chip manufacturing will begin to be used in many other technologies. Microfluidics, MEMS, etc., plus likely many other less obvious things. For example, the electrodes used by Neuralink are already being developed using these techniques.

  35. At some point, some solid-state quantum phenomena will be discovered that create compelling alternatives to transistors. They will not replace transistors, but rather will result in chips that combine transistors with other devices that can perform certain complex logic functions more efficiently.

    CMOS logic, which has been the standard technique for digital logic for decades, already uses a combination of multiple different types of transistors, as this produces enormous energy efficiency benefits even if it roughly doubles the complexity of circuits. Expanding the toolbox beyond just two types of transistors will complicate chip design even more, but may eventually prove to be a cost worth paying.

  36. Quantum computing will influence classical computing. While quantum algorithms may be hard to simulate on classical computers, they are still possible to simulate. Some will prove useful enough (at least on small scales) to either become common tools in their own right, or inspire new approaches to old problems.

  37. Reversible computing will become more common than quantum computing, perhaps even commonplace. This will be partly from RC being a subset of QC with simpler physical constraints (and thus easier to engineer/mass produce), but also from the fact that RC theoretically allows for vastly improved energy efficiency.

    RC does have its limitations though, and the full suite of tools used in general-purpose computing with it is not practical. It will instead serve as a form of accelerator for certain workloads.

  38. Heavier research into reversible computing, likely combined with all the other revolutions going on in the rest of computing, will eventually make their way into physics, likely vastly improving our understanding of phenomena such as entropy (which is very closely related to RC and its energy efficiency benefits).

  39. Classical computing will face the square-cube law; surface area scales faster than volume. Core count (and thus computing power) scales with volume, while memory bandwidth and I/O scale with surface area. Computers that perform less than 10,000 operations on a single byte fetched from external memory/storage/network/etc. will quickly become bottlenecked by I/O.

    Of course, thermal output also scales with surface area, but the constant factors regardless still favor computers where computing power vastly exceeds memory bandwidth; this is already a phenomena faced by modern hardware, and more exotic future hardware is likely to find more inefficiencies to clean up in compute than memory I/O.

    How Tall Can a Human Get?
  40. One result of this is that superlinear algorithms with time complexities such as O(n^2), O(n^3), and O(k^n) will become much more attractive. Many algorithms traditionally considered impractical to run will be opened up.

    Most notably, tools like Persistent Homology and Dimensionality Reduction will become ubiquitous, as they should be. New algorithms we can’t even dream of yet will be discovered as well.

    Persistent Homology can be thought of as the topological equivalent of a Fourier Transform; it takes a collection of points and extracts topological properites. This is proving to be one of the most interesting tools in modern data analysis and in reverse-engineering things like DNNs and the brain. It runs in roughly O(n^3) time.

  41. A wide range of new forms of dimensionality reduction will be developed focused on a variety of high-dimensional phenomena, not just locality. Dimensionality reduction algorithms will be combined with many other visualization techniques and become just one tool of many that can be composed into visualizations.

  42. Common programmers will be forced to actually learn how CPU caches and similar systems work (though the on-chip memory systems of the future will be very different from those found on modern chips). One of the biggest factors for performance and efficiency of code will be the precise geometric layout of data in memory; data takes up real space, and the further away data is the more time and energy it takes to access it.

    Optimizing code in this way by hand will not be practical, and more powerful code synthesis and optimization tooling will fill that role.

  43. A high compute-to-I/O ratio will also create an incentive to apply compression algorithms much more widely. You may get significant performance improvements in your code by compressing much of your data in RAM.

  44. SAT and SMT solvers will become extremely common tools. These tools are critical for any type of formal code analysis, but also are extremely general purpose for a vast number of domains on account of these problems being NP-complete.

    SAT, and by extension the entire P vs NP problem, is by far the most important open mathematical question for all of humanity right now. Nothing even comes close.

    I can’t say whether the P vs NP problem will be solved any time soon, but the truly bizarre properties of NP-complete problems are extremely deeply related to pretty much every problem known to modern mathematics, and are very poorly understood. Further research into this subject will accelerate over the coming decades due to their astonishing utility and wide applicability, but also due to the extremely memory-bound hardware of the future being perfectly suited toward the worst-case exponential complexity of such algorithms.

  45. Deeper research into P v NP will result in a far deeper and clearer understanding of exactly what is and is not possible in cryptography.

  46. Deeper research into P v NP will result in a far deeper and clearer understanding of exactly what is and is not possible in machine learning, and the nature of intelligence itself.

  47. Existing SAT solvers are vastly more efficient than they have any theoretical reason to be. Despite being theoretically an O(2^n) problem, many real-world SAT problems with hundreds of thousands or even millions of parameters are practical to solve while many much smaller problems with only hundreds of parameters would require billions of years to solve.

    However, exactly what constitutes a ”real world” problem currently is very poorly understood. This bizarrely better performance across a wide range of mysterious yet common cases suggests there are perhaps many computational shortcuts to problems that we don’t yet understand. A deeper understanding of the mathematics of SAT solvers will create a vast diversity of new computational techniques.

  48. Many bizarre properties of SAT solvers seem to strongly hint that there is something deeply, fundamentally wrong about our current understanding of computational complexity. Some deep insights are likely to shake up this understanding. I wouldn’t be surprised if we eventually drop big-O notation for a different, more powerful model of computational complexity.

  49. Due to the fact that computationally finding chemical equillibria is at a minimum NP-complete, it seems likely that a deeper understanding of SAT solvers will also revolutionize our understanding of biology and the computing that goes on inside cells, likely revealing that many of these bizarrely powerful computing techniques we currently fail to explain in SAT solvers are the bread and butter of biological systems.

  50. All of the bizarre new computing paradigms will create a feedback loop with tool development; a deeper understanding of computing will enable new types of tools, and new tools will make new paradigms of computing practical.

  51. The “black box” approach to machine learning, etc. will end. The profound laziness of modern ML that creates vast AI models with minimal effort into reverse-engineering them will be looked back on as a mistake.

  52. Superoptimization will become the dominant form of code optimization, fueled by advancements in theoretical CS that allow more intelligent navigation of such search spaces, combined with the previously-mentioned abundance of computing power relative to memory, making exponential searches much less scary to programmers.

  53. Software development tooling based on code synthesis, or at a minimum tool-assisted fine-tuning, will become common-place. It may become commonplace for code to include a block of precise constants generated by a tool.

  54. ML will make all non-cryptographic data easily forged. Better software tooling and a deeper understanding of P v NP will accelerate this to an extent we cannot currently imagine. Already gone are the days of “pics or it didn’t happen”.

  55. The only real defense against powerful and dangerous ML tech will be cryptography, which luckily, will also get profoundly more powerful.

    The rest of this article is behind a paywall. Past here is another 25 predictions regarding machine learning, crypto, industry, and broader social impacts of technology.

  56. Cryptographic primitives such as hashes, signatures, ZK proofs, and likely many things that have yet to be discovered will become much more common. Nearly every file shared on the internet will feature some cryptographic data.

  57. Something like the PGP Web of Trust will make a major comeback.

  58. Crypto will expand beyond finance and secure communication to wider social protocols.

  59. Better funding models for software will be developed, though largely from a very different approach to that taken in modern crypto. Advertising will become less viable, less necessary, and may eventually become much more tightly regulated due to concerns about manipulation.

    Tokenization will prove to be a mistake, and a broader range of techniques beyond “turn X into a financial asset” will be developed.

  60. Crypto will mostly move away from blockchains for scaling reasons, and will begin to more closely resemble a modern version of the khipu system used by Andean civilizations. Like in Andean civilizations, detailed tracking of supply chains will become a central function. Blockchains will still serve a valuable role, but localized consensus will generally prove much better in terms of scalability and privacy.

  61. Proof of Work will remain as the backbone of many cryptographic protocols. The energy usage will be seen as an advantage due to the problem of load balancing in the energy grid, which will become a more serious concern as renewables become a larger part of the energy grid.

  62. The unique cryptographic properties of Proof of Work will prove valuable, though applications will gradually move away from double-spend protection and toward more scalable features. One of the most useful applications of Proof of Work will turn out to be trustless timestamps and proof-of-existence protocols. These are easy to massively scale using Merkle trees, further mitigating energy usage concerns.

  63. Cryptographic protocols focused on providing plausible deniability as opposed to privacy will become more common.

  64. A theoretical Rosetta stone between cryptography and human social technologies (social tech in the way Samo Burja defines it) will be developed. Many common human social and cultural practices and will become viewed as analogous to cryptographic protocols. This may result in new cryptographic protocols, new social norms, and likely some new tools for social engineering.

    For example, the questioning of lies is closely related to interactive proof systems. Costly personal sacrifices pervasive in religions are the human equivalent of Proof of Work consensus.

  65. A deeper understanding of the brain through computational neuroscience and attempts to build brain-computer interfaces will result in a new, far deeper theory of human psychology.

  66. This theory of human psychology, combined with the aforementioned cryptographic-social tech Rosetta stone, will develop into new theories of human culture and social capital. A new age of social engineering, much wiser than previous attempts, will begin.

    Western society will likely need to learn to make a clearer distinction between race and culture, and to have much more respect for what can be learned from other cultures before this research can really take off though.

  67. Of course, when dealing with experiments on these kinds of scales, both tremendous benefits and tremendous harms can result. Cultural Engineering become an extremely controversial and politically charged field.

  68. The new theory of human psychology will have profound impacts on economics, which will be forced to give up on its “infinitely rational homo economicus” model of human decision-making. This, combined with new theories of social engineering and innovations in cryptographic computing will result in new paradigms of socioeconomics that make “capitalism”, “socialism”, “communism”, and similar models look like ideas from the stone age.

  69. Attempts to replace human social interaction with digital interactions (Metaverse, etc.) will fail both by being unpopular with the public and by being inferior to in-person interactions.

  70. Computational social technologies will instead trend toward augmenting human social networks rather than replacing them. Technologies aimed toward building social capital and new types of communities will be built and become very popular. These communities will not just improve people’s social lives, but will also rebuild much of lost civic and industry-related community, which will significantly improve the quality of politics and the efficiency of industry.

  71. A surplus in computing power makes many hard optimization problems tractable. The digitization of finance, supply chains, and social networks will allow for a sort of “homo economicus” digital assistant, advising many aspects of a person’s life with better decisions. Due to the asymmetry between computing resources and memory bandwidth, mass surveilance will remain a powerful tool for large organizations, but may get less effective and struggle to outcompete highly-optimized decisions from the masses.

  72. “Advertising” (or whatever replaces it) will become more symmetric and less manipulatory. Tools for allowing consumers to better communicate their demands will shape production.

    Information about supply chains and byproducts of production will become more accessible; industrial waste will become more commonly viewed as resources for other industries to make use of, creating a more circular economy.

  73. Digital assistants based on curating information and choices will generally be more effective than assistants that just dictate commands however, as there will always be tacit information available to the human that is not easily given to the machine.

    This may be somewhat mitigated by neural implants however.

  74. The shift toward increased feasibility of extreme optimization will impact a very large number of fields, making the late Christopher Alexander’s vision of a more generative approach to architecture and design a reality. This will result in both more beautiful and lively environments for humans, but also a more efficient use of resources and reduced environmental harm.

  75. A world more focused on local optimization will be one less focused on mass-production. Industrial capacity will stay large and likely even expand, but the focus will move away from producing vast numbers of standardized, identical items and toward producing vast varieties of goods and services tuned toward highly-specific, but different niche roles.

  76. More niche products will require improved markets and logistics, which will be made possible by extreme compuational optimization and cryptographically encoded supply chain information.

  77. Recycling will in many ways get more difficult, and the problem of repurposing old items will become more significant, likely resulting in recycling becoming a much larger portion of the economy. Many products may be made up of dismantled and recombined older products. The slowdown of Moore’s law however will have the benefit though of greatly expanding the lifespan of digital devices (or at least their components) before becoming obsolete.

  78. More niche production will also require more niche education and skills, which digital tools will accelerate. The preservation and propagation of niche expertise will become significantly more emphasized, and mass standardized education such as that seen in universities today will become significantly less valuable.

  79. A focus on more niche expertise will increase the leverage of workers, as people become much harder to replace. Quality of life will improve for many people. For situations where industries expand or decline, more intelligent and optimal ways of training and navigating people toward new jobs will be critical.

    Further, social technologies and norms will need to be developed to prevent abuse by workers who find themselves the linchpin of a large operation.

  80. Exactly which technologies and parts of society to standardize and which to allow to balkanize will become a major question and will become a highly central and political question.

 

Related Posts