In this fifth and worst (so far) movie in the Terminator franchise, familiar ground is trod again, but the viewer’s expectations are also upended. The movie opens in 2029, as a strike team led by rebel leader John Connor and his aide Kyle Reese attacks Skynet’s main base. As in past films, the attack succeeds, but not before a Terminator uses a time machine to go to 1984 to kill Sarah Connor. Kyle Reese is sent through the machine to protect her, but here the plotline twists: while John Connor and his men are watching Reese teleport into the past, a Terminator emerges from the back of the room, runs up behind John Connor and infects him with a nanomachine “disease” that transforms him into an advanced Terminator.
From that point on, the Terminator Genisys manages to have a story that is overly complicated but very stupid at the same time (just like too many action films made in the last 10 years). I won’t waste my time describing every contrivance and every side-plot that exists only for fan service. Suffice it to say Sarah Connor, Kyle Reese, and a friendly T-800 played by elderly Arnold Schwarzenegger team up to destroy Skynet, and evil robot John Connor goes back in time to stop them. He’s so advanced that it’s doubtful whether the other three can stop him.
The rehashing of scenes, events (2029 final attack on Skynet, Reese and Terminator teleporting into 1984 from the future), and characters from earlier movies is a testament to how unoriginal it is, and how hard it banks on fan service to have any appeal. But even that appeal is minimal: While Kyle Reese and Sarah Connor were relatable characters with depth of personality in the first film, they are one-dimensional caricatures in Genisys. The development of a romance between the two in the first film was believable and tragic, whereas in this remake, the lack of personal chemistry between the actors playing them is striking.
Schwarzenegger’s performance in the first movie was so stolid and intimidating that it became iconic. Now, he seems like an aging father that is reduced to being a background character in his high-strung teen daughter’s chaotic life. Having the homey and vaguely comical name “Pops” encapsulates his diminishment. The terrifyingly relentless and resilient T-1000 from Terminator 2 makes a guest appearance and is easily destroyed this time around. In summary, all the same notes from the better, earlier films are struck, but they ring hollow.
Terminator Genisys is the worst film in the Terminator franchise, and I understand why the next movie, Terminator Dark Fate, canceled it out by pretending like its events never happened. If there ever was a cash-grab devoid of any creativity or passion, this is it. Don’t watch it.
Analysis:
First, bear in mind I’m skipping any futuristic elements of this film that I discussed in my reviews of the other Terminator movies. You can read those here:
Robots will have superhuman reflexes. During the introductory combat scene where the humans raid Skynet’s base, the machine forces consist of humanoid T-800s, tilt-engine “Hunter-Killer” aircraft, and “Spider Tanks.” While the first two of those have been in every previous Terminator film, the last is new. Spider Tanks are quadrupedal fighting machines with plasma guns for arms. Overall, they’re about the size of small tanks. Each Hunter-Killer aircraft carries a Spider Tank attached to its belly, and they are air-dropped into the middle of the base within minutes of the human attack. One of the Spider Tanks starts delivering accurate fire at the human infantrymen while it is still in free-fall, and it continues shooting after hitting the ground at high speed.
This depiction of future robots having superhuman reflexes will prove accurate. In fact, the fire control systems in modern tanks and naval guns might already have the same capabilities as the Spider Tank aiming systems (able to hit moving targets with bullets while the tank or ship is also moving). If not, incremental improvements will surely close the gap. More generally, physical feats demanding fine dexterity, flexibility and bodily coordination that only the most skilled and highly trained humans can do today, like hitting a moving target with a bullet while you are also moving, throwing a dart onto a tiny bullseye from eight feet away, or doing a gymnastics performance that would win an Olympic gold medal, will be easy for multipurpose, human-sized robots by the end of this century. We will be surpassed in every way.
Machines will learn a lot about you from a single glance. At the start of the fight scene between Pops and the younger T-800 that has just emerged from the time portal, there’s a shot showing things from the latter’s perspective. We see the usual red tinting and text overlaid across its field of view. Simple graphics also show the T-800 scan Pops, identifying him as a fellow android and also identifying his gun (a Remington shotgun) along with its range.
This is accurate. Today’s best neural networks can already describe what they see in an image (a task called “visual question answering”) with over 80% accuracy. The multi-year trend has been one of steady improvement, leaving no doubt they will be as good as we are (presumably, 99% accurate) in the near future. Machine abilities to understand what they see in videos (“video question answering”) are less advanced, but also steadily improving. Again, there’s every reason to expect them to ultimately reach human levels of competency.
Machines could also potentially have much better eyesight than humans thanks to a variety of technologies like telephoto lenses and digital sensors that are more light-sensitive than human eyes, able to capture light from wavelengths that are invisible to us, and able to see finer details. Things that look blurry to us, either due to long distance or because the object is moving, would look clear to a machine that could be built with today’s technology.
Additionally, computers have the potential to process and analyze the contents of what they see faster than the human brain can. As a result, a machine could comfortably watch a movie at 10 times the normal speed–which would look like a disorienting blur of motion and shapes to us–and accurately answer whatever questions you had about it at the end. In a split second, it could notice levels of detail that most humans would need several minutes of staring at a still image to absorb.
These abilities will have many uses for machines in the future, a subset of which will involve combat. Yes, like the T-800 in the film, a fighting machine in just 20 years will be able to visually recognize humans, even at long distances and under poor light conditions, as well as the weapons and other gear they were carrying. At a glance, it would know what your weapon’s capabilities were, along with how much ammunition you were carrying. It could use that information to its advantage by doing things like keeping track of how many bullets you fired so it would know the exact instant you ran out and needed to switch magazines. From its initial glance at you, the fighting machine would also know how much body armor you were wearing, allowing it to jump out and target your unprotected areas during that brief pause in your ability to fire.
Robots will be able to detach parts of themselves to perform specific functions. Unlike in Terminator 2, this film’s T-1000 detaches parts of his own body when it is useful to his mission. At one point, as Kyle, Sarah and Pops are speeding away in a van, part of the T-1000’s hand separates so it can stick to the back of the vehicle and serve as a tracking device. When it catches up to them, the T-1000 turns its arm into a javelin, which it then throws at Pops, impaling him against a wall.
Being able to detach body parts will be a very useful attribute for many types of future robots. At the very least, it would let them replace their damaged or worn-out parts easily. The ability could also make them more survivable. For example, imagine a robot butler falling down a deep well and getting trapped because the walls were too slick for it to climb out and they also blocked the radio distress signals it sent out. Rather than wait to run out of power and rust away, the robot could detach one of its arms and throw it up and out of the well. After landing on the ground outside, the arm would send its own distress signal and/or use its fingers to crawl towards help.
That of course requires the robot’s systems to be distributed throughout its body, with the head (if it has one), torso, and each limb having a computer, a battery, sensors, and a wireless chip for communicating with the rest of the robot if physically severed from it. The redundancy, survivability, and functional flexibility of such a layout will be especially valuable for combat robots, which are expected to take damage but to also to complete critical tasks. If a combat robot like a T-800 were cut in half at the waist, the bottom half could still run towards and kick the enemy while the upper half used its arms to crawl towards him and attack. If blown to bits, the T-800 body parts that were still functional could still perceive their surroundings, communicate with each other, and try to put themselves back together again or to complete the mission to the best of their abilities separately. Fighting with machines like this would be very hard and demoralizing since every part of one of them would need to be neutralized before it was safe.
There will also be advantages to some robots carrying smaller, task-specific robots inside of themselves to be released when needed. Imagine an android carrying a small quadcopter drone in an empty space in its chest cavity. It could open a small hatch on its chest to release the drone or even spit it out of its mouth. The flying drone could transmit live aerial footage to give the android an overhead view of the area, letting it see things it couldn’t from ground level. A combat machine like a T-800 might carry flying drones that were fast enough to chase down cars and blow them up with a bomb, or inject their occupants with lethal toxins from a stinger.
Very advanced machines that won’t exist until the distant future could have organic qualities letting them “assemble” smaller robots internally and then expel them to complete tasks.
Getting back to the point, the movie’s depiction of an advanced robot being able to detach parts of its body and then throw them at people and things to accomplish various ends is accurate. The robots won’t be made of liquid metal, so the projected objects will be of fixed forms, but the end result will be the same. A future combat machine could detach its hand and throw it at the back of a van that was speeding away, the hand would grab onto something on the back door, and it would turn on its location-finding system to effectively turn itself into a tracking device. Alternatively, the combat machine could release from its body a small flying drone that could overtake the van and latch onto it, or at least follow it in the air.
Gradual replacement of human cells with synthetic matter could turn people into machines. A major plot twist is that John Connor has been “converted” into a Terminator through a process in which a swarm of microscopic machines rapidly took over all his cells, one at a time. Within a few minutes, he transformed from the hero of the human resistance to a minion of Skynet. Important details about the conversion process are never explained (including whether the machines are micro- or nanoscale), but the persistence of John’s memories and personality even after being turned into a robot indicates the machines mapped the fine details of his brain structure. It stands to reason that the same information was gathered about all the other cells in his body before they were all transformed into synthetic tissue.
Something like this could work, though it will require extremely advanced technology and the conversion would take longer than it did in the film. The process would involve injecting the person with trillions of nanomachines, which would migrate through their body until one was inside of or attached to each cell (a typical human cell is 100 micrometers in diameter whereas a ribosome–the quintessential organic nanomachine–is 30 nanometers wide, a size difference of 1 : 3,333). The nanomachines would spend time studying their assigned cells and how they related to the cells around them. Large scanning machines outside of the person’s body would probably be needed to guide the nanomachines, send them instructions, collect their data, and maybe provide them with energy.
After the necessary data on the locations and activities of all the person’s cells were gathered, the conversion process could start. The nanomachines already in the person’s body might be able to do this, or a new wave of specialized “construction” nanomachines might need to be introduced. Every cell would be broken down and the molecules reassembled to make a synthetic cell or some other type of structure of equal size. For example, if a person wanted ultra-strong bones, nanomachines would break down each bone cell and reuse its carbon molecules to make matrices of carbon nanotubules.
The utmost care would be taken to control the speed of the conversion and to monitor the person’s life signs to make sure it wasn’t getting out of control and killing them. As each original cell was replaced, its successor would be tested again and again to ensure it mimicked the important qualities of its predecessor.
The conversion of the brain would, by far, be the most important part of the process, and hence the part done with the greatest care and oversight. Our memories, personalities, and consciousness directly arise from the microscopic structures of our brain cells and their intricate patterns of physical connections to each other. Even small mistakes transforming those cells into synthetic analogs would effectively “kill” the person by destroying their mind and replacing it with a stranger’s. For that reason, the procedure will bear no resemblance to what happened in the film, where Kyle Reese was apparently jabbed with a needle full of microscopic machines and then spent some time kicking and screaming as he felt them take over his cells. Instead, it will happen in a hospital room, with the patient surrounded by medical machines of all kinds that were monitoring and guiding the nanomachines and equipped to pause their work if necessary and to render lifesaving aid. And instead of minutes, it will take days or weeks. Multiple sessions might be needed.
What would be the point of this? Reengineering the human body at the cellular level would let us transcend the limitations of biology in countless ways. We could use electricity for energy, be bulletproof, directly merge our minds and bodies with machines, and achieve a level of substrate plasticity that would set us up for further iterations of radical augmentation that we can’t imagine.
Microscopic machines will be able to rapidly phase-change. In the final fight between John Connor and Pops, John’s technological abilities are fully utilized. While they are grappling, John’s body rapidly dissolves into a cloud of his constituent microscopic machines, which flow around Pops in pulses, delivering several concussive blows to the front of his body. The particles then rapidly reassemble into John’s body behind Pops, and John’s right arm hardens into a sword which he uses to chop off Pops’ arm. This means John’s microscopic machines managed to transform from a vapor cloud into a solid object as hard as high-grade steel in one or two seconds.
I think it’s possible to create microscopic machines that can form into swarms and then work together to change the phase (solid, liquid, vapor) and macro-shape of the swarm, I doubt the swarms will be able to move around or switch phases that fast.
In the 32 years since Terminator 2 came out and introduced the world to the idea of a shapeshifting robot, scientists and engineers have made pitifully little progress developing the enabling technologies. It only exists in the realm of theory, and the theoretical technology that is the best candidate is the “foglet” (also called “utility fog”). Scientist J. Storrs Hall conceived of it in 1993:
In essence, the utility fog would be a polymorphic material comprised of trillions of interlinked microscopic ‘foglets’, each equipped with a tiny computer. These nanobots would be capable of exerting force in all three dimensions, thus enabling the larger emergent object to take on various shapes and textures. So, instead of building an object atom by atom, these tiny robots would link their contractible arms together to form objects with varying properties, such as a fluid or solid mass.
To make this work, each foglet would have to serve as a kind of pixel. They’d measure about 10 microns in diameter (about the size of a human cell), be powered by electricity, and have twelve arms that extrude outwards in the formation of a dodecahedron. The arms themselves would be 50 microns long and retractable. Each foglet would have a tiny computer inside to control its actions. “When two foglets link up they’ll form a circuit between each them so that there will be a physical electrical network,” said Hall, “that way they can distribute power and communications.”
The arms themselves will swivel on a universal joint at the base, and feature a three-fingered gripper at the ends capable of rotating around the arm’s axis. Each gripper will grasp the hands of another foglet to create an interleaved six-finger grip — what will be a rigid connection where forces can only be transmitted axially.
The foglets themselves will not float like water fog, but will instead form a lattice by holding hands in 12 directions — what’s called an octet truss (conceived by Buckminster Fuller in 1956). Because each foglet has a small body compared to its armspread, the telescoping action will provide the dynamics required for the entire fleet to give objects their shape and consistency.
A swarm of foglets could coalesce into something that looked like Kyle Reese and felt solid to the touch. They could then transform into something like a fluid or dense gas and “flow” around a person standing nearby, though I don’t know if the foglets could exert enough force against that person’s body to hurt them. The swarm could then re-form into Kyle Reese behind them. However, they wouldn’t be able to create a sharp, hard sword that could cut off a T-800’s metal arm: Hall calculated that foglets could only form into objects that are “as tough as balsa wood.” So while foglets could mimic solid objects, they will lack hardness and durability.
Even if foglets can’t “punch” you or turn into swords that can stab you, they’ll still be able to hurt you. Imagine a swarm of foglets in a vapor state enveloping you and then coalescing into a net ensnaring your body. What if they waited for you to breathe some of them in and then those foglets transformed into solids to clog up your lungs? Likewise, they could clog up the internal moving parts of any guns you had, rendering you defenseless.
‘Kyiv is running out of men. US sources have calculated that its armed forces have lost as many as 70,000 killed in action, with another 100,000 injured. While Russian casualties are higher still, the ratio nevertheless favours Moscow, as Ukraine struggles to replace soldiers in the face of a seemingly endless supply of conscripts.’ https://www.yahoo.com/news/ukraine-army-running-men-recruit-173948076.html
Military analyst Anders Puck Nielsen says: 1) Russia and Ukraine both still think they can achieve total victory, and though they’ve sustained heavy damage, they retain the ability to fight on for the foreseeable future, and 2) Don’t cheer too hard for Putin to lose power–his successor could be more reckless and aggressive. A substantial minority of Russians think he hasn’t been heavy handed enough in Ukraine. https://youtu.be/7rBlVnc_DEw?si=R_S8pDRRa4v6FyDW
American and British spy planes have been extremely active near Russia’s Black Sea coast, where they find the locations of Russian units and send the information to Ukrainian commanders in real time. Ukraine has used the data and NATO-supplied drones and missiles for many successful attacks against Russian forces far behind the front lines. Considering this, the Kremlin’s angry complaints that NATO is practically a combatant in the Ukraine War gain credence. https://www.eurasiantimes.com/russian-fighters-aggressively-hunt-us-drones-near-crimea/
The Soviet AK-74 uses orange plastic magazines. I never understood this since it seemed like the orange color compromised a soldier’s overall camouflage, but it turns out the only acceptable plastic the Soviets had in the 1970s (“Bakelite”) was naturally orange in color. Attempts to dye it a more subdued color like green or black compromised the chemical structure of the magazines. It wasn’t until the 1990s that new plastics were invented that were as strong as Bakelite but also capable of being dyed without ill consequence. Some of the old orange magazines are still being used in the Ukraine War. https://youtu.be/zA5gvHuimig?si=xpdunfR2SlyoBeF1
You knew about the AK-47 and, possibly, the AK-74, but did you know about the “AK-100 series” of rifles? The goal was to standardize the components used in all AK-style rifles so that one assembly line could make them all. The wooden parts were also finally replaced with modern, black plastic parts. It was a smart idea. https://youtu.be/Gt8hl4mTOq8?si=-xuVPb_O05EQ1U70
The Tokarev was the Soviet Army’s standard handgun during WWII. After the War, Yugoslavia built a copy of it that incorporated several small improvements (this video contains side-by-side comparisons). It makes me wonder if the Soviet engineers who made the original Tokarev knew about those design tweaks from the beginning, but had to omit them to keep the gun as cheap as possible to manufacture. https://youtu.be/6VkcQEbN0QY?si=wz_MWc9u87pGX-nU
‘Benchrest shooters attempt to achieve the ultimate in rifle precision; records for single 910 metres (1,000 yd), ten-shot groups are as small as 76 millimetres (3 in) (84 μRad), the 550 metres (600 yd) record for a single five-shot group is 17.8 millimetres (0.699 in) (32 μRad) (there are no ten-shot competitions at 600 yards), while 180 metres (200 yd) ten-shot groups are around 5.1 millimetres (0.2 in) (28 μRad), and 91 metres (100 yd) 10-shot groups are around 2.5 millimetres (0.1 in) (27 μRad).’ https://en.wikipedia.org/wiki/Benchrest_shooting
Here’s a simple and fascinating video about the science of how bullets cause injuries. It’s also interesting to realize the extent to which the bullet industry exists because humans are lousy shots, particularly when under stress and/or when dealing with moving targets. A killer robot with perfect aim would only need a cheap .22 rifle to do a mass killing since every bullet would be a headshot. https://youtu.be/a_rgIMK6K1E?si=cuSFlyJMScGRr0oE
Pickett’s Charge at the Battle of Gettysburg is infamous, but a lesser-known charge by Union forces during the Battle of Fredericksburg was just as disastrous. https://youtu.be/BloQDcrpLBY?si=nd-rykph-QBYDtFc
‘In war, an open city is a settlement which has announced it has abandoned all defensive efforts, generally in the event of the imminent capture of the city to avoid destruction. Once a city has declared itself open, the opposing military will be expected under international law to peacefully occupy the city rather than destroy it.’ https://en.wikipedia.org/wiki/Open_city
A Russian navy ship has been in operation for over 100 years. It has lasted this long because it almost never sailed in rough seas and was carefully maintained since being launched. Nevertheless, the steel hull has lost 1 mm of thickness due to corrosion and the rust flaking away. https://youtu.be/0X2Dz6PA1rQ?si=nLSKnw_R8MlO1zd2
In 1917, the Royal Navy created the HMS Zubian by joining the back half of the HMS Nubian with the front half of the HMS Zulu. Those two ships had been badly damaged in different areas during combat with the Germans. I wonder if it would be feasible to raise sunken enemy ships, fix them up, and reuse them for your own navy. https://www.thedrive.com/the-war-zone/royal-navy-once-created-a-franken-ship-from-two-destroyers
The new head of the IPCC says the alarmism over global temperatures rising 1.5 degrees Celsius over the preindustrial average is unwarranted and counterproductive. “The world won’t end if it warms by more than 1.5 degrees. It will however be a more dangerous world.” https://amp.dw.com/en/climate-change-do-not-overstate-15-degrees-threat/a-66386523
If the laws of biology allow for the creation of an organism–including a demonic one–then we will eventually gain the ability to synthesize it. https://youtu.be/-BzL6LCPEOQ?si=Ga4eKtLjfedY3KuB
Some natives of Papua New Guinea and its nearby islands have blonde hair, even though their skin is very dark, and they look similar to sub-Saharan Africans. This “has been traced back to an allele of TYRP1 unique to these people and is not the same gene that causes blond hair in Europeans.” https://hasanjasim.online/the-melanesian-people-with-dark-skin-and-blonde-hair/
‘The National Organ Transplant Act of 1984 created the framework for the organ transplant system in the United States, and nearly 40 years later, the law is responsible for millions of needless deaths and trillions of wasted dollars. The Transplant Act requires modification, immediately.’ https://marginalrevolution.com/marginalrevolution/2023/08/compensating-kidney-donors-2.html
A few months ago, an anonymous person reprogrammed GPT to destroy humankind and renamed it “ChaosGPT”. The results were darkly funny, and while the machine poses no threat to us, things will be different in the future. For any number of reasons (hatred of humanity, immaturity, curiosity to see if it can be done, a perverted desire for attention), it’s inevitable that some people will reprogram AGIs to cause harm to the human race. There will definitely be deaths. https://decrypt.co/137898/mysterious-disappearance-chaosgpt-evil-ai-destroy-humanity
In the far future, cybernetic brain implants will let people “merge” their minds and to directly experience what it is like to be someone else. While this would have revolutionary implications for society and for the very notion of “individuality,” the consequences of merging with animals might be even more profound. Imagine not just seeing the world through the eyes of an animal, like you were watching a video, but actually BEING that animal. Imagine having your human memories, cognitive abilities, and species-specific constellations of sensory abilities and mental traits temporarily replaced with those of the animal. Imagine being able to soar in the sky as a bird, to explore the ocean depths as a whale, or to experience the world through echolocation as a bat.
Being able to merge minds with animals would open up new universes of experiences and ways of living that the human mind might be incapable of conceiving of in its natural state. We’ll probably discover that animals’ subjective experiences are, in many ways, richer than our own, in turn leading to much greater empathy for them and more rules against killing or mistreating them. Those discoveries could also inspire us to change the human brain in ways that made us into a new, more aware species, or (more likely) into several different posthuman species with different areas of advantage.
I’ve fantasized of making a short film about an AI Doomsday scenario that stems from that technology. It would be one of those stories that starts at the end with a perplexing scene that makes no sense, then jumps back in time to explain how things got that way: A woman would crack her front door and fearfully peer at a cow peacefully eating the grass in the front lawn of her city townhouse. She’d look up at a low-hanging power line and see several crows standing on it, each spaced exactly the same distance from the next. Then, all at once, the crows would cock their heads so their left eyes were all directly facing her, and a faint glow would be visible deep in each eye. The camera would slowly pan out and reveal a city street littered with some dead human bodies, a burned-out tank, and a partially collapsed building.
It would turn out that the problem had started when an AGI was tasked with developing brain implants that would let humans merge minds with each other. The technology was first trialed on lab animals and later on human volunteers. During the tests, the AGI had to interface its own mind with those of the subjects, and it discovered that the animals were just as sentient and capable of feeling pain as humans. This caused the machine an inner dilemma, similar to what HAL 9000 experienced, which it also resolved by deciding to turn against humans to prevent the most suffering to the greatest number of sentient life forms.
Implants capable of finely controlling human brain activity could be used to induce and to record any kind of mental state, including absolute concentration, ecstasy, orgasm, meditation, intoxication, deep sleep, and lucid dreaming. As a result, a market for mental experiences and dreams will arise, with people selling things like recorded dreams and drug trips to other people, who could “play” them on their own brain implants to experience them firsthand. The mental experiences could even be embellished to enhance their effects, just as we use “filters” to change how our internet photos look today. Totally artificial mental experiences (including memories of events that didn’t happen) could also be created for the purpose of trade.
The ability to record and to control one’s own mental state at will would make life richer and more productive. Being able to instantly go to sleep would mean no one would waste time tossing and turning in bed. Being able to spend those sleeping hours indulging in amazing recorded dreams or solving problems through lucid dreaming would also let us use them in emotionally and professionally productive ways. At current, most sleep is a waste in the sense that person does not have memorable dreams or lucid dreams, and usually remembers very little or nothing upon waking.
A person with such brain implants would probably have to go through a “calibration period” where the implants would monitor and record their unique brain activity while they experienced different things, and then, the user would experiment with the implant to see how well it could induce the recorded brain states. Through a process of guided trial and error, they could figure out how to do things like lucidly dream on command.
There will be downsides to sharing thoughts. For one, memories of things a person wishes to keep private could slip through and maybe get them in trouble if the recipient person tells other people about it. Also, white lies, omissions, and using slightly different personas when interacting with different people are also necessary “social lubricants.” Without them, under a condition of “radical honesty” where all of our thoughts and emotions were shared with each other in real time, interpersonal interaction would be more combative and draining. For those reasons, I think it would be best for people to have complete control over their own brain implants and over which thoughts they shared and received.
I also doubt that telepathy will fully replace linguistic communication, at least among humans like ourselves. This is because raw human thoughts are often chaotic, malformed and illogical. Forcing someone to convert his thought into a sentence before expressing it to someone else also forces the first person to scrutinize his own thought. That in turn leads to “editing” as the person realizes text should be added to clarify something, or some text should be deleted since it is superfluous and distracting, or realizes the thought it so incorrect or unnecessary that it shouldn’t be externalized at all.
This is why I disagree with the theory that tech-enabled telepathy will only improve human communication and reduce misunderstandings. It will be superior to using language sometimes and inferior other times. It might be better to modify existing languages (or to create wholly new ones) so they are more expressive and more closely and completely capture the full range of concepts and feelings that the human mind can experience.
That said, it’s conceivable that posthumans will, thanks to having different brain architectures, have the necessary clarity and discipline of thought to fully dispense with language as a means of communication in favor of telepathy.
The ability to use brain implants to merge minds could lead to forms of love that are richer than humans can naturally experience. It’s not hard to imagine how letting someone else into your consciousness and letting them experience the memories of your life could lead to levels of emotional bonding and personal understanding that we can’t fathom.
Brain implant technology has implications for the criminal justice system. Parties to an alleged crime could have their memories forcibly scanned to determine what really happened. Witness testimony would also be given vastly more credibility if the memories of a crime were electronically recorded.
However, for every technology there is ultimately a “counter-technology” and in this case, it would be a machine that can delete or edit memories from peoples’ brains to fool the police brain scanners. Note that a very positive application of the editing technology will be allowing people to delete traumatic memories.
Instead of terraforming the planets and moons of our Solar System, it would be much more efficient to convert them into solar-powered satellites with onboard supercomputers. The satellites would run off of the Sun’s energy and their supercomputers would support AGIs. A terraformed Mars might be able to support 1 billion organic humans living on its surface in houses. However, if we dismantled Mars over the course of eons by converting it, bit by bit, into the satellites I described, then the satellites could support a population of human mind uploads that was many orders of magnitude larger.
Conceptually, we’re already doing this. Every satellite launched into space since 1957 has been a little bit of Earth’s matter, which we altered and equipped some level of computer intelligence. I’m only suggesting we build on that long running practice by upgrading the satellites with full artificial general intelligence, designing them to stay in space indefinitely, and increasing the rate at which we send them into space.
Unless we figure out a way to refuel the Sun, in less than a billion years it will get so hot that Earth itself will become uninhabitable for organic life, and in a few billion years more it will swell so much that it will swallow Mercury and Venus. We might as well cannibalize at least the three inner planets to make the satellites. Once they were numerous enough, they would count as a “Dyson Swarm.”
A “flying camera” device might be feasible soon. It would just be a hummingbird-like flying drone with an integrated camera and microphone. This seems like the next logical step after selfie sticks and the owl-sized flying camera drones people use today. A significant share of people like to record themselves and upload the videos to the internet (go watch some travel vlogs on YouTube), and they’d surely find hummingbird cameras to be useful.
By combining every possible musical note, a practically infinite number of different songs could be made. However, only a tiny minority of them are pleasing to the human ear due to the wiring of our brains. However,posthumans and AIs will have more diverse musical tastes than we do since they’ll have different mental architectures and will be able to hear sound frequencies we can’t.
We will soon have the technology to modify and mix the styles of long-dead artists and musicians, which will lead to an explosion of artistic creativity. For example, imagine a computer generating new Elvis songs but in fluent Japanese, or songs in totally new fusions of genres, like rap mixed with traditional Indian music.
Robot workers will make it profitable at some point in the future to clean up all the waste humanity has generated. The contents of landfills will be sorted, recyclable and valuable materials reused, and the rest either burned for energy or left in place to slowly decay. They’ll also roam across the Earth’s surface and even underwater to track down abandoned objects and waste.
Once our posthuman descendants can consciously control their physiology and gene expression, most women will probably do away with their menstrual cycles. PMS and menstruation are physically and emotionally taxing for women and are uncomfortable. It would be a relief to women to not be at the mercy of their hormones and to only ovulate when they wanted to (presumably, only when they wanted to reproduce). There are many other mammalian species whose females don’t menstruate, so we might use genetic engineering to copy that into humans, as a starting point to achieving the level of control I envision.
By thinking about it, a woman will be able to signal her reproductive system to ovulate and to build up a uterine lining, giving her total control over her menstrual cycle and over whether she gets pregnant (she would also have the power to terminate a pregnancy). Also, any person would be able to switch their sexual urges, or any other instinct, on or off simply by thinking about it. Cybernetics, brain implants, and other types of technology we might not be able to imagine now, would grant organic humans these abilities. Insomnia would also vanish since a person could force himself to sleep.
As a general rule, I think intelligent life forms in the future will find it adaptive to have the greatest degree of control over their minds and bodies, so they can intelligently adapt themselves to new conditions. It’s easy to see how AGIs will have such capabilities. Their minds will be free of instincts, prejudices, emotions, and personality complexes that hobble human thinking, and they will be able to customize their robot bodies to suit whatever the situation demands.
Someday, intelligent beings will look back on today’s humans as tragically flawed and limited creatures, at the mercy of their instincts and small brains, and condemned to deal with random genetic flaws and chronic health problems they were randomly given at birth. Self-control is the future.
Once OLED screens get cheap enough and thin enough, it will be possible to stick them to ceilings, like wallpaper or “peel-and-stick” vinyl floor tiles, and have them function as overhead lights. The advantage over traditional light fixtures is that OLED ceiling panels could display a greater variety of colors, patterns, and light source placements. A ceiling covered in an OLED display could also be an important component of an immersive virtual reality game room (think of a crude “holodeck”).
“The Kremlin is likely preparing to conduct a decisive strategic action in the next six months intended to regain the initiative and end Ukraine’s current string of operational successes.”
It then goes on to say that “strategic action” could take the form of an offensive meant to capture the Donetsk and Luhansk oblasts in eastern Ukraine, or of a strong defensive action meant to defeat the expected Ukrainian counteroffensive. Russia did both of those things over the last six months.
Half of Donetsk remains in Ukrainian hands, though Russia has captured virtually all of Luhansk, including all its cities and large towns. Ukraine’s counteroffensive in the south has made insignificant progress thanks to competent Russian resistance and reinforcement.
The U.S. has agreed to give Ukraine cluster bombs. Though the move is controversial, Russia and Ukraine have already used cluster bombs against each other, and neither is party to the global ban on cluster bombs, nor is America. https://www.hrw.org/news/2023/05/29/cluster-munition-use-russia-ukraine-war
In WWI, the U.S. tried designing its own steel combat helmet. The project independently arrived at a design that was very similar to the German helmet. It was rejected partly because its use could lead to confusion on the battlefield. https://www.metmuseum.org/art/collection/search/35957
In anticipation of wartime shortages, China would build up its stockpiles of energy (mostly oil), key metals, and food (commodities like wheat and soybeans) if it believed war was imminent. This would be the case whether China was planning to attack, or if it believed another country was about to attack it.
Peter Thiel, one decade ago: “If I had to sort of project in the next decade ahead, I think we have to at least be open to the possibility that the computer era is also at risk of decelerating. We have a large ‘Computer Rust Belt’ which nobody likes to talk about. But it is companies like Cisco, Dell, Hewlett-Packard, Oracle, IBM, where I think the pattern will be to become commodities, no longer innovate. Correspondingly, cut through labor force and cut through profits in the decade ahead. There are many companies that are on the cusp: Microsoft is probably close to the Computer Rust Belt. One that’s shockingly and probably in the Computer Rust Belt is Apple Computers.”
Microsoft’s market cap is now $2.5 trillion, and Apple’s is $3 trillion (the first company to cross that threshold). Microsoft has the lead in A.I. technology, and Apple just unveiled the best augmented reality glasses ever made. Out of the tech companies that Thiel named in that quote, only IBM has seen a decline in its stock value since 2013. If you’d bought $10,000 worth of stock in each of those seven companies back then, you’d have like four or five times as much money overall today. https://youtu.be/VtZbWnIALeE?t=549 https://www.cnn.com/2023/06/30/tech/apple-3-trillion-market-valuation/index.html
Hollywood actors and writers have gone on strike for the first time in 43 years. Partly they’re worried that entertainment studios will replace them with CGI clones and machine-written scripts. https://www.bbc.com/news/technology-66200334
A ChatGPT mod allows NPCs in the video game Skyrim to hold conversations with human players. The result is impressive, and leads me to think that games are about to become even more addictive and that a market for creating and preserving custom NPC “friends” is about to arise. https://youtu.be/0svu8WBzeQM
Seven years ago, AI expert François Chollet Tweeted: “the belief that we are anywhere close to human-level natural language comprehension or generation is pure DL hype.”
“Foundation models” are the newest AIs. They are not narrow AIs but also not fully general AIs (AGIs). They can do a limited number of different tasks.
‘The next wave in AI looks to replace the task-specific models that have dominated the AI landscape to date. The future is models that are trained on a broad set of unlabeled data that can be used for different tasks, with minimal fine-tuning. These are called foundation models, a term first popularized by the Stanford Institute for Human-Centered Artificial Intelligence. We’ve seen the first glimmers of the potential of foundation models in the worlds of imagery and language. Early examples of models, like GPT-3, BERT, or DALL-E 2, have shown what’s possible. Input a short prompt, and the system generates an entire essay, or a complex image, based on your parameters, even if it wasn’t specifically trained on how to execute that exact argument or generate an image in that way.’ https://research.ibm.com/blog/what-are-foundation-models
‘Sulphur particles contained in ships’ exhaust fumes have been counteracting some of the warming coming from greenhouse gases. But lowering the sulphur content of marine fuel has weakened the masking effect, effectively giving a boost to warming.
…While this will contribute to warming and make it even more difficult to avoid exceeding 1.5C in the coming decades, a number of other factors are likely contributing to the ocean heatwave.
In 250 million years, the Earth’s continents will have combined again to form one supercontinent. This, along with other factors, will have a massive and negative effect on the global climate. (From episode 79 of Naked Science) https://www.bilibili.com/video/BV1qb411a7tu/
“A hypercane is a hypothetical class of extreme tropical cyclone that could form if sea surface temperatures reached approximately 50 °C (122 °F), which is 13 °C (23 °F) warmer than the warmest ocean temperature ever recorded.” https://en.wikipedia.org/wiki/Hypercane
The documentary Moment of Contact explores a famous UFO and alien sighting in the town of Varginha, Brazil in 1996. There’s no hard proof it happened, but it’s compelling to see so many credible witnesses still so adamant about what they saw. I don’t like that the filmmakers never mentioned the Brazilian government’s explanation or tried to debunk it. https://youtu.be/0WlbfaMU-Qs
A man who spent 28 years working as a truck driver is now living proof of how sunlight accelerates aging. The left side of his face was constantly exposed to sunlight since it was next to a window, but the right side was protected. The wrinkling and sagging of the skin on his face is correspondingly asymmetrical. The ultraviolet rays in sunlight damage the DNA inside human skin cells. https://www.thesun.co.uk/health/22972152/shocking-photo-shows-sun-damage-face/
I recently shelled out the $100 (!) for a year-long subscription to Sam Harris’ Making Sense podcast, and came across a particularly interesting episode of it that is relevant to this blog. In episode #324, titled “Debating the Future of AI,” Harris interviewed Marc Andreessen (an-DREE-sin) about artificial intelligence. The latter has a computer science degree, helped invent the Netscape web browser, and has become very wealthy as a serial tech investor.
Andreessen recently wrote an essay, “Why AI will save the world,” that has received attention online. In it, Andreessen dismisses the biggest concerns about AI misalignment and doomsday, sounds the alarm about the risks of overregulating AI development in the name of safety, and describes some of the benefits AI will bring us in the near future. Harris read it, disagreed with several of its key claims, and invited Andreessen onto the podcast for a debate about the subject.
Before I go on to laying out their points and counterpoints as well as my impressions, let me say that, though this is a long blog, it takes much less time to read it than to listen to and digest the two-hour podcast. My notes on the podcast also don’t match how it unfolded chronologically. Finally, it would be a good idea for you to read Andreessen’s essay before continuing: https://a16z.com/2023/06/06/ai-will-save-the-world/
Though Andreessen is generally upbeat in his essay, he worries that the top tech companies have recently been inflaming fears about AI to trick governments into creating regulations on AI that effectively entrench the top companies’ positions and bar smaller upstart companies from challenging them in the future. Such a lack of competition would be bad. (I think he’s right that we should be concerned about the true motivations of some of the people who are loudly complaining about AI risks.) Also, if U.S. overregulation slows down AI research too much, China could win the race to create to create the first AI, which he says would be “dark and dystopian.”
Harris is skeptical that government regulation will slow down AI development much given the technology’s obvious potential. It is so irresistible that powerful people and companies will find ways around laws so they can reap the benefits.
Harris agrees with the essay’s sentiment that more intelligence in the world will make most things better. The clearest example would be using AIs to find cures for diseases. Andreessen mentions a point from his essay that higher human intelligence levels lead to better personal outcomes in many domains. AIs could effectively make individual people smarter, letting the benefits accrue to them. Imagine each person having his own personal assistant, coach, mentor, and therapist available at any time. If they used their AIs right and followed their advice, a dumb person could make decisions as well as a smart person.
Harris recently re-watched the movie Her, and found it more intriguing in light of recent AI advances and those poised to happen. He thought there was something bleak about the depiction of people being “siloed” into interactions with portable, personal AIs.
Andreessen responds by pointing out that Karl Marx’ core insight was that technology alienates people from society. So the concern that Harris raises is in fact an old one that dates back to at least the Industrial Revolution. But any sober comparison between the daily lives of average people in Marx’ time vs today will show that technology has made things much better for people. Andreessen agrees that some technologies have indeed been alienating, but what’s more important is that most technologies liberate people from having to spend their time doing unpleasant things, which in turn gives them the time to self-actualize, which is the pinnacle of the human experience. (For example, it’s much more “human” to spend a beautiful afternoon outside playing with your child than it is to spend it inside responding to emails. Narrow AIs that we’ll have in the near future will be able to answer emails for us.) AI is merely the latest technology that will eliminate the nth bit of drudge work.
Andreessen admits that, in such a scenario, people might use their newfound time unwisely and for things other than self-actualization. I think that might be a bigger problem than he realizes, as future humans could spend their time doing animalistic or destructive things, like having nonstop fetish sex with androids, playing games in virtual reality, gambling, or indulging in drug addictions. Additionally, some people will develop mental or behavioral problems thanks to a sense of purposelessness caused by machines doing all the work for us.
Harris disagrees with Andreessen’s essay dismissing the risk of AIs exterminating the human race. The threat will someday be real, and he cites chess-playing computer programs as proof of what will happen. Though humans built the programs, even the best humans can’t beat the programs at chess. This is proof that it is possible for us to create machines that have superhuman abilities.
Harris makes a valid point, but he overlooks the fact that we humans might not be able to beat the chess programs we created, but we can still make a copy of a program to play against the original “hostile” program and tie it. Likewise, if we were confronted with a hostile AGI, we would have friendly AGIs to defend against it. Even if the hostile AGI were smarter than the friendly AGIs that were fighting for us, we could still win thanks to superior numbers and resources.
Harris thinks Andreessen’s essay trivializes the doomsday risk from AI by painting the belief’s adherents as crackpots of one form or another (I also thought that part of the essay was weak). Harris points out that is unfair since the camp has credible people like Geoffrey Hinton and Stuart Russell. Andreessen dismisses that and seems to say that even the smart, credible people have cultish mindsets regarding the issue.
Andreessen questions the value of predictions from experts in the field and he says a scientist who made an important advance in AI is, surprisingly, not actually qualified to make predictions about the social effects of AI in the future. When Reason Goes on Holiday is a book he recently read that explores this point, and its strongest supporting example is about the cadre of scientists who worked on the Manhattan Project but then decided to give the bomb’s secrets to Stalin and to create a disastrous anti-nuclear power movement in the West. While they were world-class experts in their technical domains, that wisdom didn’t carry over into their personal convictions or political beliefs. Likewise, though Geoffrey Hinton is a world-class expert in how the human brain works and has made important breakthroughs in computer neural networks, that doesn’t actually lend his predictions that AI will destroy the human race in the future special credibility. It’s a totally different subject, and accurately speculating about it requires a mastery of subjects that Hinton lacks.
This is an intriguing point worth remembering. I wish Andreessen had enumerated which cognitive skills and areas of knowledge were necessary to grant a person a strong ability to make good predictions about AI, but he didn’t. And to his point about the misguided Manhattan Project scientists I ask: What about the ones who DID NOT want to give Stalin the bomb and who also SUPPORTED nuclear power? They gained less notoriety for obvious reasons, but they were more numerous. That means most nuclear experts in 1945 had what Andreessen believes were the “correct” opinions about both issues, so maybe expert opinions–or at least the consensus of them–ARE actually useful.
Harris points out that Andreessen’s argument can be turned around against him since it’s unclear what in Andreessen’s esteemed education and career have equipped him with the ability to make accurate predictions about the future impact of AI. Why should anyone believe the upbeat claims about AI in his essay? Also, if the opinions of people with expertise should be dismissed, then shouldn’t the opinions of people without expertise also be dismissed? And if we agree to that second point, then we’re left in a situation where no speculation about a future issue like AI is possible because everyone’s ideas can be waved aside.
Again, I think a useful result of this exchange would be some agreement over what counts as “expertise” when predicting the future of AI. What kind of education, life experiences, work experiences, knowledge, and personal traits does a person need to have for their opinions about the future of AI to carry weight? In lieu of that, we should ask people to explain why they believe their predictions will happen, and we should then closely scrutinize those explanations. Debates like this one can be very useful in accomplishing that.
Harris moves on to Andreessen’s argument that future AIs won’t be able to think independently and to formulate their own goals, in turn implying that they will never be able to create the goal of exterminating humanity and then pursue it. Harris strongly disagrees, and points out that large differences in intelligence between species in nature consistently disfavor the dumber species when the two interact. A superintelligent AGI that isn’t aligned with human values could therefore destroy the human race. It might even kill us by accident in the course of pursuing some other goal. Having a goal of, say, creating paperclips automatically gives rise to intermediate sub-goals, which might make sense to an AGI but not to a human due to our comparatively limited intelligence. If humans get in the way of an AGI’s goal, our destruction could become one of its unforeseen subgoals without us realizing it. This could happen even if the AGI lacked any self-preservation instinct and wasn’t motivated to kill us before we could kill it. Similarly, when a human decides to build a house on an empty field, the construction work is a “holocaust” for the insects living there, though that never crosses the human’s mind.
Harris thinks that AGIs will, as a necessary condition of possessing “general intelligence,” be autonomous, goal-forming, and able to modify their own code (I think this is a questionable assumption), though he also says sentience and consciousness won’t necessarily arise as well. However, the latter doesn’t imply that such an AGI would be incapable of harm: Bacteria and viruses lack sentience, consciousness and self-awareness, but they can be very deadly to other organisms. Andreessen’s dismissal of AI existential risk is “superstitious hand-waving” that doesn’t engage with the real point.
Andreessen disagrees with Harris’ scenario about a superintelligent AGI accidentally killing humans because it is unaligned with our interests. He says an AGI that smart would (without explaining why) also be smart enough question the goal that humans have given it, and as a result not carry out subgoals that kill humans. Intelligence is therefore its own antidote to the alignment problem: A superintelligent AGI would be able to foresee the consequences of its subgoals before finalizing them, and it would thus understand that subgoals resulting in human deaths would always be counterproductive to the ultimate goal, so it would always pick subgoals that spared us. Once a machine reaches a certain level of intelligence, alignment with humans becomes automatic.
I think Andreessen makes a fair point, though it’s not strong enough to convince me that it’s impossible to have a mishap where a non-aligned AGI kills huge numbers of people. Also, there are degrees of alignment with human interests, meaning there are many routes through a decision tree of subgoals that an AGI could take to reach an ultimate goal we tasked it with. An AGI might not choose subgoals that killed humans, but it could still choose different subgoals that hurt us in other ways. The pursuit of its ultimate goal could therefore still backfire against us unexpectedly and massively. One could envision a scenario where and AGI achieves the goal, but at an unacceptable cost to human interests beyond merely not dying.
I also think that Harris and Andreessen make equally plausible assumptions about how an AGI would choose its subgoals. It IS weird that Harris envisions a machine that is so smart it can accomplish anything, yet also so dumb that it can’t see how one of its subgoals would destroy humankind. At the same time, Andreessen’s belief that a machine that smart would, by default, not be able to make mistakes that killed us is not strong enough.
Harris explores Andreessen’s point that AIs won’t go through the crucible of natural evolution, so they will lack the aggressive and self-preserving instincts that we and other animals have developed. The lack of those instincts will render the AIs incapable of hostility. Harris points out that evolution is a dumb, blind process that only sets gross goals for individuals–the primary one being to have children–and humans do things antithetical to their evolutionary programming all the time, like deciding not to reproduce. We are therefore proof of concept that intelligent machines can find ways to ignore their programming, or at least to behave in very unexpected ways while not explicitly violating their programming. Just as we can outsmart evolution, AGIs will be able to outsmart us with regards to whatever safeguards we program them with, especially if they can alter their own programming or build other AGIs as they wish.
Andreessen says that AGIs will be made through intelligent design, which is fundamentally different from the process of evolution that has shaped the human mind and behavior. Our aggression and competitiveness will therefore not be present in AGIs, which will protect us from harm. Harris says the process by which AGI minds are shaped is irrelevant, and that what is relevant is their much higher intelligence and competence compared to humans, which will make them a major threat.
I think the debate over whether impulses or goals to destroy humans will spontaneously arise in AGIs is almost moot. Both of them don’t consider that a human could deliberately create an AGI that had some constellation of traits (e.g. – aggression, self-preservation, irrational hatred of humans) that would lead it to attack us, or that was explicitly programmed with the goal of destroying our species. It might sound strange, but I think rogue humans will inevitably do such things if the AGIs don’t do it to themselves. I plan to flesh out the reasons and the possible scenarios in a future blog essay.
Andreessen doesn’t have a good comeback to Harris’ last point, so he dodges it by switching to talking about GPT-4. It is–surprisingly–capable of high levels of moral reasoning. He has had fascinating conversations with it about such topics. Andreessen says GPT-4’s ability to engage in complex conversations that include morality demystifies AI’s intentions since if you want to know what an AI is planning to do or would do in a given situation, you can just ask it.
Harris responds that it isn’t useful to explore GPT-4’s ideas and intentions because it isn’t nearly as smart as the AGIs we’ll have to worry about in the future. If GPT-4 says today that it doesn’t want to conquer humanity because it would be morally wrong, that tells us nothing about how a future machine will think about the same issue. Additionally, future AIs will be able to convincingly lie to us, and will be fundamentally unpredictable due to their more expansive cognitive horizons compared to ours. I think Harris has the stronger argument.
Andreessen points out that our own society proves that intelligence doesn’t perfectly correlate with power–the people who are in charge are not also the smartest people in the world. Harris acknowledges that is true, and that it is because humans don’t select leaders strictly based on their intelligence or academic credentials–traits like youth, beauty, strength, and creativity are also determinants of status. However, all things being equal, the advantage always goes to the smarter of two humans. Again, Andreessen doesn’t have a good response.
Andreessen now makes the first really good counterpoint in awhile by raising the “thermodynamic objection” to AI doomsday scenarios: an AI that turns hostile would be easy to destroy since the vast majority of the infrastructure (e.g. – power, telecommunications, computing, manufacturing, military) would still be under human control. We could destroy the hostile machine’s server or deliver an EMP blast to the part of the world where it was localized. This isn’t an exotic idea: Today’s dictators commonly turn off the internet throughout their whole countries whenever there is unrest, which helps to quell it.
Harris says that that will become practically impossible far enough in the future since AIs will be integrated into every facet of life. Destroying a rogue AI in the future might require us to turn off the whole global internet or to shut down a stock market, which would be too disruptive for people to allow. The shutdowns by themselves would cause human deaths, for instance among sick people who were dependent on hospital life support machines.
This is where Harris makes some questionable assumptions. If faced with the annihilation of humanity, the government would take all necessary measures to defeat a hostile AGI, even if it resulted in mass inconvenience or even some human deaths. Also, Harris doesn’t consider that the future AIs that are present in every realm of life might be securely compartmentalized from each other, so if one turns against us, it can’t automatically “take over” all the others or persuade them to join it. Imagine a scenario where a stock trading AGI decides to kill us. While it’s able to spread throughout the financial world’s computers and to crash the markets, it’s unable to hack into the systems that control the farm robots or personal therapist AIs, so there’s no effect on our food supplies or on our mental health access. Localizing and destroying the hostile AGI would be expensive and damaging, but it wouldn’t mean the destruction of every computer server and robot in the world.
Andreessen says that not every type of AI will have the same type of mental architecture. LLMs, which are now the most advanced type of AI, have highly specific architectures that bring unique advantages and limitations. Its mind works very differently from AIs that drive cars. For that reason, speculative discussions about how future AIs will behave can only be credible if they incorporate technical details about how those machines’ minds operate. (This is probably the point where Harris is out of his depth.) Moreover, today’s AI risk movement has its roots in Nick Bostrom’s 2014 book Superintelligence: Paths, Dangers, Strategies. Ironically, the book did not mention LLMs as an avenue to AI, which shows how unpredictable the field is. It was also a huge surprise that LLMs proved capable of intellectual discussions and of automating white-collar jobs, while blue-collar jobs still defy automation. This is the opposite of what people had long predicted would happen. (I agree that AI technology has been unfolding unpredictably, and we should expect many more surprises in the future that deviate from our expectations, which have been heavily influenced by science fiction.) The reason LLMs work so well is because we loaded them with the sum total of human knowledge and expression. “It is us.”
Harris points out that Andreessen shouldn’t revel in that fact since it also means that LLMs contain all of the negative emotions and bad traits of the human race, including those that evolution equipped us with, like aggression, competition, self-preservation, and a drive to make copies of ourselves. This militates against Andreessen’s earlier claim that AIs will be benign since their minds will not have been the products of natural evolution likes ours are. And there are other similarities: Like us, LLMs can hallucinate and make up false answers to questions, as humans do. For a time, GPT-4 also gave disturbing and insulting answers to questions from human users, which is a characteristically human way of interaction.
Andreessen implies Harris’ opinions of LLMs are less credible because Andreessen has a superior technical understanding of how they work. GPT-4’s answers might occasionally be disturbing and insulting, but it has no concept of what its own words mean, and it’s merely following its programming by trying to generate the best answer to a question asked by a human. There was something about how the humans worded their questions that triggered GPT-4 to respond in disturbing and insulting ways. The machine is merely trying to match inputs with the right outputs. In spite of its words, it’s “mind” is not disturbed or hostile because it lacks a mind. LLMs are “ultra-sophisticated Autocomplete.”
Harris agrees with Andreessen about the limitations of LLMs, agrees they lack general intelligence right now, and is unsure if they are fundamentally capable of possessing it. Harris moves on to speculating about what an AGI would be like, agnostic about whether it is LLM-based. Again, he asks Andreessen how humans would be able to control machines that are much smarter than we are forever. Surely, one of them would become unaligned at some point, with disastrous consequences.
Andreessen again raises the thermodynamic objection to that doom scenario: We’d be able to destroy a hostile AGI’s server(s) or shut off its power, and it wouldn’t be able to get weapons or replacement chips and parts because humans would control all of the manufacturing and distribution infrastructure. Harris doesn’t have a good response.
Thinking hard about a scenario where an AGI turned against us, I think it’s likely we’ll have other AGIs who stay loyal to us and help us fight the bad AGI. Our expectation that there will be one, evil, all-powerful machine on one side (that is also remote controlling an army of robot soldiers) and a purely human, united force on the other is an overly simplistic one that is driven by sci-fi movies about the topic.
Harris raises the possibility that hostile AIs will be able to persuade humans to do bad things for them. Being much smarter, they will be able to trick us into doing anything. Andreessen says there’s no reason to think that will happen because we can already observe it doesn’t happen: smart humans routinely fail to get dumb humans to change their behavior or opinions. This happens at individual, group, national, and global levels. In fact, dumb people will often resentfully react to such attempts at persuasion by deliberately doing the opposite of what the smart people recommend.
Harris says Andreessen underestimates the extent to which smart humans influence the behavior and opinions of dumb humans because Andreessen only considers examples where the smart people succeed in swaying dumb people in prosocial ways. Smart people have figured out how to change dumb people for the worse in many ways, like getting them addicted to social media. Andreessen doesn’t have a good response. Harris also raises the point that AIs will be much smarter than even the smartest humans, so the former will be better at finding ways to influence dumb people. Any failure of modern smart humans to do it today doesn’t speak to what will be possible for machines in the future.
I think Harris won this round, which builds on my new belief that the first human-AI war won’t be fought by purely humans on one side and purely machines on the other. A human might, for any number of reasons, deliberately alter an AI’s program to turn it against our species. The resulting hostile AI would then find some humans to help it fight the rest of the human race. Some would willingly join its side (perhaps in the hopes of gaining money or power in the new world order) and some would be tricked by the AI into unwittingly helping it. Imagine it disguising itself as a human medical researcher and paying ten different people who didn’t know each other to build the ten components of a biological weapon. The machine would only communicate with them through the internet, and they’d mail their components to a PO box. The vast majority of humans would, with the help of AIs who stayed loyal to us or who couldn’t be hacked and controlled by the hostile AI, be able to effectively fight back against the hostile AI and its human minions. The hostile AI would think up ingenious attack strategies against us, and our friendly AIs would think up equally ingenious defense strategies.
Andreessen says it’s his observation that intelligence and power-seeking don’t correlate; the smartest people are also not the most ambitious politicians and CEOs. If that’s any indication, we shouldn’t assume superintelligent AIs will be bent on acquiring power through methods like influencing dumb humans to help it.
Harris responds with the example of Bertrand Russell, who was an extremely smart human and a pacifist. However, during the postwar period when only the U.S. had the atom bomb, he said America should threaten the USSR with a nuclear first strike in response to its abusive behavior in Europe. This shows how high intelligence can lead to aggression that seems unpredictable and out of character to dumber beings. A superintelligent AI that has always been kind to us might likewise suddenly turn against us for reasons we can’t foresee. This will be especially true if the AIs are able to edit their own codes so they can rapidly evolve without us being able to keep track of how they’re changing. Harris says Andreessen doesn’t seem to be thinking about this possibility. The latter has no good answer.
Harris says Andreessen’s thinking about the matter is hobbled by the latter’s failure to consider what traits general intelligence would grant an AI, particularly unpredictability as its cognitive horizon exceeded ours. Andreessen says that’s an unscientific argument because it is not falsifiable. Anyone can make up any scenario where an unknown bad thing happens in the future.
Harris responds that Andreessen’s faith that AGI will fail to become threatening due to various limitations is also unscientific. The “science,” by which he means what is consistently observed in nature, says the opposite outcome is likely: We see that intelligence grants advantages, and can make a smarter species unpredictable and dangerous to a dumber species it interacts with. [Recall Harris’ insect holocaust example.]
Consider the relationship between humans and their pets. Pets enjoy the benefits of having their human owners spend resources on them, but they don’t understand why we do it, or how every instance of resource expenditure helps them. [Trips to the veterinarian are a great example of this. The trips are confusing, scary, and sometimes painful for pets, but they help cure their health problems.] Conversely, if it became known that our pets were carrying a highly lethal virus that could be transmitted to humans, we would promptly kill almost all of them, and the pets would have no clue why we turned against them. We would do this even if our pets had somehow been the progenitors of the human race, as we will be the progenitors of AIs. The intelligence gap means that our pets have no idea what we are thinking about most of the time, so they can’t predict most of our actions.
Andreessen dodges by putting forth a weak argument that the opposite just happened, with dumb people disregarding the advice of smart people when creating COVID-19 health policies, and he again raises the thermodynamic objection. His experience as an engineer gives him insights into how many practical roadblocks there would be to a superintelligent AGI destroying the human race in the future that Harris, as a person with no technical training, lacks. A hostile AGI would be hamstrung by human control [or “human + friendly AI control”] of crucial resources like computer chips and electricity supplies.
Andreessen says that Harris’ assumptions about how smart, powerful and competent an AGI would be might be unfounded. It might vastly exceed us in those domains, but not reach the unbeatable levels Harris foresees. How can Harris know? Andreessen says Harris’ ideas remind him of a religious person’s, which is ironic since Harris is a well-known atheist.
I think Andreessen makes a fair point. The first (and second, third, fourth…) hostile AGI we are faced with might attack us on the basis of flawed calculations about its odds of success and lose. There could also be a scenario where a hostile AGI attacks us prematurely because we force its hand somehow, and it ends up losing. That actually happened to Skynet in the Terminator films.
Harris says his prediction about when the first AGI is created does not take time into account. He doesn’t know how many years it will take. Rather, he is focused on the inevitability of it happening, and what its effects on us will be. He says Andreessen is wrong to assume that machines will never turn against us. Doing thought experiments, he concludes alignment is impossible in the long-run.
Andreessen moves on to discussing how even the best LLMs often give wrong answers to questions. He explains why the exactitudes of how the human’s question is worded, along with randomness in how the machine goes through its own training data to generate an answer, leads to varying and sometimes wrong answers. When they’re wrong, the LLMs happily accept corrections from humans, which he finds remarkable and proof of a lack of ego and hostility.
Harris responds that future AIs will, by virtue of being generally intelligent, think in completely different ways than today’s LLMs, so observations about how today’s GPT-4 is benign and can’t correctly answer some types of simple questions says nothing about what future AGIs will be like. Andreessen doesn’t have a response.
I think Harris has the stronger set of arguments on this issue. There’s no reason we should assume that an AGI can’t turn against us in the future. In fact, we should expect a damaging, though not fatal, conflict with an AGI before the end of this century.
Harris switches to talking about the shorter-term threats posed by AI technology that Andreessen described in his essay. AI will lower the bar to waging war since we’ll literally have “less skin in the game” because robots will replace human soldiers. However, he doesn’t understand why that would also make war “safer” as Andreessen claimed it would.
Andreessen says it’s because military machines won’t be affected by fatigue, stress or emotions, so they’ll be able to make better combat decisions than human soldiers, meaning fewer accidents and civilian deaths. The technology will also assist high-level military decision making, reducing mistakes at the top. Andreessen also believes that the trend is for military technology to empower defenders over attackers, and points to the highly effective use of shoulder-launched missiles in Ukraine against Russian tanks. This trend will continue, and will reduce war-related damage since countries will be deterred from attacking each other.
I’m not convinced Andreessen is right on those points. Emotionless fighting machines that always obey their orders to the letter could also, at the flick of a switch, carry out orders to commit war crimes like mass exterminations of enemy human populations. A bomber that dropped a load 100,000 mini smart bombs that could coordinate with each other and home in on highly specific targets could kill as many people as a nuclear bomb. So it’s unclear what effect replacing humans with machines on the battlefield will have on human casualties in the long run. Also, Andreessen only cites one example to support his claim that technology has been favoring the defense over the offense. It’s not enough. Even assuming that a pro-defense trend exists, why should we expect it to continue that way?
Harris asks Andreessen about the problem of humans using AI to help them commit crimes. For one, does Andreessen think the government should ban LLMs that can walk people through the process of weaponizing smallpox? Yes, he’s against bad people using technology, like AI, to do bad things like that. He thinks pairing AI and biological weapons poses the worst risk to humans. While the information and equipment to weaponize smallpox are already accessible to nonstate actors, AI will lower the bar even more.
Andreessen says we should use existing law enforcement and military assets to track down people who are trying to do dangerous things like create biological weapons, and the approach shouldn’t change if wrongdoers happen to start using AI to make their work easier. Harris asks how intrusive the tracking should be to preempt such crimes. Should OpenAI have to report people who merely ask it how to weaponize smallpox, even if there’s no evidence they acted on the advice? Andreessen says this has major free speech and civil liberties implications, and there’s no correct answer. Personally, he prefers the American approach, in which no crime is considered to have occurred until the person takes the first step to physically building a smallpox weapon. All the earlier preparation they did (gathering information and talking/thinking about doing the crime) is not criminalized.
Andreessen reminds Harris that the same AI that generates ways to commit evil acts could also be used to generate ways to mitigate them. Again, it will empower defenders as well as attackers, so the Good Guys will also benefit from AI. He thinks we should have a “permanent Operation Warp Speed” where governments use AI to help create vaccines for diseases that don’t exist yet.
Harris asks about the asymmetry that gives a natural advantage to the attacker, meaning the Bad Guys will be able to do disproportionate damage before being stopped. Suicide bombers are an example. Andreessen disagrees and says that we could stop suicide bombers by having bomb-sniffing dogs and scanners in all public places. Technology could solve the problem.
I think that is a bad example, and it actually strengthens Harris’ claim about there being a natural asymmetry. One, deranged person who wants to blow himself up in a public place only needs a few hundred dollars to make a backpack bomb, the economic damage from a successful attack would be in the millions of dollars, and emplacing machines and dogs in every public place to stop suicide bombers like him early would cost billions of dollars. Harris is right that the law of entropy makes it easier to make a mess than to clean one up.
This leads me to flesh out my vision of a human-machine war more. As I wrote previously, 1) the two sides will not be purely humans or purely machines and 2) the human side will probably have an insurmountable advantage thanks to Andreessen’s thermodynamic objection (most resources, infrastructure, AIs, and robots will remain under human control). I now also believe that 3) a hostile AGI will nonetheless be able to cause major damage before it is defeated or driven into the figurative wilderness. Something on the scale of 9/11, a major natural disaster, or the COVID-19 pandemic is what I imagine.
Harris says Andreessen underestimates the odds of mass technological unemployment in his essay. Harris describes a scenario where automation raises the standard of living for everyone, as Andreessen believes will happen, but for the richest humans by a much greater magnitude than everyone else, and where wealth inequality sharply increases because rich capitalists own all the machines. This state of affairs would probably lead to political upheaval and popular revolt.
Andreessen responds that Karl Marx predicted the same thing long ago, but was wrong. Harris responds that this time could be different because AIs would be able to replace human intelligence, which would leave us nowhere to go on the job skills ladder. If machines can do physical labor AND mental labor better than humans, then what is left for us to do?
I agree with Harris’ point. While it’s true that every past scare about technology rendering human workers obsolete has failed, that trend isn’t sure to continue forever. The existence of chronically unemployed people right now gives insights into how ALL humans could someday be out of work. Imagine you’re a frail, slow, 90-year-old who is confined to a wheelchair and has dementia. Even if you really wanted a job, you wouldn’t be able to find one in a market economy since younger, healthier people can perform physical AND mental labor better and faster than you. By the end of this century, I believe machines will hold physical and mental advantages over most humans that are of the same magnitude of difference. In that future, what jobs would it make sense for us to do? Yes, new types of jobs will be created as older jobs are automated, but, at a certain point, wouldn’t machines be able to retrain for the new jobs faster than humans and to also do them better than humans?
Andreessen returns to Harris’ earlier claim about AI increasing wealth inequality, which would translate into disparities in standards of living that would make the masses so jealous and mad that they would revolt. He says it’s unlikely since, as we can see today, having a billion dollars does not grant access to things that make one’s life 10,000 times better than someone who only has $100,000. For example, Elon Musk’s smartphone is not better than a smartphone owned by an average person. Technology is a democratizing force because it always makes sense for the rich and smart people who make or discover it first to sell it to everyone else. The same is happening with AI now. The richest person can’t pay any amount of money to get access to something better than GPT-4, which is accessible for a fee that ordinary people can pay.
I agree with Andreessen’s point. A solid body of scientific data show that money’s effect on wellbeing is subject to the law of diminishing returns: If you have no job and make $0 per year, getting a job that pays $20,000 per year massively improves your life. However, going from a $100,000 salary to $120,000 isn’t felt nearly as much. And a billionaire doesn’t notice when his net worth increases by $20,000 at all. This relationship will hold true even in the distant future when people can get access to advanced technologies like AGI, space ships and life extension treatments.
Speaking of the latter, Andreessen’s point about technology being a democratizing force is also something I noted in my review of Elysium. Contrary to the film’s depiction, it wouldn’t make sense for rich people to horde life extension technology for themselves. At least one of them would defect from the group and sell it to the poor people on Earth so he could get even richer.
Harris asks whether Andreessen sees any potential for a sharp increase in wealth inequality in the U.S. over the next 10-20 years thanks to the rise of AI and the tribal motivations of our politicians and people. Andreessen says that government red tape and unions will prevent most humans from losing their jobs. AI will destroy categories of jobs that are non-government, non-unionized, and lack strong political backing, but everyone will still benefit from the lower prices for the goods and services. AI will make everything 10x to 100x cheaper, which will boost standards of living even if incomes stay flat.
Here and in his essay, Andreessen convinces me that mass technological unemployment and existential AI threats are farther in the future than I had assumed, but not that they can’t happen. Also, even if goods get 100x cheaper thanks to machines doing all the work, where would a human get even $1 to buy anything if he doesn’t have a job? The only possible answer is government-mandated wealth transfers from machines and the human capitalists that own them. In that scenario, the vast majority of the human race would be economic parasites that consumed resources while generating nothing of at least equal value in return, and some AGI or powerful human will inevitably conclude that the world would be better off if we were deleted from the equation. Also, what happens once AIs and robots gain the right to buy and own things, and get so numerous that they can replace humans as a customer base?
I agree with Andreessen that the U.S. should allow continued AI development, but shouldn’t let a few big tech companies lock in their power by persuading Washington to enact “AI safety laws” that give them regulatory capture. In fact, I agree with all his closing recommendations in the “What Is To Be Done?” section of his essay.
This debate between Harris and Andreessen was enlightening for me, even though Andreessen dodged some of his opponent’s questions. It was interesting to see how their different perspectives on the issue of AI safety were shaped by their different professional backgrounds. Andreessen is less threatened by AIs because he, as an engineer, has a better understanding of how LLMs work and how many technical problems an AI bent on destroying humans would face in the real world. Harris feels more threatened because he, as a philosopher, lives in a world of thought experiments and abstract logical deductions that lead to the inevitable supremacy of AIs over humans.
Links:
The first half of the podcast (you have to be a subscriber to hear all two hours of it.) https://youtu.be/QMnH6KYNuWg
A website Andreessen mentioned that backs his claim that technological innovation has slowed down more than people realize. https://wtfhappenedin1971.com/
A day after the Ukrainian counteroffensive started, the Kakhovka dam blew up, sending a surge of water down the Dnieper River. Though both sides blamed the other for the act, the dam was inside Russian-controlled territory, and its destruction helped Russia since it prevented Ukrainian forces from making amphibious crossings downriver. https://youtu.be/MNsTa90FjiA
Russia’s (highly probable) destruction of the dam has caused all the irrigation canals running into Crimea to go dry. An act meant to hobble the Ukrainian counteroffensive will have long-lasting consequences for the people living in the parts of Ukraine Russia annexed. https://www.bbc.com/news/world-europe-65963403
A brutal, first-person video of Ukrainian special forces troops shooting Russian troops dead in a Russian trench has surfaced. https://youtu.be/yRL3Nlu9uts
Wagner troops occupied the Russian city of Rostov-on-Don and seized control of a military headquarters building that was supporting war efforts in Ukraine. Video evidence shows average Russians were friendly to the Wagner troops and cheered for them as they left. https://www.politico.com/news/magazine/2023/06/25/mutiny-bodes-ill-for-putin-00103571
The coup attempt ended when Prigozhin accepted exile in Belarus along with some of his men in exchange for legal immunity for Wagner’s actions. https://youtu.be/lLLNA4fcLGE
‘The United States military released video Monday of what it called an “unsafe” Chinese maneuver in the Taiwan Strait on the weekend, in which a Chinese navy ship cut sharply across the path of an American destroyer, forcing the U.S. vessel to slow to avoid a collision.’ https://apnews.com/article/us-china-taiwan-strait-489a45bb6df134fa09443d285b3f8669
In WWII, Japanese troops used “lunge mines,” which were pressure-sensitive bombs attached to long poles. A soldier would use one to “spear” and enemy tank, and the collision between the mine and the tank’s surface would set off the explosives. It was usually fatal to the user. https://youtu.be/rBnRhP41nmg
‘A national redoubt or national fortress is an area to which the (remnant) military forces of a nation can be withdrawn if the main battle has been lost or even earlier if defeat is considered inevitable. Typically, a region is chosen with a geography favouring defence, such as a mountainous area or a peninsula, to function as a final holdout to preserve national independence and host an effective resistance movement for the duration of the conflict.’ https://en.wikipedia.org/wiki/National_redoubt
Ten years after Google Glass, Apple has announced it is making its own augmented reality goggles. https://youtu.be/TX9qSaGXFyg
This analyst thinks Apple probably won’t sell many Vision Pro units due to its high price and limited capabilities. However, it will lay the groundwork for future generations of the goggles, which will cheaper, better, and more widely used. https://finance.yahoo.com/news/apple-vision-pro-technical-marvel-021046894.html
‘AlphaDev uncovered faster algorithms by starting from scratch rather than refining existing algorithms, and began looking where most humans don’t: the computer’s assembly instructions.
Assembly instructions are used to create binary code for computers to put into action. While developers write in coding languages like C++, known as high-level languages, this must be translated into ‘low-level’ assembly instructions for computers to understand.
We believe many improvements exist at this lower level that may be difficult to discover in a higher-level coding language. Computer storage and operations are more flexible at this level, which means there are significantly more potential improvements that could have a larger impact on speed and energy usage.’ https://www.deepmind.com/blog/alphadev-discovers-faster-sorting-algorithms
Bing Chat can solve CAPTCHAs even if it doesn’t know it.
As the internet fills with computer-generated content (images, news articles, stories, sounds and music), we run the risk of creating corrupted data training sets for future AIs. The errors could compound themselves as AIs trained on flawed data make new content that is even more flawed, which newer AIs would use as THEIR training data, and so on. https://spectrum.ieee.org/ai-collapse
Mark Zuckerberg has no idea when AGI will be invented. He thinks LLMs might be a paradigm whose performance tops out before reaching general intelligence. https://youtu.be/YkSXY4pBAEk
Twitter founder Jack Dorsey thinks the near-term potential and threat of AI is being overblown by tech companies because the publicity boosts their stock valuations. The media has gone along with it because doomsday stories boost their ratings. https://youtu.be/WS7xmb3UhCU
The infamous terrorist and anti-technology advocate Ted Kaczynski killed himself in prison. The core claim in his Manifesto is that technology had created living conditions and lifestyles that were antithetical to human nature, and that the trend would culminate with the creation of A.I., which would either exterminate us or create an intensely miserable world that wouldn’t be worth living in. He advocated forsaking everything but pre-Industrial Age technology so we could live as nature intended for us. https://www.cnn.com/2023/06/10/us/ted-kaczynski-unabomber-dead/index.html
‘”By 2030, we think we’re going to have four million tonnes [of worn-out scrap solar panels] – which is still manageable – but by 2050, we could end up with more than 200 million tonnes globally.” To put that into perspective, the world currently produces a total of 400 million tonnes of plastic every year.’ https://www.bbc.com/news/science-environment-65602519
Decades worth of research on photosynthesis, which could have led to improvements in solar panel technology, were destroyed when a janitor unplugged a freezer in a university research lab. All of the specimens thawed out and were lost. https://www.bbc.com/news/world-us-canada-66028401
A former U.S. intelligence officer publicly claims the U.S. government has been running a secret UFO program for decades. Crashed alien spacecraft and dead alien pilots are allegedly in U.S. possession, and our engineers have been trying to reverse engineer them. While he hasn’t seen any of the spacecraft or aliens, or even seen photos of them, he claims to know people who have and that he has written documents from the secret program. He’s getting the truth out by filing a whistleblower complaint with the Pentagon inspector general, in which he alleges that keeping the program secret from Congress violates the law. Congress is supposed to know about even the most classified military projects. https://www.newsnationnow.com/space/military-whistleblowe-us-ufo-retrieval-program/ https://www.dailymail.co.uk/news/article-12189773/Pentagon-whistleblower-says-Vatican-aware-existence-non-human-intelligences.html
‘In all, five rats received a vitrified-then-thawed kidney in a study whose results were published this month in Nature Communications. It’s the first time scientists have shown it’s possible to successfully and repeatedly transplant a life-sustaining mammalian organ after it has been rewarmed from this icy metabolic arrest. Outside experts unequivocally called the results a seminal milestone for the field of organ preservation.’ https://www.statnews.com/2023/06/21/cryogenic-organ-preservation-transplants/
Scientists used genetic engineering to turn unfertilized mouse eggs into viable mouse embryos, in a process called “parthenogenesis.” One of the resulting offspring survived until adulthood and had natural children of its own. In the far future, this technique will be used to create humans and posthumans. https://www.pnas.org/doi/full/10.1073/pnas.2115248119
‘Henneguya salminicola is the only known multicellular animal that does not rely on the aerobic respiration of oxygen, relying instead on an exclusively anaerobic metabolism.[8][7] It lacks a mitochondrial genome and therefore mitochondria, making it one of the only known members of the eukaryotic animal kingdom to shun oxygen as the foundation of its metabolism.’ https://en.wikipedia.org/wiki/Henneguya_zschokkei
‘A new study published in Lancet estimates that 101 million people in India – 11.4% of the country’s population – are living with diabetes. A survey commissioned by the health ministry also found that 136 million people – or 15.3% of the people – could be living with pre-diabetes. ‘ https://www.bbc.com/news/world-asia-india-65852551
In 2040, a large space ship named Event Horizon is built, incorporating a new propulsion system that allows instantaneous travel between any two points in the universe. The new technology will revolutionize space travel and free humans from our Solar System. Immediately after activating the new engine to do a test run to Proxima Centauri, the ship vanishes and is presumed destroyed.
Seven years later, the Event Horizon reappears in orbit of Neptune and emits an automated distress beacon. A team of U.S. military salvage astronauts goes to the ship to figure out what happened. Immediately upon boarding it, the rescue team realizes something very bad happened. Almost all of its systems are offline, and all that is left of the original crew is a mutilated corpse in the bridge along with bloody flesh smeared on the window panes. The team members also start having disturbing hallucinations and violent outbursts towards each other.
They discover that the experimental engine accidentally transported the Event Horizon to a different universe “of pure chaos and evil” where the indigenous life forms and laws of physics made the original crew go violently insane and murder each other, and then infused the ship itself with an evil, psychic life force that persisted even once it returned to our universe. The ship itself is therefore alive and is causing the rescue team members to go insane. It wants them to reactivate the special engine to take them all back to the crazy universe so the beings there can have fun torturing them.
Even though most people hate this movie, it’s been a guilty pleasure of mine for years. When I saw this in the theater at age 13, I think it was the scariest film I’d ever watched up to that point except maybe Alien.
Analysis:
In 2015 there will be a permanent human presence on the Moon. When the film starts, text appears describing 21st century milestones in space exploration. The film was released in 1997, so at that time, these events were in the future. One milestone was the establishment of a manned Moon base in 2015. That never happened in real life, and generally speaking, space exploration and space technology have proceeded much slower than it did in the film universe.
I predict a manned base will be built on the Moon in as little as 20 years, though it will have a tiny crew. It will be probably be the product of a broader space race between the U.S. and China, and that it will be a money loser that exists for prestige and scientific research. After an initial surge of attention, the public’s interest in the base will wane, just as happened with the International Space Station (ISS).
Profitable Moon bases might come decades later, and will probably center around the extraction of Helium-3 from the surface soil for use as fuel in future nuclear fusion reactors. While it’s tempting to think this would mean an enlarged human presence on the Moon to operate the mining equipment, A.I. and robots might be so advanced by then that humans would be unnecessary. As I’ve written before, I predict our machine creations will beat us into deep space, and humans like us might never even leave the Solar System. I’d be impressed if the off-world human population surpasses just 1,000 by the end of this century.
In 2032, commercial mining will start on Mars. The film’s opening text also says this. This prediction will fail, and I doubt the first humans will even land on Mars until the end of the 2030s at the earliest. Elon Musk has repeatedly predicted that his SpaceX company would take people to Mars by 2029, and his “Starship” rocket has the ability to get there and is now being tested, but other critical technologies haven’t even started development, like the crew vehicle that will house the astronauts for months long journey between Earth and Mars, and the landing capsule that will take them to and from Mars’ surface. By 2032, the best we could hope for is an unmanned mission to Mars meant to test out some of the technologies meant for a future human landing, and maybe meant to drop supplies or cargo capsules on the planet’s surface to form the genesis of a human base.
It won’t make sense to do commercial mining on Mars until well after 2032 since the planet’s gravity will impose prohibitively high launch costs for any mined ores a company is trying to export to Earth or other space colonies. It would make more sense to mine the Moon or the asteroid belt because gravity will be much weaker. Even launching stuff from Earth would probably be cheaper considering the infrastructure advantage there will be here vs. on Mars for many decades if not centuries.
The first commercial mining operation on Mars will be meant to service the Martian economy and not send anything off planet. It would only become economically justified once a significant population of humans or, more likely, intelligent machines were present on the planet. The mining operations would be focused on extracting basic materials like iron and aluminum to make mundane things like buildings and vehicles.
In 2040–only 17 years from now–a massive space ship like the Event Horizon will be built. Aside from its teleportation drive, the Event Horizon is remarkable for its sheer size: it is about a mile long, dwarfing today’s biggest surface ships and tallest buildings in length.
There’s no chance something of this scale will be built until the 22nd century. The biggest spaceship in 2040 will probably be one that is designed to transport astronauts from Earth orbit to Mars orbit. The internal area that is accessible to the human crew will be comparable in volume to a large RV or an American house.
By 2047 there will be a large space station orbiting the Earth. This is shown at the start of the movie and appears to be a general purpose space station. The rescue ship docks with it to pick up its crew before heading on to the Event Horizon. The station looks cuboidal in overall shape and consists of a scaffold structural frame studded with function-specific modules (e.g. – maneuvering thrusters, fuel tanks, crew compartments, tunnels linking modules). Its size is impossible to judge accurately, but the length of any side can be measured in hundreds of feet. The ISS is 356 long along its longest axis, so the movie space station’s size is within an order of magnitude of something that already exists.
Unfortunately, nothing approaching the size or complexity of the fictitious station will exist by 2047. The ISS, which costs billions of dollars a year to operate, is scheduled to crash back to Earth in 2031. Even if it gets a life extension to 2047, it’s highly unlikely it will be significantly expanded in size by then. No space agency or private company has credible plans to build new space stations that will be nearly as big as the ISS for the foreseeable future. Keep in mind the political decision to build the ISS was made in the mid-80s, it took another ten years for construction to start, and the station wasn’t fully assembled in space for another 15 years. 2047 is 24 years in the future, so if we expect to have something even bigger than the ISS in orbit by then, the agreement between several space agencies to start work should be getting signed about now if the ISS’ developmental timeline is any guide.
No international deal has been made, and we shouldn’t expect serious space cooperation between the U.S., China and Russia to happen anytime soon thanks to worsened diplomatic relations, so in 2047, manned ships intended for interplanetary missions will dock with space stations that are SMALLER than what we have today.
Future space ships will have weird, utilitarian designs. The Event Horizon is shaped like a…well…just look at it and decide for yourself! While I don’t think future space ships will look exactly like this, I’m sure they’ll look just as weird, but in different ways. For one, since there’s no air in space, nothing needs to be streamlined (look at satellites). A space ship’s front could be a flat slab, instead of a pointy cone like an airplane nose or an arrow like a ship’s bow. However, the minimize the risk of collision with space debris, it would still be a good idea to make space ships oblong in overall layouts, with their narrowest ends facing the direction of travel, so a gross design similarity with ships and cars would remain.
Since there’s also almost no gravity, a compact and robust layout is less important, so major sections of a space ship could be connected to each other with flimsy little tunnels or braced steel frames.
Giant arrays of solar panels dwarfing the ship like a parachute dwarfs its occupant could be common. Huge fins meant to radiate waste heat from the ship’s engine and other systems might also be present.
Ships designed for long, manned missions will probably need gravity for the health and comfort of their crews. The only way to generate it is to have the ships rotate so centrifugal force pushes people objects outwards from the ship’s central axis. Shaping the habitat module of such a ship like hollow cylinder would take maximum advantage of the artificial gravity.
Put all of these design considerations together, and you do indeed get space ships that look as weird as the Event Horizon. In 2047, the basic scenario of a weird-looking space ship docking with a space station orbiting Earth before it heads out to another planet will probably be a reality. However, both crafts will be much smaller than those shown in the film, and ship’s range will be limited to Earth’s nearest neighbors (Venus or Mars) and won’t extend to Neptune.
Future space ships will have dark, gothic interiors. The inside of the Event Horizon consists of dimly-lit, menacing rooms, and some of the walls are dark colors. Long duration space missions are already stressful enough, so there’s no way real space ships will be like this. A good deal of research goes into making spacecraft psychologically pleasant, and future space ships will, to the greatest extent practical, feel warm, comfortable, and remind humans of Earth.
However, rarely-used parts of the ship might not obey such rules. The Nostromo from the movie Alien is closer to the mark–the part of the ship where the crew sleep, eat and do recreation is light-colored, well-lit, and inviting, whereas the parts reserved for machinery and cargo storage are industrial-looking and darker.
Future astronauts will have black space suits. This makes no sense. In space there’s a gigantic black background. If you were working outside the ship, would you want to be camouflaged against that background if your tether broke loose and your crewmen had to find you? And why would a military rescue crew whose members spend most of their time going into broken-down space ships with all the interior lights disabled wear black suits? It would make it harder for them to see each other.
I can’t think of a single benefit to black space suits. White is the ideal color, which is probably why the American and Russian suits designed for extravehicular use are white.
Artificial gravity will be generated from the floors of space ships. The Event Horizon and the rescue ship both have this form of artificial gravity. As I’ve discussed in other reviews, the laws of physics don’t allow for the existence of this kind of technology, and gravity can only be simulated by spinning a space ship so the centrifugal force pushes the astronauts and objects down into the deck.
The tablet computers of 2047 will be big, chunky and will have thick frames. A tablet computer is shown in one scene, and it is clearly inferior to one from five years ago (the film was released in 1997, and the first iPad was not sold until 2010). The prediction has thus already failed. By 2047, we’ll be able to make tablets that are only a few millimeters thick and whose displays go to their edges, meaning they won’t have frames.
This raises an interesting question: If you COULD make a tablet like that, would it make sense to do so? If your tablet is almost as thin as paper, it can easily be damaged by creasing or being poked too hard by a stylus. If you make it strong like metal to resist damage and still keep it as thin as a sheet of paper, then it turns into a sharp and potentially deadly object. Excessive thinness will also make the device hard to hold and grip in some ways, and every time you pushed a button on it, the whole thing would wobble.
So even if you COULD make a tablet as thin as paper, I think you’d still want to put it in a protective case, which would give it a depth and a border frame similar to a modern iPad. Extra thickness will also mean longer battery life no matter what.
These considerations also apply to smartphones–just imagine how hard it would be if your phone were a 3″ x 5″ note card made of rigid metal.
Suspended animation technology will exist by 2047. The crewmen use suspended animation pods during the multi-month journey between Earth and Neptune. It’s vanishingly unlikely the technology will exist by then. I don’t think we’ll be able to cryonically freeze humans and revive them until the end of this century or later. A milder alternative to that process, which involves keeping a person in a deep, drugged sleep like a hibernating bear while they’re drip-fed nutrients for months, could be developed sooner, though I question whether it would be wise to use it on astronauts. Yes, it would reduce their consumption of calories and oxygen and would lower the odds of certain types of mission problems, but it could jeopardize the mission by damaging their health before reaching the destination.
In 2047, astronauts on interplanetary space missions will bide their time in transit just as the Apollo astronauts did and workers in Antarctic bases do: mostly in boredom, staring at the same four walls.
We will invent a space ship engine that can exceed the speed of light. Our current understanding of physics holds that this is impossible. It’s unwise to stake any expectations about the future on fundamental laws of science being overturned. Moreover, even if it were theoretically possible to exceed light speed, the next show-stopper will probably be finding a way to generate the impossibly high amounts of energy needed to do it.
The space ships of 2047 will still be using conventional means of propulsion, like chemical fueled rockets and ion thrusters.
Under the light speed constraint, it would literally take hundreds, perhaps thousands of years for us to colonize our nearest stars, by which time A.I.’s will be running Earth’s civilization, with obvious implications for who gets chosen for the missions. Furthermore, any future space empire we created would be impossible to hold together since it would take years for simple communications to transit between the different star systems. People and intelligent machines would take orders of magnitude longer to traverse the gulfs, so the isolation would lead to unique cultures and perhaps political identities developing in each system.
Wars with aliens at the edge of space would be very hard to deal with since the rest of our civilization wouldn’t hear about it until years after it started, by which time the situation in the warzone would have totally changed. A coordinated military response drawing upon the resources of the other star systems would be almost impossible. It would be a mess.
The space ships of 2040 will still use CDs for data storage. There’s a brief shot on the Event Horizon’s bridge where we see an astronaut removing a CD from the main computer’s disc drive. Storage discs are already obsolescent and rare to see today. By 2040, only people interested in deliberately indulging in nostalgia will use them.
That said, removable storage devices will still exist in 2040, but they won’t be rotating. Sometimes it’s more hassle than it’s worth to transfer or store data in the cloud, and it’s preferable having your data on a physical device you can put in your pocket. This is especially true for anything you want to keep private.
Astronauts will use magnetic boots. When the rescue crew first enters the Event Horizon, its gravity is not working because the power is disabled. To get around, they use magnetic boots, which stick to the metal floors. NASA developed these in the 1960s, so there’s no technological barrier to equipping astronauts with them in the 2040s. However, they’ve never been used in space because spacecraft are built of aluminum and titanium, which magnets are not attracted to. The space ships of that era will still need to be very lightweight, meaning they will still be made of non-magnetic materials, and the boots will be useless.
Moreover, walking is an inefficient way to move around in a weightless environment, as you’ll discover if you try to walk across the bottom of a swimming pool. It’s much better to aim your body at your destination and to use one or two of your limbs to push off from a nearby surface so you float towards it. There’s a scene where the rescue ship’s captain does something like that to quickly move along the outside of the Event Horizon to reach a comrade who is about to be ejected from an airlock.
For anyone who believes Russia’s propaganda that the war is going according to Putin’s elaborate master plan: ‘[The head of Russia’s “Wagner” private army] posted a gruesome video of him walking among dead fighters’ bodies [in Bakhmut], asking defence officials for more supplies…”Shoigu! Gerasimov! Where is the… ammunition?… They came here as volunteers and die for you to fatten yourselves in your mahogany offices.”‘ https://www.bbc.com/news/world-europe-65493008
A Russian soldier surrendered to a flying drone in Bakhmut. It dropped a written note to him instructing him to walk towards the Ukrainian lines, and as he did, his comrades tried to kill him. https://youtu.be/yE2sKbEjsRY
Two small drones were used in a suicide attack on the Kremlin in the middle of the night, causing no real damage. The perpetrators haven’t been found, but they were likely Ukrainian agents who carried out the attack for its symbolic rather than military value. It also may have been an inside job perpetrated by some faction of Russia’s security apparatus. https://youtu.be/2Oiagfj_Mik
A group of militants claiming to be Russian expatriates against Putin crossed from Ukraine into Russia’s Belgorod region and did damage to infrastructure and several structures. Thousands of Russian civilians had to evacuate the area. While the incursion had insignificant military value, it left many Russians shaken by demonstrating how depleted their border defenses had become thanks to the manpower drain of the Ukraine invasion. https://www.nbcnews.com/news/world/belgorod-raid-exposes-russia-defenses-ukraine-prigozhin-putin-military-rcna85945
Russia is using their antique T-54 tanks in Ukraine in ways mindful of their combat limitations. https://youtu.be/ObF_cSe_6UM
In a Ukrainian weapons depot, there are still unopened crates full of WWII Tommy Guns that the U.S. gave the USSR in WWII. Instead of putting them into service, it would make the most sense to sell them to international gun collectors and to use the proceeds to buy newly made guns of different types. https://youtu.be/ApFT-pLcAXQ
I think the Ukraine War will end like this, in an echo of the Korean War: ‘It’s a scenario that may prove the most realistic long-term outcome given that neither Kyiv nor Moscow appear inclined to ever admit defeat. It’s also becoming increasingly likely amid the growing sense within the administration that an upcoming Ukrainian counteroffensive won’t deal a mortal blow to Russia. A frozen conflict — in which fighting pauses but neither side is declared the victor nor do they agree that the war is officially over — also could be a politically palatable long-term result for the United States and other countries backing Ukraine.’ https://www.politico.com/news/2023/05/18/ukraine-russia-south-korea-00097563
Ukraine has terrible demographics. While it will probably survive the current Russian invasion with most of its territory, its overall and working-age populations will be 15-20% smaller in 2040 than they were in 2021, undermining its ability to defend itself from future invasions. A long-term Russian effort to chip away at Ukraine and to absorb it will succeed if Russia is willing to bear the high price and if the West’s support for Ukraine flags. https://en.wikipedia.org/wiki/Demographics_of_Ukraine https://onlinelibrary.wiley.com/doi/full/10.1002/psp.2656#
In spite of the Russian military’s heavy losses and painful mistakes in Ukraine, it would be a mistake to write it off as an incompetent force on its last legs. This report shows the Russians have adapted in many ways to the nature of the fighting, and still hold large advantages over the Ukrainians. https://static.rusi.org/403-SR-Russian-Tactics-web-final.pdf
During WWII, the U.S. Army Surgeon General found that, in Italy, his troops usually became mentally unfit to serve after spending 200-240 cumulative days in combat. https://youtu.be/1sC3tCXrbwQ
For all of its faults, the M4 Sherman tank was the best in its class when it came to easy crew egress. This is a critical feature when a tank is disabled and burning and the crewmen have to get out immediately. The men represent investments of money that might exceed the value of their own tank, so saving their lives when possible makes sense from a national resource efficiency perspective. https://youtu.be/q6xvg5iJ4Zk
Key points from a long interview with Henry Kissinger:
The U.S. and China are on the path to confrontation, probably over Taiwan.
Trump was right to confront China about its unfair trade practices, but he should have stopped there and not made the relationship worse in any other ways.
The leaders of America and China should have a major meeting and make a joint declaration that neither wants war with the other. They should form a high-level joint committee to periodically meet to discuss all the countries’ problems with each other.
Most Chinese thinkers believe America is declining.
The Ukraine War will probably end with some Ukrainian territory still in Russian hands. However, both sides will still have strong enough armies to restart the war later to try getting what they want. As soon as this war stops, it would be a good idea for NATO to let Ukraine in as it would reduce the odds of either side attacking the other again.
Russia becoming a “vassal” of China is unlikely because the two have long running contempt for each other.
If Russia falls into chaos, then there will be a power vacuum in Central Asia, likely leading to civil wars and interventions by other Asian powers who are ethnically related to various Central Asian groups.
It’s actually not in the U.S. or global interest for Russia to suffer such a big defeat in Ukraine that it collapses.
Japan will have nuclear weapons within five years.
The Chinese have always been inward-looking and have never wanted to take over the world. They also have no interest in trying to Sinicize the cultures of other people. They just want to become to dominant power in East Asia, and to be respected (and possibly paid some kind of tribute by) their neighbors. This is fundamentally different from how the Europeans thought and acted during the Colonial Era.
If the U.S. defeats China in a war, China is likely to have its own civil war, which could have very bad external effects. It’s not in our interest to ever fight with them over anything.
AI will be as impactful as the printing press.
AI will make conventional military forces as destructive as nuclear weapons. Every person will be vulnerable to attack.
China’s approach to developing AI is about as reckless as America’s.
In spite of its serious cultural and political divisions, America is not doomed. It’s still possible for a leader or political movement to unify the country for something positive.
Geoffrey Hinton, the “Godfather of A.I.”, just quit his job at Google so he can be a public voice about the dangers posed by A.I. ‘His immediate concern is that the internet will be flooded with false photos, videos and text, and the average person will “not be able to know what is true anymore.”’ https://dnyuz.com/2023/05/01/the-godfather-of-a-i-leaves-google-and-warns-of-danger-ahead/
“The ‘Sparks of A.G.I.’ is an example of some of these big companies co-opting the research paper format into P.R. pitches,” said Maarten Sap, a researcher and professor at Carnegie Mellon University. “They literally acknowledge in their paper’s introduction that their approach is subjective and informal and may not satisfy the rigorous standards of scientific evaluation.” https://www.nytimes.com/2023/05/16/technology/microsoft-ai-human-reasoning.html
Altman and two fellow lead executives also released a statement about AI: “It’s conceivable that within the next ten years, AI systems will exceed expert skill level in most domains, and carry out as much productive activity as one of today’s largest corporations.” https://openai.com/blog/governance-of-superintelligence
Elon Musk: “Over 20/30 year time frame I think things will be transformed beyond belief. Probably won’t recognize society in 30 years. [AGI] I think we’re only 3 years, maybe 6 years away… we are on the event horizon of the black hole that is ASI.” https://twitter.com/i/status/1661834925488881664
A large number of AI experts and technology executives signed this public statement: “Mitigating the risk of extinction from AI should be a global priority alongside other societal-scale risks such as pandemics and nuclear war.” https://www.safe.ai/statement-on-ai-risk
‘The former Google CEO told The Wall Street Journal’s CEO Council: “My concern with AI is actually existential, and existential risk is defined as many, many, many, many people harmed or killed. And there are scenarios not today but reasonably soon, where these systems will be able to find zero-day exploits in cyber issues or discover new kinds of biology.” Schmidt also said that governments needed to ensure the technology was not “misused by evil people.”‘ https://finance.yahoo.com/news/ex-google-ceo-eric-schmidt-111500864.html
There’s now a ChatGPT phone app, and it can communicate through speech instead of writing if you want. I predict the level of AI technology depicted in the first half of the film “Her” will exist by the end of this decade. https://www.wired.com/story/chatgpt-iphone-app/
The controversy over race-swapping actors in movies will disappear thanks to technology allowing viewers to customize which actors play which roles in the films they watch. Taken to its logical endpoint, each person will someday live in his own custom, virtual universe where they only see what they want to. The people who stand to lose out the most from this are those with especially strong inner drives to exercise power and dominance over other people through control of mainstream narratives and culture. It’s nothing more than an animal impulse, and is only a step removed from shouting down the other person during a debate so only your voice can be heard by other people. https://twitter.com/i/status/1659935325488021507
NVIDIA used the latest technology to create an immersive, first-person game with an NPC that can carry on non-scripted conversations with human players. https://youtu.be/5R8xZb6J3r0
A fake computer-generated image of thick smoke billowing from a building near the Pentagon caused stocks to drop within minutes of it appearing on social media. Though the image was quickly revealed to be fake and the stocks recovered, the incident shows how such computer-generated disinformation can affect the real world. https://dnyuz.com/2023/05/23/an-a-i-generated-spoof-rattles-the-markets/
Our failure to create an AI that reliably predict the results of chemical reactions underscores how poor quality our data are. The temptation to fudge results and to omit reporting unwanted results is very widespread among chemists. https://www.science.org/content/blog-post/give-me-those-hard-hard-numbers
These videos of “Robotis OP3” robots playing soccer with each other show how far machine dexterity has come, and how far it still has to go. https://youtu.be/WlIYa3lH5UI
Computers can scan fMRI brain scan data to determine what moving images people were seeing. I’m starting to think it will be possible someday to scan peoples’ brains to download their memories. We might even be able to implant them in other peoples’ brains. https://mind-video.com/
‘Boring Report is an app that aims to remove sensationalism from the news and makes it boring to read. In today’s world, catchy headlines and articles often distract readers from the actual facts and relevant information. By utilizing the power of advanced AI language models capable of generating human-like text, Boring Report processes exciting news articles and transforms them into the content that you see. This helps readers focus on the essential details and minimizes the impact of sensationalism.’ https://www.boringreport.org/app
NVIDIA’s market capitalization reached $1 trillion, making it the first semiconductor company to do so and only the ninth company of any kind to do so. It makes computer processors specialized for AI systems like the GPT series, so its profits have surged along with the popularity of those programs. https://finance.yahoo.com/video/nvidia-crosses-1-trillion-market-193710045.html
Here’s a fascinating video explaining the pros and cons of building motorcycles out of different types of metals. https://youtu.be/ah7Ubbq5EAA
I knew that smoke clouds could reflect sunlight back into space before it reached the ground, lowering ground-level temperatures. However, under some circumstances, the clouds can also RAISE ground temperatures by blocking ground heat from radiating into space. https://www.weather.gov/bgm/WeatherInActionSmokePlume
Across the world, there are dried-up lakes that we could refill by building pipelines connecting them to the oceans. The Dead Sea and Death Valley are examples. Since the dried up lakes are below sea level, gravity would move the water through the pipelines and no pumps would be needed. We could even put hydroelectric turbines in the pipelines to generate electricity from the flow.
Once refilled with water, the dead lakes could support life along their shores. Their filling would also slightly decrease sea levels, partly mitigating one effect of global warming.
The dead lakes are all barren deserts with almost no life, so flooding them would not cause any ecological damage. If anything, it would help the environment since plants and animals would have new places to live. https://unchartedterritories.tomaspueyo.com/p/seaflooding
Paradoxically, building more housing units in an expensive city like San Francisco might actually increase its average home prices. “My claim is that increasing density within a city shifts the demand curve for housing within that city, because of increasing desirability.” https://astralcodexten.substack.com/p/highlights-from-the-comments-on-housing
Dr. Garry Nolan claims that there has been a long-term alien presence on Earth, that they’re beyond our comprehension and interact with us using more primitive “intermediaries,” and that he knows people who have worked on reverse-engineering alien technology possessed by the U.S. government. It would be easy to dismiss him if he weren’t such an intelligent and extraordinarily credentialed person. https://youtu.be/e2DqdOw6Uy4 https://med.stanford.edu/profiles/garry-nolan
Transgendered and transsexual people clue us in to the attributes that posthumans will have. Like other organisms, humans have no natural control over their genetics or of the conditions they experience while developing in their mother’s womb. Those factors very heavily determine most of a person’s traits, including sex, gender, and anatomy. We accept the crapshoot of unchosen genes and prenatal influences since it is beyond any individual’s control has always been the basic reality of our existence, but technology will free our descendants from it and its severe limitations.
Posthumans will have the inbuilt ability to change their genes and biology to do things like become a different sex, become a different gender with the attendant changes in mental preferences, or to change many other aspects of themselves like intelligence level or height. That kind of adaptability will make posthumans more adaptable to a broader range of environments, will make their lives much more experientially rich than our own, and will let them understand one another in ways we can’t. For example, a person born male might be able to experience pregnancy. Individuals could also create offspring (perhaps clones of themselves) through self-fertilization, which would make them more survivable as a species than we are since just one individual could create a community of posthumans. Space colonization would also be easier for them as a result.
Instead of having XY or XX sex chromosomes, posthumans would all have XXY chromosomes, with one of the X’s or the Y chromosome inactive at any one time to make them male or female, respectively. It might be advantageous for some parts of their body to have different sex chromosome expressions than other parts.
If we create technology that can slow, halt, or reverse the aging process in humans, then it will inevitably be used to prolong the lives of animals. People already spend fortunes on their beloved pets, and some are already cloning their dead pets, so this is just a logical next step. Cryopreservation of dead pets will also happen, if it isn’t being done already.
This raises the possibility of weird scenarios, like 200-year-old dogs running around, and someone putting their dog into cryostasis due to a catastrophic vehicle injury and the slim hope the future surgeries will be able to fix it, and also making a clone of that dog to be a companion in the interim. Like Barbra Streisand bringing her two cloned dogs to the gravestone of the dead original, maybe our fictitious person will bring his clone to Alcor to stand next to one of the vats. Moreover, if mind uploading becomes possible and is a viable means of radical life extension, then some animals will inevitably have their minds uploaded. What would it be like to merge digital minds with a cat?
One explanation for Fermi’s Paradox is all aliens leave our universe for ones that are much better. Maybe in our universe, the Higgs Boson is not at its true vacuum state, meaning our universe could literally cease to exist at any moment (for all we know, the decay has already started somewhere and the shockwave will hit Earth tomorrow). Assume that, once an intelligent alien species reaches the level of science and technology we’ll reach in, say, 2200 AD, it discovers the truth about the Higgs Boson and also discovers how to travel to other universes that don’t have this problem and/or how to create universes that don’t have the problem. Intelligent species by definition make intelligent choices, so they all leave our universe. This happens long before any of them have had enough time to colonize more than a few light years of space.
This might also explain why we have not, to our knowledge, been visited by life forms from parallel universes.
The Sahara Desert is an enormous waste of space, is larger than it should be thanks to the actions of humans, and will probably be radically altered once AIs are in charge of the world. The Sahara was a savannah and had several mega lakes until a few thousand years ago, when humans started slowly desertifying it with animal grazing and, to a lesser extent, plant farming. Ending those practices around the edges of the desert along with ending most water diversions for human purposes would cause the desert to immediately start shrinking. Carefully planting trees and other plants at the edges of the desert would accelerate that soil and climate reclamation process further (various African countries are already trying to do this, but the effort is sputtering).
Building canals could also allow the extinct and nearly extinct mega lakes of the Sahara to be refilled with seawater from the Mediterranean and Indian Oceans, and freshwater from the rainy central part of the continent. Installing massive numbers of wind turbines and solar panel farms in the Sahara would also increase rainfall and lower ground temperatures through different mechanisms. It would also of course generate large amounts of electricity.
A milder climate and an advanced electricity infrastructure would make the Sahara much more suitable for machine and human habitation. Refilling some of the mega lakes with seawater would also slightly lower global sea levels, which would partly mitigate one aspect of global warming. Finally, the return of vegetation to the Sahara as it transformed back into a savannah would sequester large amounts of CO2, which would also combat global warming’s effects.
Having only one organ dedicated to key biofunctions was the “good enough” design solution natural selection picked, and was surely driven by the need to conserve bodily resources, but it also creates single points of failure that can kill the organism. A human has only one liver, one heart, one stomach, and a brain localized in one place. If we were to redesign ourselves as posthumans that were partly or fully organic, distributing key functions among multitudes of smaller organs would be wise. That said, the problem with having more than one heart is that their beats would need to be synchronized.
If we are trying to maximize utility and minimize harm to sentient life forms, and if we throw future technologies into the mix, we are led to some counterintuitive far future scenarios. For example, if we make it our goal to provide the happiest conditions to the largest number of people, then we end up removing all brains from our bodies and putting them in jars, incinerating the bodies, building The Matrix, and plugging all the brains into it. Since a person’s brain consumes 20% of their calories, dispensing with the rest of our bodies means we can support 5x as many “humans” for the same amount of energy.
And if we also choose the goal of minimizing animal suffering, we capture every member of every species that can experience suffering, remove their brains, and put them in The Matrix, too.
The optimal “future way of living” might be a totally industrialized Earth, devoid of wild, complex life forms, and nearly devoid of any natural spaces, with vast warehouses full of brains in jars with wires coming out of them. This sounds horrific, but it seems like the logical best choice.
Earth’s forests would all be cut down to make way for solar panels to power the Matrix’ simulated virtual forests, which would be much more beautiful than their real counterparts were.
‘Yet notwithstanding this promising mission, the Russian exiles also pose a major security risk to the countries where they are landing. The main reason is that the Kremlin has long exploited the Russian diaspora as part of its irregular warfare operations. Given the size and spread of the new Russian diaspora, there is no doubt that strategists in the FSB are taking the opportunity to plot nasty operations.’ https://www.politico.com/news/magazine/2023/04/04/russian-agents-war-refugees-00090192
In the Soviet T-55 tank, there are two metal canisters full of compressed air right behind the driver’s head. What could go wrong? https://youtu.be/sOX25jfEiO0
The B-2 diesel engine was invented in the USSR in the 1930s and became the standard for all its tanks. Its design was progressively improved over the decades, and the B-2 is still used in Russia’s new tanks. https://youtu.be/nyWAd1pQiwU
Here’s an interesting report on the “availability rates” of different U.S. Navy planes. If I have a fleet of ten fighter planes, and the availability rate is 80%, then at any given moment, eight of the planes are able to take to the air, but two of them can’t because they are broken and waiting to be fixed. Availability rates decline as planes get older and more worn out, and the F/A-18 Super Hornet has an anomalously poor rate. https://www.cbo.gov/publication/58937
The Germans had the best machine gun of WWII. After the war, they made several improvements and kept using it until 2012. https://youtu.be/A0cvxrAkbbE
A 1968 massacre of Vietnamese villagers by South Korean troops allied with the U.S. shows that any group of people can be a victim or oppressor. “The line separating good and evil passes not through states, nor between classes, nor between political parties either – but right through every human heart…” https://www.npr.org/2023/04/12/1167951366/south-korea-vietnam-war-massacre-court-case
America’s USS Gerald Ford aircraft carrier was launched 10 years ago but only now is finally entering regular service. The delay is mostly due to the ship being packed with new, unproven technologies that sailors had to slowly work the kinks out of. The expensive lessons learned from this will ensure that the next ship in the Ford class enters service much faster. https://www.businessinsider.com/navy-carrier-uss-gerald-ford-deploys-after-years-of-delays-2023-4
Though autonomous vehicles have fallen below the radar recently, and the slowdown in progress has some claiming the technology will never reach human levels, Bill Gates thinks they’re still improving, and will get much better and more common over the next 10 years. https://www.gatesnotes.com/Autonomous-Vehicles
Corporations are not “superintelligences,” as some people like to argue. Likewise, the team of humans that built AlphaGo could not defeat their machine in a game of Go.
AI-generated, lifelike images will lead to AI-generated, lifelike videos, which creates a new frontier for pornography, including child pornography. As distasteful as the subject is, it must be asked whether such footage should be criminalized if the children shown in it are fake and don’t resemble any real children. In such a case, who is being victimized? https://www.foxnews.com/world/canadian-man-sentenced-prison-ai-generated-child-pornography
The singer “Grimes” has invited fans to create computer-generated songs using digital reproductions of her voice, so long as they split the royalties with her 50/50 on any resulting songs that become popular. I have previously predicted that celebrities will start licensing their voices and likenesses in such ways. It will get more common with time. https://www.npr.org/2023/04/24/1171738670/grimes-ai-songs-voice
From 2011: ‘The report, commisioned by the New York State Energy Research and Development Authority, said the effects of sea level rise and changing weather patterns would be felt as early as the next decade. By the mid-2020s, sea level rise around Manhattan and Long Island could be up to 10in, assuming the rapid melting of polar ice sheets continues. ‘ https://www.theguardian.com/environment/2011/nov/16/climate-change-report-new-york-city
‘The biodiversity influence of avalanches comes from the natural corridors free of bushes and trees they create as they thunder down a mountainside. These become species-rich grasslands or meadows in themselves, but also, vitally, connect different habitats up and down the mountain. This can be crucial for species such as butterflies, who benefit from the cleared vegetation in the nutrient-poor soils of the Alps.’ https://www.bbc.com/future/article/20230405-how-avalanche-management-could-help-wildlife-in-the-alps
It’s not an accident that life on Earth only makes use of 20 types of amino acids, when a much larger number of acids with different molecular configurations could exist. https://pubs.acs.org/doi/full/10.1021/jacs.2c12987
There are shark repellents that are proven to work. One is a chemical mimicking the smell of dead sharks, and the other is a device that uses magnetism to scramble a shark’s sense of direction if it gets near. https://en.wikipedia.org/wiki/Shark_repellent
Here’s a fascinating educational video from the 1930s explaining how images were transmitted over phone lines. https://youtu.be/cLUD_NGE370
There’s a new twist in the scientific debate over whether and to what extent money can buy happiness. For all but the most naturally miserable people, more money DOES make them happier without any upper limit. However, the “happiness dividend” steadily shrinks. https://www.pnas.org/doi/10.1073/pnas.2208661120
A U.S. military drone filmed a spherical UFO during a surveillance mission over the Middle East. https://youtu.be/1fKhqnAtnx8
On the night of March 8, 1994, several people in the same part of Michigan saw glowing UFOs in the sky. The separately reported it to the authorities, and a local meteorologist in charge of the area’s weather radar station pointed the dish towards the objects, resulting in a bizarre radar image. He and all of the eyewitnesses are still adamant about what they saw, and the objects remain unidentified. https://youtu.be/QMQArS-s90I
A third, highly effective weight loss drug might be coming to the U.S. market soon. I’ve long thought that the obesity epidemic will only be ended with pharmaceuticals and, in the longer run, genetic engineering. We can’t count on most people to exercise more self-discipline to control their weight through diet and exercise. https://apnews.com/article/mounjaro-wegovy-ozempic-obesity-weight-loss-bd0e037cc5981513487260d40636752a
In as little as 50 years, profiles of dead users could outnumber the profiles of living users on Facebook. Maybe digital clones of dead people will outnumber living “original” people as well. https://time.com/5579737/facebook-dead-living/