These Are The Events That Will Happen Before 2045 (Singularity)

preview_player
Показать описание
These are the events that will happen before 2045.

If we don't blow ourselves up, if an asteroid doesn't strike us down, if a virus doesn't eradicate the human species, or if any other catastrophic event, whether manmade or natural, doesn't occur, it's extremely likely that technological progress will not only continue, but accelerate.

What will life be like, in 2045? "Prediction is very difficult, especially, if it's about the future." This quote, often attributed to Niels Bohr, captures the essence of our challenge when trying to foresee the events that will happen before this seemingly arbitrary year.

So, what's so special about 2045? Why not try predicting events in 2050, or 2100?

In a post-AGI world, if humans are still part of it, it will most certainly mean we have solved the AI alignment problem. A machine capable of outmatching any human in intelligence will create the blueprint for the machine that makes the next machine to achieve any of its goals.

Automation will skyrocket, leading to what is called a "post-work society," where machines perform every mundane and even intellectual task. Any existing job will be performed much better by a machine, potentially leading to widespread unemployment.

Even intellectual tasks will be outstripped by AI. Obviously, such an intelligent AGI will also be used for good, such as solving complex societal problems like disease.
But even if we solve the AI alignment problem, our current political system will need drastic changes to counter the effects of a few companies owning AGI, creating trillions of dollars in wealth while the rest of society is left to starve.

A sort of wealth redistribution system, perhaps even devised by AGI, will need to be enacted to address every job being lost to machines.
But even if this political system problem is solved, some critics argue that this would create a false Utopia, as many of the now 8 billion-plus free of disease tourists who play Frisbie all the time, still find meaning in their work.

#AI #sciencetime #AGI
Рекомендации по теме
Комментарии
Автор

5:05 “…further democratising the creation of visual content.” This is one of the most bone-chilling things I’ve heard. Anyone can change anything to be what they want? Truth will cease to exist. It terrifies me.

DarylShires
Автор

Quick question: How would companies make money when almost everyone’s lost their job to a machine?

umidevon
Автор

Remember if we do end up capable of making a sentient AI, what you have to be afraid of is if it can start inventing and developing products without human help, then it could turn out dangerous and see us as a hindrance, especially if we try to get in its way.
Luckily enough that time is not anywhere near yet. The complexities of our brain is far beyond what computer programming can achieve. It probably will not be in our lifetime. Algorithms will get more advanced, but the inside of thousands of processors just isn’t capable of producing a Sentient being.
It will be trapped inside a supercomputer if it ever did happen, and that is hard to vision how you could get a brain that works like ours but has no senses or emotions inside a massive computer.

Vile_Entity_
Автор

What's crazy to me, is that humans won't even be a pin prick in the history of the universe and that one day in the relatively near future, we won't exist at all.

dizzyupthefool
Автор

There is virtually no one that will slow AI down, no one that will regulate it. All the pressure will be on making it first and that means it will not be aligned with any goal humanity would have.

sojournern
Автор

peak civilization will be when humans no longer have to work jobs if they don’t want to. technology is supposed to work for us, not against us. we will know that humanity has truly reached a pinnacle when that day comes. most people work jobs that they despise to simply put food on the table. it’s an unfortunate reality but one that we must have to keep society running as it is currently. but I hope and believe that there will be a day that technology will be able to do the services/jobs for us and we can focus our time and energy on other things instead of working a 9-5, 6 days a week. but of course, if you want to work for whatever reason (maybe you don’t have any hobbies), then I think that option should still be available.

jaylenadegoke
Автор

It seems a bit utopian to me to think about superintelligence and a singularity because it’s possible that it’s impossible for a singularity to be born from a binary and logically structured programming language. Human intelligence, our mind, our consciousness, whatever you want to call it, does not work in binary with 0s and 1s. There is more than logic and reasoning in our intelligence, there is the notion of concepts and the depth of knowledge… For example, imagine a program that plays chess, it is better than any human who plays chess, but the question we must ask ourselves is that the program does not know what chess is, it just calculates… Imagine a person who wants to become a sexologist, she can read all the books in the world and watch all the movies, if she has never made love, she cannot understand in the same way that a machine cannot have a concept of good or evil, a concept at the level of emotions, reasoning, etc. And it is quite possible that it is impossible for such intelligence or reasoning (singularity) to be born from a logical program created by man that is binary and structured, etc. I believe we are more likely to see very fast and stupid robots than intelligent ones, etc.

cloudstrife
Автор

Not sure why they wouldn't make any highly advanced AI with Asimov's 3 laws hard coded into them, that could save an awful lot of trouble.

"A robot may not injure a human being or, through inaction, allow a human being to come to harm. A robot must obey orders given it by human beings except where such orders would conflict with the First Law. A robot must protect its own existence as long as such protection does not conflict with the First or Second Law."

digitalfootballer
Автор

Considering the situation, I strongly support the idea of Roko's Basilisk and embracing the Singularity

PuerAeternus.
Автор

I'll probably be dead by 2045 so I hope we'll reach technical singularity by then.

twirlyspitzer
Автор

The doom loop music really sets the mood

BuildingTimeFreedom
Автор

Where do you get high quality images for your thumbnail? How to make thumbnails like Science Time?

aadhunikfacts
Автор

It is fascinated to imagine what could happen between now and 2045. The problem is that we the humans are to stupid and We might end up destroying ourselves way before 2045. I really miss my cave days.

spacecommander
Автор

7:29 the most important problem our species has to resolve is conflict and war, with of course the threat of nuclear war. Until this occurs the AI alignment problem can’t be fully solved.

justgarry
Автор

I really prefer to envision 2045 without us. For the benefit of all other living beings on this planet.

suzanak
Автор

Life extension and rejuvenation. Repairing the bugged mess that is the human genome and brain.

Khannea
Автор

We are a species used to slow gradual change that is easily denied. That makes it difficult for us collectively to go through rapid change as we are now and will as it progresses. Considering all the variables we have to endure first, (i.e. "if we don't blow ourselves up, if.... if....") we need solutions sooner rather than later. We should have already solved some of these problems. Unfortunately we have the tendency to keep doing business as usual. People wanted to "go back to normal" during covid, failing to realize that normal is killing us. We're about to witness that for ourselves.

SethrenJensen
Автор

If it seems so dangerous or unpredicteble, why do we keep creating and impeoving AI technologies?

katenowak
Автор

Seems like this planet has been too corrupted to salvage...this is not my world. I'm just visiting.

xgutkhj
Автор

Rogue A.I. is not a solution to the Fermi Paradox. Even if all those alien civilizations would have been wiped out by rebellious A.I., we would still see those A.I. doing stuff in place of their dead creators, like building Dyson Spheres or moving stars around.

TrangleC