All of Donald Hobson's Comments + Replies

The Price of Progress

Our primitive monkey brains are good at over-estimating very unlikely risks.[2]

 

I think this is presupposing the question isn't it. 

If a risk is indeed very unlikely, then we will tend to overestimate it. (If the probability is 0 it's impossible to underestimate) 

But for risks that are actually quite likely, then we are more likely to underestimate them.

And of course, bias estimates cut both ways. "Our primitive monkey brains are good at ignoring and underestimating abstract and hard to understand risks".

The real data wall is billions of years of evolution

Ok. Firstly I do think your "Embodied information" is real. I just think it's pretty small. You need the molecular structure for 4 base pairs of DNA, and for 30 ish protiens. And this wikipedia page. https://en.wikipedia.org/wiki/DNA_and_RNA_codon_tables

That seems to be in the kilobytes. It's a rather small amount of information compared to DNA.

Epigenetics is about extra tags that get added. So theoretically the amount of information could be nearly as much as in the DNA. For example, methyization can happen on A and C, so that's 1 bit per base pair, in th... (read more)

Will we ever run out of new jobs?

https://www.lesswrong.com/posts/ZiRKzx3yv7NyA5rjF/the-robots-ai-and-unemployment-anti-faq

 

Once AI does get that level of intelligence, jobs should be the least of our concerns. Utopia or extinction, our future is up to the AI.

Safe Stasis Fallacy

 > It also seems vanishingly unlikely that the pressures on middle class jobs, artists, and writers will decrease even if we rolled back the last 5 years of progress in AI - but we wouldn't have the accompanying productivity gains which could be used to pay for UBI or other programs.  

 

When plenty of people are saying that AGI is likely to cause human extinction, and the worst scenario you can come up with is middle class jobs, your side is the safe one. 

The Case for Positive-Sum Environmentalism

I think your notion of "environmental progress" itself is skewing things. 

When humans were hunter gatherers, we didn't have much ability to modify our surroundings. 

Currently, we are bemoaning global warming, but if the earth was cooling instead, we would bemoan that too. 

Environmentalism seems to only look at part of the effects.

No one boasts about how high the biodiversity is at zoos. No one is talking about cities being a great habitat for pigeons as an environmental success story. 

The whole idea around the environmentalist movement ... (read more)

What if government worked like Wikipedia?

One system I think would be good is issue based voting.

 

So for example, there would be several candidates for the position of "health minister", and everyone gets to vote on that. 

And independently people get to vote for the next minister for education.

 

Other interesting add ons include voting for an abstract organization, not a person. One person that decides everything is an option on the balot, but so are various organizations, with various decision procedures. You can vote for the team that listens to prediction markets, or even some sub... (read more)

The Bridge to Boundless Human Development

Solving global warming

Most of the current attempts that interact with everyday people are random greenwashing trying to get people to sort recycling or use paper straws. Yes solar panel tech is advancing, but that's kind of in the background to most peoples day to day lives. 

And all this goal is promising is that things won't get worse via climate change. It isn't actually a vision of positive change. 

A future with ultra cheap energy, electric short range aviation in common use etc. 

 

Building true artificial intelligence (AGI, or artifi

... (read more)
1Coleman Foley24dThanks Donald, good feedback. I agree about maximizing good over minimizing bad. Curing aging, or extending healthspan, is a great one. Certainly an easier sell than becoming a multiplanetary species.
Dude, where is my self-driving train?

within human laws

I have no idea what superintelligence following existing laws even looks like. 

Take mind uploading. Is it 

  1. Murder
  2. A (currently unapproved) medical procedure
  3. Not something the law makes any mention of, so permitted by default. 

The current law is very vague with respect to tech that doesn't exist yet. And there are a few laws which, depending on interpretation, might be logically impossible to follow. 

 

ASI by default pushes right to the edge of what is technically possible, not a good fit with vague rules. 

1Kevin Kohler2moYou are raising good questions, though they are probably beyond the scope for me to answer. My high-level take would be that there are quite a few existing laws that could apply in such a scenario (eg Neuralink-implants to record brain-activity need FDA approval) and that we should expect laws to be adapted to new circumstances caveated with the pacing problem.
Dude, where is my self-driving train?

So, even if Sam Altman would declare tomorrow that he has built a digital God and that he will roll it out for free to everyone, this would not immediately lead to full automation.

Superintelligent AI (whether friendly or not) may not feel obliged to follow human laws and customs that slow down regular automation.  

1Kevin Kohler2moThat is a very fair point! I guess even within human laws there is some point before "God-level" where the "automation overhang" is reduced when AI becomes so good that it can compete with the product/services of many companies end-to-end rather than relying on integration into business processes. Still, I think it's fair to say that a) business integration can be/is a bottleneck to automation and b) "automation overhang" differs between products/service based on market structure (eg lower in management consulting, higher in public transport)
Don’t be scared by the AI fearmongering.

>Throughout history, fearmongering has been used to justify a lot of extreme measures.

And throughout history, people have dismissed real risks and been caught with their pants down. What, in 2018 or Feb 2020 would appear to be pretty extreme measures at pandemic prevention would make total sense from our point of view.

 

Countries can and do spend a huge pile of money to defend themselves from various things. Including huge militaries to defend themselves from invasion etc. 

 

All sorts of technologies come with various safety measures. ... (read more)

The Offense-Defense Balance Rarely Changes

Well 

>Invader countries have to defend their conquests and hackers need to have strong information security.

One place where offense went way ahead of defense is with nukes. 

However nukes are sufficiently hard to make that only a few big powers have them. Hence balance of power MAD. 

If destruction is easy enough, someone will do it.

In the war example, as weapon lethality went up, the fighters moved further apart. So long as both sides have similar weapons and tactics, there exists some range at which you aren't so close as to be instakilled, nor are you so far as to have no hope of attacking. This balance doesn't apply to civilian casualties. 

Report on the Desirability of Science Given Risks from New Biotech

The thing is, we have many options that aren't just accelerating or decelerating the whole thing. Like we can choose gain of function research and cutting edge AI capabilities, and accelerate everything except that. 

Science is lots of different pieces, differential technological development.

 

 "25% probability that the domain experts are right x 50% chance that it’s not too late for science to affect the onset of the
time of perils x 50% chance that science cannot accelerate us to safety = 6.25%"

This smells of the "multistage fallacy"

You think... (read more)

Neither EA nor e/acc is what we need to build the future

The thing about e/acc is it's a mix of the reasonable and the insane doom cult. 

The reasonable parts talk about AI curing diseases ect, and ask to speed it up. 

Given some chance of AI curing diseases, and some chance of AI caused extinction, it's a tradeoff. 

Now where the optimal point of the tradeoff lands on depend on whether we just care about existing humans, or all potential future humans. And also on how big we think the risk of AI extinction is.

If we care about all future humans, and think ai is really dangerous, we get a "proceed wit... (read more)

Neither EA nor e/acc is what we need to build the future

I don't see any specific criticism of effective altruism other than "I don't like the vibes".

And the criticism from "acrimonious corporate politics". 

"Helen Toner was apparently willing to let OpenAI be destroyed because of a general feeling that the organization was moving too fast or commercializing too much."

Between the two of them, a philosophy that aims to prevent catastrophic risk in the future seems to be creating its own catastrophes in the present.

Shutting down a company and some acrimonious board room discussion is hardly "catastrophic"... (read more)

Why Governments Can't be Trusted to Protect the Long-run Future

It could be an externality, if the land was randomly reassigned a new owner every year or something. But if the land is sold, that is taken into account. It isn't an externality. Capitalism has priced this effect in.

Why Governments Can't be Trusted to Protect the Long-run Future

For something that is long term, but only effects the property of 1 person, like the field example, the market prices it in and it's not an externality.

No one is significantly incentivized to stop climate change, because no one person bears a significant fraction of the damage caused.

Politicians are far from perfect, but at least they have any incentive to tackle these big problems at all.

1Maxwell Tabarrok1yWell if it affects one plot of land that is currently the property of just 1 person it can still be an externality because lots of different people will own this land in the future.
Radical Energy Abundance

I have yet to see a good case against AI doom. 

The Next Einstein Could Be From Anywhere: Why Developing Country Growth Matters for Progress

I don't think most of these "next einstein" arguments prove what you think they do. 

If you want to increase the chances of string theory breakthroughs, you want to find the sort of people that have a high chance of understanding string theory, and push them even further. If any genetic component is relatively modest, then it becomes mostly pick someone, and throw lots of resources at educating them. If genetics or randomness control a lot, but are easily observed, then it's looking out for young maths prodigies and helping them. 

Ensuring widely s... (read more)

Small signs you live in a complacent society

Have you considered printing off a few sheets of paper, getting some glue, and just adding a few signs yourself? ;-)

Why slow progress is more dangerous than fast progress

Some tech, like seatbelts, are almost pure good. Some techs, like nukes are almost pure bad. Some, like cars, we might want to wait until we develop seatbelts and traffic lights for before we use widely. It depends on the technology. 

Philosophy in Space

Elon musk is very good at making himself the center of as many conversations about technology as possible. 

He should not be taken as a source of information of any reliability. 

 

Living on mars with tech not too far beyond current tech is like living in antarctica today. It's possible, but it isn't clear why you would want to. A few researchers on a base, not much else. 

Think ISS but with red dust out the windows. 

At some point, which might be soon or not so soon, tech is advanced enough that it becomes easy to get to mars. But at t... (read more)

1Kassi Dick10moThanks for your comment! The goal should be to colonize our solar system before advancing to another. Learning to develop an atmosphere would be man's greatest achievement because then we would not be dependent on our planet but we could colonize any planet with the right composition/distance from its star etc. Space travel is Man's next journey and I'm here for it, if that means downloading my consciousness onto a self-replicating robot, count me in haha although I'm only studying to be a pilot; I don't pretend to be a computer engineer or neurologist.
Will Technology Keep Progressing? (A Happier World video)

I don't think the "aside from the internet, nothing much". Firstly comuter and internet tech have been fairly revolutionary across substantial chunks of industry and our daily lives. This is the "a smartphone is only 1 device so doesn't count as much progress" thinking. Without looking at the great pile of abacuses and slide rules and globes and calculators and alarm clocks and puzzle toys and landline phones and cameras and cassette tapes and ... that it replaced and improved on. 

Secondly, there are loads of random techs that were invented recently, ... (read more)

Against Altruism

If longtermists existed back when blacks were widely regarded as morally inferior to whites, would the moral calculus of the longtermists have included the prosperity of future blacks or not? It seems like it couldn't possibly have included that. More generally, longtermism can't take into account progress in moral knowledge, nor what future generations will choose to value. Longtermists impose their values onto future generations.

It is true that we can't predict future moral knowledge. However.

  1. An intervention by someone from that time period that helps mo
... (read more)