Moritz Wallawitsch

Wiki Contributions

Comments

A plea for solutionism on AI safety

Why would it not be relevant to the question? What's the value of only looking at eliminating the potential risk?

Regulating a technology is not just about eliminating the risks of it but about reducing the risks to some extent while still enabling the upside. the upsides need to be clearly analyses and acknowledged.

A plea for solutionism on AI safety

Also, as Marc Andreessen points out in his piece, AI can also increase safety (this point seems unaddressed in your essay): https://pmarca.substack.com/p/why-ai-will-save-the-world

A plea for solutionism on AI safety

I don't understand the point you're trying to make. Is it "safety is good"? That seems pretty obvious? 

I think the problem is that some people think the state should regulate/interfere with how safe something can or should be. Related: https://worksinprogress.co/issue/anti-growth-safetyism

How can we classify negative effects of new technologies?

The existence of most of the ones you listed sounds questionable.  

How about economic risk exposure (for a given person/city/state)? I think there is already a ton of research on this.

E.g. funding some new nuclear power research could provide a 10000x ROI but .0000X% danger of destroying the city/area of the research facility.

We Should Break Up Elite Colleges

Sounds reasonable.  However, a better long-term strategy seems to be complete privatization. I.e. to remove the subsidies and tax breaks. I think Brian Caplan would support this strategy (see his book on the education system).